tRelational / DPS Adabas-to-Oracle Success in South Africa

by Hans-Peter Will, Senior Technical Representative and Joseph Brady, Manager of Marketing and Technical Documentation at Treehouse Software, Inc.

Recently, Hans-Peter Will, Senior Technical Representative for Treehouse Software, traveled to South Africa to assist our partner Bateleur Software (pty) Ltd. with setting up a large public-sector customer’s data replication implementation using our Adabas-to-RDBMS tool set, tRelational / DPS (Data Propagation System).

Arriving in Johannesburg, Peter met with representatives from Bateleur and the IT Organization, where the key players discussed how tRelational / DPS was going to be used in the project. The customer initially wanted to populate sample data into Oracle, so Peter configured tRelational / DPS to process one of the smaller Adabas files to generate some data. He also recommended running an analysis with tRelational to determine whether the file contained a unique key. Peter took this opportunity to show the customer what other benefits they could realize out of the analysis information. Interestingly, they used a personnel file for analysis, and Peter was immediately able to show that 23 of the records had no gender entry and 180 of records had no surname. The customer was very pleased to see these revelations in the first analysis, and looked forward to identifying other data quality issues before commencing data replication.

Customer Replication Scenario with Treehouse Software Product Set…

_0_tReDPS_Replication_Scenario

The next step was to build the target structure in accordance with the Oracle DBA’s requirements. The DBA had specified that all columns were to be defined as VARCHAR2, except the date information. After the first model was completed, DDL and DPS parameters were generated and a quick materialization of data accomplished the desired result.

At the subsequent kickoff meeting, Peter provided a complete tRelational / DPS overview and discussed the target structure with the attendees and the Oracle DBA. The rest of the day was spent doing Adabas file implementations, analysis and modeling.

Setup was then completed for transferring extracted and transformed Adabas data into the customer’s Windows environment. Adabas Vista is used, so that one logical Adabas file was actually split into two files stored on different Adabas databases, and the customer wanted to combine them into the same target table in Oracle. While there was no unique descriptor, it was discovered that three fields in combination would make a unique key, enabling the model to be created to combine data from the separate physical files into a single Oracle table.

The team proceeded with file implementation, modeling, mapping, DPS executions, and resolving data issues. Various issues that were encountered, like invalid tab characters within the data, negative personnel numbers, duplicates in unique keys (maintained by the application) and the need to add an extra column to the output. These issues were resolved quickly by the customer’s staff.

Within a day, all the files were materialized and the PLOG copy process was modified so that from that point forward, every PLOG copy would automatically be processed through DPS Propagation to update the RDBMS on the target Windows machine.

The next day, Peter was asked by the customer, “How many of the files have been processed so far?”. Peter was pleased to report that every file was processed and was propagating successfully. The happy customer remarked that they never had a project that was completed this far ahead of the deadline.

Throughout the project, Peter never personally laid hands on a customer keyboard, but instead sat with staff, effectively training them and handing over comprehensive knowledge of tRelational / DPS. The customer was very excited to learn that their personnel can now easily use the product set to do any remaining work on their own.

A few days later, we received an e-mail from Bateleur:

“I had a very pleasant meeting with the customer today. They used tRelational to reject the non-unique keys, reran the Materialization, and reran DPS plus update into Oracle. The month-end update of Oracle that was taking nearly three days to complete, now takes five Minutes! Everyone delighted!”

Bjørn (Sam) Selmer-Olsen, Managing Director, Bateleur Software (pty) Ltd


About Treehouse Software’s tRelational / DPS Product Set

tReDPS_DIAGRAM

tRelational / DPS is a robust product set that provides modeling and data transfer of legacy Adabas data into modern RDBMS-based platforms for Internet/Intranet/Business Intelligence applications. Treehouse Software designed these products to meet the demands of large, complex environments requiring product maturity, productivity, feature-richness, efficiency and high performance.

The tRelational component provides complete analysis, modeling and mapping of Adabas files and data elements to the target RDBMS tables and columns. DPS (Data Propagation System) performs Extract, Transformation, and Load (ETL) functions for the initial bulk RDBMS load and incremental Change Data Capture (CDC) batch processing to synchronize Adabas updates with the target RDBMS.

Visit the Treehouse Software website for more information on tRelational / DPS, or contact us to discuss your needs.

TREETIP: Integrate Mainframe Data Sources In Your Big Data Initiatives

tcVISION supports a vast array of integration scenarios throughout the enterprise, providing easy and fast data migration for mainframe application modernization projects and enabling bi-directional data replication between mainframe, Linux, Unix and Windows platforms. This innovative technology offers comprehensive abilities to identify and capture changes occurring in mainframe and relational databases, then publish the required information to an impressive variety of targets, both on-premise and Cloud-based.

Analysts have observed that perhaps 80 percent of the world’s corporate data still resides on mainframes. So it’s no surprise that Bloor Research (http://www.bloorresearch.com/research/spotlight/big-data-and-the-mainframe/), notes that “it is necessary today to place the mainframe as a ‘first-class player’ in any enterprise Big Data strategy.”

In February 2017 we highlighted tcVISION’s support for replication to the leading NoSQL database MongoDB. MongoDB continues to increase in popularity as a back end for operational applications with real-time requirements.

tcVISION also supports analytics and “mainframe offload” Big Data use cases that generally leverage Hadoop HDFS and/or streaming data transport. With tcVISION, data from a wide variety of IBM mainframe data source can be quickly and easily replicated to Big Data targets, requiring minimal mainframe know-how and having minimal impact on the mainframe.

___tcVISON_Big_Data_001

Boost the return on investment for your Big Data initiatives using tcVISION!


Find out more about tcVISION — Enterprise ETL and Real-Time Data Replication Through Change Data Capture

tcVISION provides easy and fast data migration for mainframe application modernization projects and enables bi-directional data replication between mainframe, Linux, Unix and Windows platforms.

_0_tcVISION_Simple_Diagram

tcVISION acquires data in bulk or via change data capture methods, including in real time, from virtually any IBM mainframe data source (Software AG Adabas, IBM DB2, IBM VSAM, IBM IMS/DB, CA IDMS, CA Datacom, even sequential files), and transform and deliver to virtually any target. In addition, the same product can extract and replicate data from a variety of non-mainframe sources, including Adabas LUW, Oracle Database, Microsoft SQL Server, IBM DB2 LUW and DB2 BLU, IBM Informix, and PostgreSQL.


__TSI_LOGO

Visit the Treehouse Software website for more information on tcVISION, or contact us to discuss your needs.

Treehouse Software Onsite Training Classes are Available

New installation of a Treehouse Software product? New employees? Need a refresher course? Want to explore untapped product features? Three-to-five-day training sessions are available for tRelational / DPS and other Treehouse products.

_0_Software_Training

Treehouse Software training sessions are customized to meet your site’s unique environment and needs. As an example, the following three-day tRelational / DPS training class was conducted at a large University where they are using tRelational / DPS to create an archive of their legacy Adabas data to Oracle in preparation for their upcoming mainframe system retirement.

Over the course of the three-day session, the Treehouse instructor covered:

  • tRelational AUTOGEN feature (auto-generation of complete RDBMS schemata – Tables, Columns, Primary Key, and Foreign Keys – based upon existing ADABAS file structures). AUTOGEN was quite attractive since it will be a very time efficient way to create the Adabas-to-Oracle mappings.
  • The customer wanted to use Relational / DPS to automate as much of the process as possible, and take advantage of the batch functions of tRelational. Generally, the workshops of the class are all online, but in this case, the class ran through one online example after the File Implementation section of the training, and then jumped to the batch section of the training to show how that works. The class then returned to run additional File Implementations, Analysis, and Reporting in batch.
  • The class continued through the Modeling and Mapping section, and ran more batch jobs with some additional Adabas files.
  • Admin and Configuring DPS Parameters sections were next on the agenda. Because of the desire to automate as much as possible, the work focused on how to set up a single job stream that could be generated programmatically, where it was fed in the name of the file to be implemented and then do the AUTOGEN all in the same job. The tRelational AUTOGEN user exit was set up so the customer could identify and cater to the possibility of duplicate table names where multiple Adabas files have the same named MU (like “COMMENTS”); and handle the addition of an ETR (external transformation routine) being automatically added (there were some known HEX values in some fields) and then update the DPSCOLLGTH.
  • The class finished up the Materialization (initial loading of data) training section and then set up the jobs to run Materialization, and installed DPSSPLIT (used to separate the control and data files into one control and one data file per table materialized) on one of their servers. Generally, most class attendees are not the ones responsible for running the Materialization and Propagation (ongoing synchronization of data) jobs, so the instructor doesn’t usually set them up, but in this case, it made sense. The class started out by running the Materialization from a full database ADASAV taken over the weekend. The output of the Materialization job was FTPed to their Server and DPSSPLIT run against it to create the individual loader files. GENDDL (generate the control statements that can be passed to the RDBMS to define the tables and columns) was also run and that output transferred. After setting up the SQL job to process the DDL, the Oracle SQL*LOADER was successfully run to load the data.

Finally, the class went through some additional tRelational / DPS features, and as a final exercise, set up a Propagation job and ran it against a current PLOG just to make sure it worked satisfactorily.


__TSI_LOGO

If you are interested in finding out more about Treehouse training classes for your product, contact Treehouse Software today for more information, or to schedule onsite training.

TREETIP: tcVISION Supports Data Replication to MongoDB

tcVISONv6

The tcVISION cross-system integration platform is a robust, proven, and mature solution that is constantly under development to meet the requirements of new technologies, including support for MongoDB.

mongodb

MongoDB is among the leading NoSQL databases in the market and has been developed for the needs of today’s information technology. MongoDB supports a data model with dynamic schemata and is especially suitable to store large amounts of data using GridFS. It contains automatic failure protection using an integrated replication. MongoDB also offers native, idiomatic drivers for nearly all programming languages and frameworks.

Find out more about MongoDB here: https://www.mongodb.com

In addition to support for MongoDB, tcVISION features connectivity to other output targets, such as Hadoop (see previous blog about Hadoop support), Adabas LUW, DB2 BLU, and EXASOL. Additionally, new input sources include z/OS VSAM Logstream (CICS and Coupling Facility / Shared VSAM), z/OS VSAM Batch Extension, z/OS DBMS to Logstream, CA IDMS v17, CA Datacom CDC, IMS Active Log, and SMF data.

Find out more about tcVISION — Enterprise ETL and Real-Time Data Replication Through Change Data Capture

tcVISION provides easy and fast data migration for mainframe application modernization projects and enables bi-directional data replication between mainframe, Linux, Unix and Windows platforms.

_0_tcVISION_Simple_Diagram

tcVISION acquires data in bulk or via change data capture methods, including in real time, from virtually any IBM mainframe data source (Software AG Adabas, IBM DB2, IBM VSAM, IBM IMS/DB, CA IDMS, CA Datacom, even sequential files), and transform and deliver to virtually any target. In addition, the same product can extract and replicate data from a variety of non-mainframe sources, including Adabas LUW, Oracle Database, Microsoft SQL Server, IBM DB2 LUW and DB2 BLU, IBM Informix and PostgreSQL.


__TSI_LOGO

Visit the Treehouse Software website for more information on tcVISION, or contact us to discuss your needs.

Treehouse Software partner, Cognistx to present at The Big Data Summit 2017 in Germany

florian_metze_bigdatasummit

Dr. Florian Metze, Partner and Chief Speech Scientist at Cognistx, will be presenting “Speech and Deep QA (Question Answering) Developments” as part of a special Cognitive Computing track at the Big Data Summit 2017 to be held in Germany, February 16, 2017.

Dr. Metze will be showcasing how Cognitive Computing, Deep QA and context sensitive end-to-end ASR can be utilized to provide extremely intelligent solutions with an extremely simple user interface. The user can ask questions about e.g. their wireless devices and bills, power tools and home appliances, etc. The system proactively engages the user via a Question Generation (QG) module and then provides the right solution or content based on the user queries. The strength of the solution lies in an intuitive voice user interface, an interactive QG module, and its ability to synthesize large quantities of structured and un-structured data to provide the right answer based on the user query.

He will also discuss recent trends in US academia and industry that have been enabled by the rapid development of Artificial Intelligence using large amounts of data and novel machine learning algorithms.

Dr. Florian Metze is an Associate Research Professor at Carnegie Mellon University. He holds a PhD from Universität Karlsruhe (TH), and has previously worked at Deutsche Telekom Laboratories. Dr. Florian is currently a Partner and Chief Speech Scientist at Cognistx.

For more information on this event, visit the BITKOM Big Data Summit website.


The Treehouse Software / Cognistx Connection

_0_treehouse_to_cognistx002

Additionally, representatives from Treehouse will be in attendance with Cognistx at the Big Data Summit. Rolf Vollmer, Treehouse Software European Sales Manager and Hans-Peter Will, Treehouse Software Senior Technical Representative will be available to meet with those interested in learning about how the Cognistx platform complements Treehouse Software’s tcVISION. tcVISION is the comprehensive product that can acquire data in bulk or via change data capture methods, including in real time, from virtually any IBM mainframe data source (Software AG Adabas, IBM DB2, IBM VSAM, IBM IMS/DB, CA IDMS, CA Datacom, even sequential files), and transform and deliver to virtually any target. In addition, the same product can extract and replicate data from a variety of non-mainframe sources, including Adabas LUW, Oracle Database, Microsoft SQL Server, IBM DB2 LUW and DB2 BLU, IBM Informix and PostgreSQL.


Treehouse Software Partners with Stone Bond Technologies to offer comprehensive enterprise Data Virtualization

Press Contacts: Joe Brady (+1.724.759.7070 x110; jbrady@treehouse.com); Irma Barrientos (+1 713.300.8882; ibarrientos@stonebond.com)

Pittsburgh, PA; Houston, TX, November 17, 2016. ­ Treehouse Software, Inc., of Sewickley, PA and Stone Bond Technologies, L.P. of Houston, TX today announced a partnership to bring agile integration and data virtualization capabilities to every enterprise. Now mainframe data sources can be combined into virtual schemata with relational databases, ERP data, Cloud sources and social media by combining Stone Bond’s Enterprise Enabler® with Treehouse’s tcACCESS.

Since the mid-1990s, Treehouse has been a global leader in mainframe data migration, replication and integration, offering robust and flexible solutions for ETL, CDC and real-time, multidirectional replication between databases on various platforms. With Enterprise Enabler, Treehouse customers benefit from a single platform for enterprise and Cloud integration to deliver integration projects with up to 90% time savings over other solutions.

Treehouse will provide worldwide product sales, marketing and support for Enterprise Enabler, bringing leading-edge integration capabilities to its hundreds of existing customers and to new customers pursuing initiatives such as Agile BI, Master Data Management, Logical Data Warehouses and Web Services.

One of the fastest-growing data virtualization platforms on the market, Enterprise Enabler has been leading the innovation of data management, data federation and data virtualization for nearly 15 years. To serve its expanding base of customers, Stone Bond has grown its set of AppComm™ connectors to accommodate the widest range of enterprise and line-of-business data sources. Treehouse’s tcACCESS effectively becomes the “Mainframe AppComm”, providing full read/write access to a wide range of mainframe data sources on both z/OS and z/VSE platforms.

Wayne Lashley, Treehouse Chief Business Development Officer, said, “Our reputation for capably solving mainframe data integration challenges now extends to other modern enterprise data sources including SAP, Salesforce.com and Microsoft SharePoint. With the push of a button, our customers can securely expose these disparate virtualized sources to popular BI tools like Tableau and to other applications in standard formats like REST and JSON. Current, past and new customers can look to Treehouse to address any imaginable integration scenario.”

Commented Tom Sieger, Stone Bond SVP, Corporate Strategy and Business Development, “Embracing mainframe data sources bolsters the “enterprise” in Enterprise Enabler. We are excited to work with Treehouse to grow Enterprise Enabler sales into new markets and geographies.”


__tsi_logo_400x200

About Treehouse Software, Inc.
Privately-held Treehouse Software was founded in 1982, and has hundreds of customers worldwide that benefit from the company’s industry-leading products and outstanding technical support. The traditional strengths of the company in performance management, security and software configuration management have been complemented by visionary leadership in mainframe data replication and integration and application modernization.

stonebond_logo

About Stone Bond Technologies, L.P.
Headquartered in the heart of Houston, Texas, Stone Bond Technologies was founded in 2002 with a commitment to innovation that continues today. More than just adding features to existing offerings, the company’s approach to information business systems is to make an enterprise’s data agile: to move data efficiently and easily, without barriers. Stone Bond’s client base continues to expand from its roots in the Energy industry to Financial Services, Life Sciences, Information Technology, Hospitality, Manufacturing and more.

TREETIP: tcVISION Allows for Surprisingly Innovative Uses

tcVISONv6

Treehouse Professional Services consultants help Adabas / Natural customers in a variety of ways, including DBA services, performance tuning, change management implementation and training, and data replication planning and training. Our experience and long history of service to the Adabas / Natural community helps us create innovative solutions for our customers’ challenges.

Recently Treehouse Senior Technical Representative Chris Rudolph assisted a customer with a tricky data replication problem. The customer uses tcVISION to perform bi-directional replication between Adabas and an RDBMS during the phase-in of a new application. Unfortunately, the new application incorrectly updated certain columns in the RDBMS, which were then replicated to Adabas. The customer attempted to address the issue by running a series of ADASEL reports against the Adabas PLOG and manually checking for “bad” transactions, which was a very time consuming process that pulled the Adabas DBA away from her normal duties.

Chris explained that tcVISION could expedite the process by replicating all transactions for the Adabas file to a journal table capturing the “before” and “after” values of the problematic columns. The developers working on the new application could then identify invalid values, correct the application, and patch the data themselves. This also allowed the Adabas DBA to return to their normal duties.

The journal table now includes columns to display the “before” and “after” values of the corrupted column, Adabas transaction time, end transaction time, operation and Adabas userid. The customer’s developers immediately recognized immense value from being able to query the journal table to find bad data, patch the data, prove that corruption is no longer taking place, and verify that all corrupted instances of the data have been patched. Journal tables have been added for all replicated Adabas files, and the developers now rely on the journal tables for all of their data patches.


Find out more about tcVISION — Enterprise ETL and Real-Time Data Replication Through Change Data Capture

tcVISION provides easy and fast data migration for mainframe application modernization projects and enables bi-directional data replication between mainframe, Linux, Unix and Windows platforms.

_0_tcVISION_Simple_Diagram

tcVISION acquires data in bulk or via change data capture methods, including in real time, from virtually any IBM mainframe data source (Software AG Adabas, IBM DB2, IBM VSAM, IBM IMS/DB, CA IDMS, CA Datacom, even sequential files), and transform and deliver to virtually any target. In addition, the same product can extract and replicate data from a variety of non-mainframe sources, including Adabas LUW, Oracle Database, Microsoft SQL Server, IBM DB2 LUW and DB2 BLU, IBM Informix and PostgreSQL.


__TSI_LOGO

Visit the Treehouse Software website for more information on tcVISION, or contact us to discuss your needs.