Treehouse Software, Inc. helps US Foods® with their Mainframe Modernization Initiative

by Joseph Brady, Director of Business Development / Cloud Alliance Leader at Treehouse Software, Inc.

 

Treehouse_USFoods_Logos

Treehouse Software, Inc. is pleased to announce that we were chosen by US Foods for their mainframe data modernization initiatives using the tcVISION Mainframe-to-Cloud and Open Systems data replication product.  

Treehouse Software is the worldwide distributor of tcVISION, the leading tool for using change data capture (CDC) for synchronizing mainframe data with real-time and bi-directional data replication. tcVISION’s intuitive data modeling and mapping, and ease of migrating data, made it the ideal choice for helping to modernize the large mainframe environment at US Foods.  


“The entire Treehouse Software team is excited about working with US Foods to make their modernization initiatives a success!” – George Szakach, CEO and President at Treehouse Software


About US Foods

With a promise to help its customers Make It, US Foods is one of America’s great food companies and a leading foodservice distributor, partnering with approximately 250,000 restaurants and foodservice operators to help their businesses succeed. With 70 broadline locations and more than 80 cash and carry stores, US Foods and its 28,000 associates provides its customers with a broad and innovative food offering and a comprehensive suite of e-commerce, technology and business solutions. US Foods is headquartered in Rosemont, Ill. Visit https://www.usfoods.com/  to learn more.


__TSI_LOGO

Interested in seeing a live, online demo of tcVISION?

tcVISION_Overall_Diagram_Cloud_OS

Simply fill out our tcVISION Demonstration Request Form and a Treehouse representative will be contacting you to set up a time for your requested demonstration.

Treehouse Software – 40 Years and Still Moving Forward (Part 3)

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc. 

__TSI_LOGO_40th_Transp

Introduction

Many readers know that Treehouse Software has been around since 1983, serving enterprises worldwide with industry-leading software products and outstanding technical support. This blog series has discussed Treehouse Software’s origins and the growth of the software company from the early 1980s up to the present.

Change is in the Air, and in the Clouds…

Cloud_Data_Replication

Parts 1 and 2 of this series illustrated the solid beginnings of Treehouse Software in the 80’s and 90’s.  Several products were developed and introduced.  Marketing representatives were acquired in several countries around the globe.  Also, other companies with valuable products sought out Treehouse Software to sell and support their offerings.

In the late 90’s and the early 2000s, we began to experience certain customers’ needs to have Adabas data moved (migrated/copied/converted/distributed) to other database systems.  We developed the tRelational/DPS product set to analyze their Adabas data and structure, and move this data to other relational database systems (RDBMS) such as Oracle, Db2, etc.  This complicated product was our foray into the mainstream of our customers’ mainframe IT processing.  Millions of records of initial data needed to be materialized (ETL) efficiently, and a Change Data Capture (CDC) capability was imperative.  Mission accomplished.

In the early 2000s, some customers required real-time CDC.  This led to the development of DPSync, so named because it kept Adabas data in Sync with the target RDBMS; but this was limited to uni-directional replication.

Requests began coming in for bi-directional (e.g., moving data back to Adabas from Oracle).  Then it was Oracle-to-Db2.  And vice versa – and more variations.  We investigated companies purporting to do data migration from/to various database systems.  In 2007, we found a company with a product already developed and proven, that could in fact move data from/to practically all known database systems at that time.  We partnered with B.O.S. of Germany for Treehouse to do worldwide sales, marketing, support, demos, POCs, training, etc., for that impressive, growing product set, tcVISION.

tcVISION_Overall_Diagram_Cloud_OS

tcVISION caught on quickly with some of our existing customers of Adabas, but the significant interest commenced when the Cloud took hold.  The Cloud was not just a remote data center, or a place to archive large amounts of data, but has capabilities and features that would attract our types of customers with mainframes and terabytes of data. Enterprise cutomers needed tools that allowed them the connectivity to take advantage of Cloud-based technologies, such as highly available and scalable databases; advanced analytics and security; machine learning and artificalial intelligence; data warehouses and stores; and the list goes on.  This shaped the future direction of Treehouse Software.  More on this in the next blog.


__TSI_LOGO_40th_Transp

About Treehouse Software

Since 1983, Treehouse Software has been serving enterprises worldwide with industry-leading mainframe software products and outstanding technical support. Today, Treehouse Software is a global leader in providing data replication, and integration solutions for the most complex and demanding heterogeneous environments, as well as feature-rich, accelerated-ROI offerings for information delivery, and application modernization.

Contact Treehouse Software

Treehouse Software – 40 Years and Still Moving Forward (Part 2)

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc. 

__TSI_LOGO_40th_Transp

Introduction

Many readers know that Treehouse Software has been around since 1983, serving enterprises worldwide with industry-leading software products and outstanding technical support. This blog series will dig a little deeper into Treehouse Software’s origins and explore how founder and president, George Szakach blazed a trail from being a programmer and manager in the 1960s and 1970s, to creating and growing his own software company from the early 1980s up to the present.

Treehouse First Generations… 1980s – 90s

As mentioned in Part 1 of this series, George’s foundational mainframe experience from 1960 to 1975, combined with skills honed from 1975 to 1982 at Software AG, planted the seeds for the next step, building Treehouse Software. George explains:

“Treehouse Software was started during Christmas of 1982, with a consulting assignment I decided to take in the Kingdom of Saudi Arabia doing ADABAS performance analysis and tuning.  In addition, I was asked to teach the customer how to use ADAMINT (which I get credit for developing from July 1975 through May 1979 – or blame, take your pick).  I thoroughly enjoyed my 10 weeks in the Kingdom, where I became a big fan of pistachio nuts. 

Treehouse Software was incorporated in mid-1984 in Sewickley, PA.  The plan then was to provide consulting and educational services to SAG sites, while developing software products related to the SAG enterprise.  The services might sell the products, and vice versa.  It worked and still works.  I started hiring help almost immediately as the demand for products and services was growing.

By the way, Sewickley is a small town near Pittsburgh where the oil, steel, coal barons of the 1800s and early 1900s had their summer homes up the hill from the river.  The 100-year-old horse-and-wagon trails and stone walls around the estates can still be seen.”

Sewickley_Bridge

The Sewickley Bridge spans the Ohio River between Sewickley and Moon Township, PA.

Some interesting facts about Treehouse Software, Inc.

By the end of the 20th century, Treehouse designed, developed, and released ten products, most notably:

Treehouse also marketed several additional products from companies around the world.  Most were related to ADABAS.

Other miscellaneous Treehouse happenings last century:

  • Taught classes for over 4000 students, ADABAS and Natural related, and related to Treehouse products. 
  • Partnered with 20 affiliates or marketing representatives in various countries, topping out at 14 around 1990.
  • Produced 55 issues of the popular Treetips newsletter, with a hard-copy circulation of 13,000 per issue.
  • Attended 11 Software AG conferences.
  • Presented at 90 regional and local Software AG meetings in various countries.
  • Visited affiliates, customers, and partners in at least 24 countries on all continents except Antarctica.
  • Invited to Oracle Conference in 1995, in Philadelphia, where Emilie Szakach entertained the crowd with an on-stage dance with Chubby Checker.
  • Bought four IBM mainframes for the company.
  • Hired many top notch programmers and analysts, tech writers, support personnel.
  • Along with most of the world, we awaited and prepared for the feared Y2K debacle, which didn’t happen. However, as the year 2000 approached, one of Treehouse’s most popular and most requested give-aways ever, was our Y2K desk clock…

_0_Treehouse_Y2K_Clock

Cha, Cha, Cha, Changes…

Along with many new products, innovations, and memorable, contributing staff moving through the doors of the Treehouse, we’ve also had several iterations of the company logo and colors, as seen in the following graphic. Do you have a favorite?

40Yearstsi_logos_all

Coming soon… Part 3: Change is in the Air, and in the Clouds…    


__TSI_LOGO_40th_Transp

About Treehouse Software

Since 1983, Treehouse Software has been serving enterprises worldwide with industry-leading mainframe software products and outstanding technical support. Today, Treehouse Software is a global leader in providing data replication, and integration solutions for the most complex and demanding heterogeneous environments, as well as feature-rich, accelerated-ROI offerings for information delivery, and application modernization.

Contact Treehouse Software

Treehouse Software – 40 Years and Still Moving Forward (Part 1)

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc. 

__TSI_LOGO_40th_Transp

Introduction

Many readers know that Treehouse Software has been around since 1983, serving enterprises worldwide with industry-leading software products and outstanding technical support. However, this blog series will dig a little deeper into Treehouse Software’s origins and explore how founder and president, George Szakach blazed a trail from being a systems programmer in the early 60s, to creating and growing his own software company from the early 80s up to the present.

The beginnings… 1960’s.  Moon Landing, Flower Power, the Righteous Brothers, and Punched Cards

George is a Vietnam-era veteran and started working with mainframes in 1960 while in the Army. 

After programming school in Fort Monmouth, NJ, George was assigned to Fort Huachuca, Arizona where he wrote army related applications on the IBM 709.

Before leaving the army in 1963, George had many job offers.  Three years of programming experience was unheard-of back then, so his skillset was very valuable.  He was even offered a job by the president of Informatics, working in Houston at the NASA Johnson Space Center to “put a man on the moon.”  He declined.

Throughout the rest of the 60s, George worked at Burroughs, Univac, and Leasco. During the 70s through 1982, George worked for Ocean Data Systems, Data General, Optical Recognition Systems, Software AG (his longest stint – 7 years), and Superior Oil.

Punched card…

01_2023_Treehouse_Chronology_Blog_Punched_Card

IBM 709 Computer System…

IBM709

George’s archeological finds from his time at Univac…

____01_Chron_UNIVAC_Card

With all of this foundational mainframe experience combined with his skills honed at Software AG, the seeds were planted for George’s future: take those roots, move to the trees, and build a house… 

Coming soon… Part 2: Treehouse Software’s first generations.    


__TSI_LOGO_40th_Transp

About Treehouse Software

Since 1983, Treehouse Software has been serving enterprises worldwide with industry-leading mainframe software products and outstanding technical support. Today, Treehouse Software is a global leader in providing data replication, and integration solutions for the most complex and demanding heterogeneous environments, as well as feature-rich, accelerated-ROI offerings for information delivery, and application modernization.

Contact Treehouse Software

AWS Services Provide Advanced Monitoring and Analytics of tcVISION’s Mainframe CDC Processing

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

____AI_Data_Monitoring_And_Analytics

Many Treehouse Software mainframe modernization customers have requirements for continuous near-real-time replication of mainframe data in order to keep a copy of the data synchronized on the Cloud. These customers are using tcVISION from Treehouse Software for changed data capture (CDC) for this synchronization, which allows changes occurring in any mainframe application data to be tracked and captured, and then published to a variety of AWS targets, including Amazon Simple Storage Service (S3). Some of these customers are also now asking us to recommend the best Cloud-based tools and methods to monitor and gain insights to these complex data processes. Coincidentally, while working with a current tcVISION customer, our technicians are testing out two particularly good, fully managed AWS services that can work hand-in-hand to address this need:

Amazon Athena

Since tcVISION supports Amazon S3 as a target, customers modernizing their mainframe systems on AWS can use Amazon Athena for monitoring and analysis of CDC processing from an S3 bucket.

Amazon Athena is a serverless, interactive analytics service built on open-source frameworks, supporting open-table and file formats. Athena provides a simplified, flexible way to analyze data from an S3 Bucket, as well as many other data sources, including on-premises data sources or other Cloud systems. Athena is built on open-source Trino and Presto engines and Apache Spark frameworks, with no provisioning or configuration effort required.

Figure 1: Example of an Athena query showing bulk-load statistics per table

____01_Amazon_Athena_Query

Amazon QuickSight

____01_Amazon_QuickSight

Once Athena is setup for monitoring an S3 Bucket, users can easily view their CDC processing and analytics with Amazon QuickSight. QuickSight utilizes advanced machine learning-powered insights and intuitive dashboards, so end users can make the best and quickest data-driven business decisions.

Figure 2: Example of Amazon QuickSight monitoring the throughput of our data to Snowflake

____01_Amazon_QuickSight02

Figure 3: Example of Amazon QuickSight pie chart showing the resulting rows loaded for each Snowflake table:

____01_Amazon_QuickSight03

Figure 4: Example of Amazon QuickSight chart showing statistics for our data bulk-load into Snowflake:

____01_Amazon_QuickSight04

Figure 5: Example of Amazon QuickSight chart showing our load time into Snowflake per table:

____01_Amazon_QuickSight05

View the Amazon QuickSite video here…


__001_TSI_LOGO

Interested in seeing a live, online demo of tcVISION?

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.


Treehouse Software Salutes Franco Harris

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc. 

With the recent passing of Pittsburgh Steelers great running back and Pro Football Hall of Famer, Franco Harris, we would like to revisit April of 1993, when Treehouse Software held an international consultant’s symposium. The symposium brought together attendees and speakers from many consulting and technology companies, and schools from around the world. Since Treehouse Software is located in the greater Pittsburgh area, company president George Szakach was acquainted with Franco and invited him to deliver a fascinating and entertaining address, where he spoke about his career, several business ventures he was pursuing, as well as his budding interest in computer technology.

____01_Chron_Franco

Franco Harris at Treehouse Software’s Consultant’s Symposium (April 1993)

A few years ago, George reminded Franco about his visit to Treehouse back in 1993. He remembered and they shared some laughs and memories. 

We would also like to mention Franco’s well-known sense of community and accessibility in Pittsburgh. Many staff members have met Franco over the years and have fond memories of his friendliness and willingness to spend time engaging in conversations. Those who come in to the Pittsburgh International airpot can see a sculpture depicting Franco’s famous “Immaculate Reception” from 1972.  Thousands of people, especially recently, have selfies taken with the sculpture. Franco will be missed by his many friends and the community.

____01_Chron_Franco_Airport_Sculpt02

Franco Harris sculpture at Pittsburgh International Airport.


__001_TSI_LOGO

About Treehouse Software

Since 1982, Treehouse Software has been serving enterprises worldwide with industry-leading mainframe software products and outstanding technical support. Today, Treehouse Software is a global leader in providing data replication, and integration solutions for the most complex and demanding heterogeneous environments, as well as feature-rich, accelerated-ROI offerings for information delivery, and application modernization.

Contact Treehouse Software

Considerations for Planning Bi-Directional Mainframe Data Replication with tcVISION

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

Data_Modrnization

Many medium-to-large size enterprises use mainframe systems that are housing vast amounts of mission-critical data encompassing historical, customer, logistics, etc. information.  Each mainframe site is unique and can have decades worth of customizations requiring innovative approaches to establishing data replication on Cloud and open systems platforms. Fortunately for these customers, Treehouse Software has been in the mainframe software market since 1982, bringing deep experience in mainframe, Cloud, and open systems technologies, as well as delivering the tcVISION mainframe data replication product. Today, Treehouse Software is helping many enterprise mainframe customers accelerate digital transformation and successfully leverage Hybrid Cloud initiatives on the IBM Z platform, storing sensitive data on a private Cloud or local data center and simultaneously leveraging leading technologies on a managed public Cloud.

Treehouse Software’s tcVISION solution focuses on changed data capture (CDC) when transferring information between mainframe data sources and Cloud and open systems-based databases and applications. Changes occurring in the mainframe application data are then tracked and captured, and published to a variety of targets. Additionally, tcVISION supports bi-directional data replication, where changes on either platform are reflected on the other platform (e.g., a change to a PostgreSQL table in the Cloud is reflected back on mainframe), allowing the customer to modernize their application on the Cloud or open systems without disrupting the existing critical work on the legacy system. tcVISION’s bi-directional replication writes directly to the mainframe database, thereby bypassing all mainframe business logic, so this architecture requires careful planning, as well as thorough and repeated testing.

Plan carefully…

The following section offers some real-world customer examples, as well as considerations and recommendations when planning bi-directional replication for any mainframe/RDBMS environments. Bi-directional replication by its nature is a very complicated undertaking, so it is necessary that customers are fully educated in all environments, software, and processes before attempting to write data back to a mainframe database. It is always recommended that customers use a minimally effective measure of bi-directional replication required to accomplish their goal — and no more. An overblown project with unnecessary bi-directional data replication invites undue complexity and delays.

Real-world customer examples…

Treehouse Software has many customers performing bi-directional data replication, and each scenario is vastly different from the others, even if some have the same sources and targets as each other.  For example, some customers utilize a Master/Master, collision-heavy proposition, while others use uni-directional one way, then “flip a switch” uni-directional the other way. Another example is a customer who has a “grand circle,” where data hits multiple applications before it finally makes its way back to an RDBMS staging database that tcVISION replicates to the mainframe.

Example of a Treehouse customer’s bi-directional data replication environment using tcVISION:

tcVISION_Adabas_To_AWS_RDS

There are many planning and implementation stages that go into a successful mainframe replication environment, and performance testing is a vital part of a successful project.  For example, customers should do performance tests on how long it takes tcVISION to read a database log, transfer data, process data, etc.  During testing at one of our reference customer sites we found a significant difference in how long it took for their test and prod LPARs to transmit data to the Cloud, based on whether the mainframe TCP/IP stack used a 32-bit or 128-bit setting.

At another site, where we are helping a large government agency perform bi-directional replication on mainframe data, their original goal was for a significant percentage of mainframe objects to have bi-directional replication. It was determined that it would be impossible to extract business logic from the existing mainframe application for usage in the downstream application. Therefore, they have decided to use a middleware product to perform the “write-back” to the mainframe database.  Given the complexity of the mainframe application, this has proven the safest way for them to proceed.

Because of the variety of customer scenarios as described above, before any site can attempt bi-directional data replication, it is crucial that they have a well-tested uni-directional process with operational controls in place for a significant time period.  “Operational controls” means processes to restart scripts, evaluation of failed transactions, orchestration of mainframe/non-mainframe DBMS changes, etc.

Please contact Treehouse Software to discuss your Mainframe-to-Cloud and Open Systems modernization plans. We can help put in place a roadmap to modernization success.


__TSI_LOGO

Contact Treehouse Software Today for a tcVISION Demo…

No matter where you want your mainframe data to go – the Cloud, open systems, or any LUW target – tcVISION from Treehouse Software is your answer.

_0_Treehouse_tcV_Cloud_OpenSystems

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.


Providing a High Availability Framework for Mainframe-to-AWS Data Replication

by Dan Vimont, Cloud Solutions Architect at Treehouse Software, Inc.

tcV_HA_on_AWS

Treehouse Software customers are using tcVISION to enable mission-critical mainframe-to-AWS data replication pipelines.  Some of these production pipelines are providing vital near-real-time synchronization between source and target, and thus can’t afford any significant downtime in the event of failure.  So it’s only natural that a number of our customers have been asking for advice in setting up a high availability configuration for their tcVISION components that run on AWS EC2 instances.  The High Availability Framework discussed here provides for a Failover EC2 instance to automatically pick up tcVISION processing should the Primary instance (running in another Availability Zone) go down.

The Core Components:  Primary Instance & Failover Instance

The core components of a tcVISION high availability framework consist of two EC2 instances running in different Availability Zones:  a Primary EC2 instance and a Failover EC2 instance.  Both identically-configured EC2 instances are attached to a shared working-storage file system (either an EFS or FSx volume), which allows the Failover instance to seamlessly and quickly pick up tcVISION processing should the Primary instance suddenly become unavailable.

HA1

Use a Step Function to Automate the Failover Process

In the event of failure of the Primary instance, the recommended framework calls for automatic triggering of a Step Function for reliable failover processing, with steps that include the following:

  • verify that the Primary instance is unavailable (The tcVISION service cannot be active on both instances simultaneously, so this verification is vital.)
  • redirect all network traffic from the Primary instance to the Failover instance (via Route 53)
  • start tcVISION processing on the Failover instance

HA2

When Ready, Use a Step Function to Automate the Restoration Process

After operations personnel have completed recovery of the Primary EC2 instance, another Step Function may be manually triggered to reliably transfer tcVISION processing back to the Primary instance.

HA3.jp

Many More Details are Available Upon Request to Treehouse Customers

Full details regarding our recommended High Availability Framework for tcVISION are available upon request to Treehouse customers.  AWS services utilized in the complete recommended framework include Step Functions, Lambda Functions, EventBridge rules, CloudWatch alarms, SNS topics, a Route 53 Private Hosted Zone, and more.  The following diagram is a partial visual inventory of the recommended framework components.

HA5

Interested in seeing a live, online demo of tcVISION?

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.


__001_TSI_LOGO

How to Synchronize Data in Real Time Between the Mainframe and AWS with Treehouse Software’s Enterprise CDC Tool

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

Bidirectional_Data_Replication

Many mainframe integration scenarios require continuous near-real-time replication of relational data to keep a copy of the data synched in the Cloud. Change Data Capture (CDC) is used for this near-real-time transactional replication by capturing change log activity to drive changes in the target dataset.

Just what is CDC anyway?

Simply put, and in relation to Mainframe-to-Cloud and open systems data replication, CDC is the use of processes to identify when data has been changed in a source system, so the replicated upstream or downstream (depending on how you look at it) target can be kept in sync with the changes.

In a recent AWS Architecture Blog, readers learn about integration using mainframe data to build Cloud native services with AWS, including transactional replication-based integration via CDC.

____AWS_Mainframe_CDC_Diagram

As mentioned in the blog, AWS Partner CDC Tools are available for connecting data center mainframes to the various data targets, and Treehouse Software’s tcVISION is one of those tools available in the AWS Marketplace.

tcVISION allows changes occurring in any mainframe application data to be tracked and captured, and then published to a variety of target AWS databases and applications. tcVISION provides an easy and fast approach for Hybrid Cloud projects, enabling real-time and bi-directional data replication between the hardware and AWS.

Example of Db2-to-AWS CDC using tcVISION Mainframe Manager:

tcVISION_Db2_To_AWS_CDC

tcVISION supports several CDC methods available, depending on each customer’s use case:

Bulk Transfer

  • Efficient transfer of entire databases
  • Analysis for data consistency (verification)
  • Initial load (ETL) and periodic mass data transfer
  • One-step data transfer

Log Processing

  • Transfer of changed data near-realtime or scheduled time frame
  • Reads both active logs and archived logs

Batch Compare

  • Comparison of data snapshots using checksums
  • Efficient transfer of changed data since last processing
  • Flexible processing options (SORT etc.)
  • Automatic creation of deltas by tcVISION

DBMS Extension

  • Real-time capture of changed data directly from the DBMS
  • Secure data storage even across DBMS restart
  • Flexible propagation methods

Interested in seeing a live, online demo of tcVISION CDC?

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.


__001_TSI_LOGO

Treehouse Software Customer Case Study: A State Government Agency’s Real-time Data Synchronization Between IBM Mainframe Adabas and AWS

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

Mainframe_to_AWS_Graphic

Software AG’s Adabas is a mainframe database that is still heavily used by government sites throughout the U.S. and the world, and this blog focuses on a current Treehouse Software customer – a U.S. State Government Agency that uses Adabas on their mainframe system.

Business Issue

The Agency’s modernization team was looking for a Change Data Capture (CDC) technology solution that enables them to synchronize their mainframe Adabas data on AWS, particularly an Amazon RDS. As with most Treehouse customers, the State’s mainframe contains vital data that must always be highly available, so rather than attempting a complete migration from the mainframe, the modernization teams decided to implement a multi-year data replication plan. This allows the mainframe legacy teams to maintain existing critical applications, while the modernization team develops new applications on AWS.

After researching various technologies, the Agency discovered tcVISION on the AWS Parter Network Blog and contacted Treehouse Software to discuss their project and to see a demonstration of Mainframe-to-AWS data replication.

Addressing the Uniqueness of Adabas

Having specialized in tools and services complementary to Adabas/Natural applications since 1982, Treehouse Software has successfully encountered and addressed many unique scenarios within the Adabas environment. The Treehouse technical team documented three primary issues with Adabas/Natural that the Agency needed to consider when they began planning data replication on AWS:

  1. Adabas has no concept of “transaction isolation”, in that a program may read a record that another program has updated, in its updated state, even though the update has not been committed.  This means that programmatically reading a live Adabas database—one that is available to update users—will almost inevitably lead to erroneous extraction of data.  Record modifications (updates, inserts and deletes) that are extracted, and subsequently backed out, will be represented incorrectly—or not at all—in the target. Because of this, at Treehouse we say “the only safe data source is a static data source”—not the live database.
  2. Many legacy Adabas applications make use of “record typing”, i.e., multiple logical tables stored in a single Adabas file.  Often, each must be extracted to a separate table in the target RDBMS.  The classic example is that of the “code-lookup file”.  Most shops have a single file containing state codes, employee codes, product-type codes, etc.  Records belonging to a given “code table” may be distinguished by the presence of a value in a particular index (descriptor or superdescriptor in ADABAS parlance), or by a range of specific values.  Thus, the extraction process must be able to dynamically assign data content from a given record to different target tables depending on the data content itself.
  3. Adabas is most often used in conjunction with Software AG’s Natural 4GL, and “conveniently” provides for unique datatypes (“D” and “T”) that appear to be merely packed-decimal integers on the surface, but that represent date or date-time values when interpreted using Software AG’s proprietary Natural-oriented algorithm. The most appropriate way to migrate such datatypes is to recognize them and map them to the corresponding native RDBMS datatype (e.g., Oracle DATE) in conjunction with a transformation that decodes the Natural value and formats it to match the target datatype.

The tcVISION Technology Solution...

Adabas_To_AWS

After technical discussions and a successful proof of concept (POC) that proved out a set of use cases, all teams at the Agency determined that tcVISION real-time mainframe data replication capabilities were the perfect fit for meeting their goals.

tcVISION‘s modeling and mapping facilities are utilized to view and capture logical Adabas structures, as documented in Software AG’s PREDICT data dictionary, as well as physical structures as described in Adabas Field Definition Tables (FDTs).  Given that PREDICT is a “passive” data dictionary (there is no requirement that the logical and physical representations agree), it was necessary to scrutinize both to ensure that the source structures were accurately modeled.

Furthermore, tcVISION generates appropriate mappings and transformations for converting Adabas datatypes and structures to corresponding target datatypes and structures, including automatic handling of the proprietary “D” and “T” source datatypes.

The teams examined the three ways that tcVISION can access Adabas data:

  1. ETL – read the active database nucleus
  2. ETL – read datasets containing unloaded Adabas files created by the ADAULD utility
  3. CDC – read the active and archived PLOGs datasets

It was decided to access the data by reading the active and archived PLOGs datasets. The schema, mappings, and transformations from the metadata import were tailored to the customer’s specific requirements.  It is also now possible to import an existing RDBMS schema and retrofit it, via drag-and-drop in tcVISION, to the source Adabas elements.

Additionally, the Agency’s teams are very pleased with tcVISION‘s minimal usage of mainframe resources. The product’s “staged processing” methodology accomplishes this, whereby the only processing occurring on the mainframe is the capture of changes from Adabas PLOGs. The bulk of the processing occurs on the AWS side, minimizing tcVISION’s footprint on the mainframe as seen in this diagram:

tcVISION_Staged_Processing

The user defines on which platform stage their processing should be done. Do as little as possible on the mainframe: Stage 0 – capture data and send data (internal format) to target, and process data in Stages 1 – 3 in AWS.

Customer Outcome

All requirements were met by tcVISION, which led to a successful project implementation.


__001_TSI_LOGO
Contact Treehouse Software for a tcVISION Demo Today…

No matter where you want your mainframe data to go – the Cloud, open systems, or any LUW target – tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.


Further reading:

Many more mainframe data migration and replication customer case studies can be read on the Treehouse Software Website.