Replicating Enterprise Mainframe Data to Cloud-based SQL Databases with tcVISION

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

Treehouse Software has been helping enterprise mainframe customers since 1982, and in recent years, we have been developing a strong presence in the Mainframe-to-Cloud data replication market space. This blog takes a quick look at three of the most popular Treehouse-supported Cloud-bases SQL database services…

Amazon RDS, a collection of managed services that makes it simple to set up, operate, and scale databases in the Cloud. Users can control the type of database, as well as where data is stored. Specific database formats that are supported include Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server:

Google Cloud SQL, a fully managed relational database service for MySQL, PostgreSQL, and SQL server. You can connect with nearly any application, anywhere in the world. Cloud SQL automates backups, replication, and failover to ensure your database is reliable, highly available, and flexible to your performance needs:

Microsoft Azure SQL, a part of the Azure SQL family, Azure SQL Database is an always-up-to-date, fully managed relational database service built for the Cloud:


Wherever you want to target your mainframe data on the Cloud, Treehouse Software helps to make the process easy…

Treehouse Software is the worldwide distributor of tcVISION, the leading tool for using changed data capture (CDC) when transferring information between most mainframe data sources (IBM Db2, IBM VSAM, IBM IMS/DB, Software AG Adabas, CA IDMS, CA Datacom, or even sequential files) and Cloud and open systems-based databases and applications. Changes occurring in the mainframe application data are then tracked and captured, and published to a variety of targets.

tcVISION_Overall_Diagram_Cloud_OS

Additionally, tcVISION supports bi-directional data replication, where changes on either platform are reflected on the other platform (e.g., a change to a PostgreSQL table in the Cloud is reflected back on mainframe), allowing the customer to modernize their application on the Cloud or open systems without disrupting the existing critical work on the legacy system. tcVISION’s bi-directional replication writes directly to the mainframe database, thereby bypassing all mainframe business logic, so this architecture requires careful planning, as well as thorough and repeated testing.

Sales and technical leaders at the major Cloud platform companies, as well as systems integrators are engaging with Treehouse Software to take advantage of our tcVISION data replication solution to help them tap into the mainframe data that customers want to be made available on new technologies.


Further reading: tcVISION is featured on the AWS Partner Network Blog showing a walk-through of data replication between Mainframe DB2 z/OS and Amazon Aurora…

AWS Partner Network (APN) Blog: Real-Time Mainframe Data Replication to AWS with tcVISION from Treehouse Software.


__tsi_logo_400x200

Interested in seeing a live, online demo of tcVISION?

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

Treehouse Software Customer Success: BMF uses tcVISION for Real-Time Data Replication Between Mainframe Adabas and PostgreSQL

BMF_Building

The Bundesministerium der Finanzen (BMF) is Germany’s Ministry of Finance and establishes sustainable fiscal policy that ensures financial empowerment of the federal budget. From tax policy via development of federal budget, to regulation of national and international financial markets – for these and other fiscal and economic questions of principle, the BMF creates strategies and concepts, and implements them. The Federal Tax Administration is part of BMF, and controls not only the cross-border goods traffic, but acts against illegal employment and other crimes. The tax administration also imposes consumer taxes (e.g., energy and tobacco tax, car tax, etc.). Financial relations between federation, countries, and communities are also coordinated by BMF.

Department II (federal budget) is part of the German government in charge of establishing the budget and financial planning of the federation. Throughout the year, it monitors execution of the budget for eventual intervention (e.g., with a budget freeze, or supplementary budget). After closing the fiscal year, the budget and balance sheet will be presented. The budget is a supplement of the budget act, legally binding.

The central service organization of BMF is the Informationstechnikzentrum Bund – ITZBund (Information technic center).

BUSINESS BACKGROUND

Drawing up the budget is a yearly, highly time consuming, and formalized business process. All departments are involved in nearly every sub-process, and budgeting and financial planning is supported by the application, “Haushaltsaufstellung / Budgetgeneration”. Using the generated reports, various addressees/receivers are supported (e.g., German Federal Government, German Federal Parliament, Federal Council of Germany, finance department in BMF, the employees in the departments, and the public).

Technically, the budget plan of the federation is based on technologies, including the IBM Mainframe with z/OS running Adabas and Natural.

The challenge was to provide an environment for employees in all departments that enables them to do their work quickly, easily, and efficiently. In the BMF, users must have an editorless, end-user driven, and real-time creation of ready-to-print products. An informative description of the workflow is shown on the website of the BMF.

The federal budget is available as download, or one can directly navigate through the data using the online application.

BUSINESS ISSUE

Some time ago, BMF decided to re-engineer the application for budget planning and port it to Open Source. To guarantee a seamless transition, the first step is propagation of data out of Adabas on z/OS to PostgreSQL, concluding with permanent synchronization.

The difficulties of this task are the complexities of setting up data definitions for the data structures in Natural and the propagation of data from Adabas on z/OS to PostgreSQL.

TECHNOLOGY SOLUTION: tcVISION

____Adabas_to_PostgreSQL_Diagram

After an analysis of the project, Treehouse Software proposed creating an extension to tcVISION’s change data capture (CDC) functionality for integration, so that tcVISION could enable BMF to continue using the implemented data definitions in a format suitable for the RDBMS.

The extension was developed within a few days, and a two-day on premise test demonstrated the solution fit the requirements of BMF.

BMF can now provide its data definitions from Natural LDA to the extension of tcVISION, and after the transformation, onto the PostgreSQL load process for processing. Another advantage of the tcVISION solution is that when needed, other targets can be integrated for propagation of data from the mainframe (e.g., Kafka, which BMF indicated is a future target environment).

Additionally, bi-directional propagation can be added in budget planning when BMF is ready.

Data structures are held in LDA, because this provides the advantages of higher flexibility in development and the adaption of new requirements to the data definitions. If definitions would have to be ported manually, in part, to PostgreSQL, it would have been a much bigger and error-prone effort.

Subsequent changes to Adabas structures can now use tcVISION’s newly developed extension to easily regenerate and load the correct definitions to the RDBMS, and tcVISION completely covers the customer’s requirements for special usage of *PEs and *MUs.

After thorough preparation and extensive testing, the solution was released to selected users first, then made available to all users.

* PEs and MUs are special Adabas formats for definition of tables. PE = Periodic Group, MU = Multiple Value Field.


__tsi_logo_400x200

Contact Treehouse Software for a Demo Today…

No matter where you want your mainframe data to go – the cloud, open systems, or any LUW target – tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software Product Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.


Further reading: Treehouse Software Customer Success – ETS: tcVISION for Real-Time Synchronization Between Mainframe IDMS and AWS RDS for PostgreSQL

Enterprise Mainframe Change Data Capture (CDC) to Apache Kafka with tcVISION and Confluent

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc. and Ram Dhakne, Solutions Engineer at Confluent

___Mainframe_To_Kafka_Confluent

This blog focuses on using Treehouse Software’s tcVISION to replicate data in real time between mainframes and Confluent, allowing for new use cases and truly setting data in motion.

Why mainframe modernization? Benefits and use cases

Mainframe data stores often hold large amounts of complex and critical data in proprietary legacy formats, making this data difficult to extract and incompatible with modern databases, data types, and data tools.

Enterprises are looking to take advantage of the latest cloud services, such as analytics, artificial intelligence (AI) and machine learning, scalable storage, security, high availability, etc., or move data to a variety of newer databases. Additionally, many customers want to modernize their application on a cloud or open systems platform without disrupting the existing critical work on the legacy system.

How tcVISION syncs legacy data for the cloud

tcVISION is a data replication software product that performs real-time synchronization of mainframe data sources and cloud and open systems, allowing critical mainframe data to be consumed by a variety of leading cloud services.

tcVISION supports many mainframe data sources for both online and offline scenarios. Data can be replicated from IBM Db2 z/OS, Db2 z/VSE, VSAM, IMS/DB, CA IDMS, CA Datacom, or Software AG ADABAS. tcVISION can replicate data to many targets including Confluent Platform, Apache Kafka®, AWS, Google Cloud, Microsoft Azure, PostgreSQL, Snowflake, etc. To learn more, see the complete list of supported tcVISION sources and targets.

tcvision-mainframe-to-confluent-cloud-data-replication-1536x1042

tcVISION focuses on CDC (change data capture) when transferring information between mainframe data sources and cloud and open systems databases and applications. Through innovative technology, changes occurring in any mainframe application data are tracked and captured, and then published to a variety of cloud and open systems targets.

tcVISION stores metadata in a relational database and the tcVISION manager components are administered by the tcVISION control board, a Windows GUI interface, which can be installed on premises or in the cloud. This allows tcVISION users to create metadata, create and control replication scripts, and control database interactions. tcVISION’s architecture is designed to minimize mainframe resource utilization.

Using the tcVISION control board, the most complex transformations can be specified, and it facilitates the mapping of the mainframe copybooks, redefines, data dictionaries, data catalogs, codepages, data type mapping, and more via the user-friendly interface. The repository editor allows users to control data transformations.

What is Confluent?

Confluent Cloud is a real-time data in motion platform that can be deployed in any public cloud, in any region of your choice. It comes with an SLA and uptime of 99.95%, and fully managed components like ZooKeeper, Kafka brokers, 120+ Kafka connectors, Schema Registry, and ksqlDB so you can leverage it on any cloud without having to worry about how it runs and scales.

Kafka Connect, Connect API, connectors, and tcVISION IBM Db2 connector

Kafka comes with three core APIs:

  • Kafka producer/Consumer API
  • Connect API
  • KStreams API

Kafka Connect is a tool for scalably and reliably streaming data between Kafka and other data systems. It makes it simple to quickly define connectors that move large data sets into and out of Kafka. Kafka Connect can ingest entire databases or collect metrics from all your application servers into Kafka topics, making the data available for stream processing with low latency. Kafka Connect connects APIs under the hood with fully managed connector support in Confluent Cloud.

Step-by-step guide on how to use tcVISION and Confluent

This example discusses the integration of tcVISION replication of data from Db2 to Confluent Cloud.

Set up tcVISION access to Confluent

Create an account with Confluent to make a Confluent user ID/password; the user ID is generally your email address. To sign on to Confluent, go to the Confluent Cloud login and enter your user ID:

Confluent Cloud welcome page

Then, enter your password:

Enter your password

When you log in, you’ll be in a Confluent environment called “default”:

Confluent environment called “default”

A Confluent environment is a type of container that holds clusters which in turn hold topics. If you are familiar with messaging systems, Confluent/Kafka will seem familiar. A cluster will need to be created to serve as a target for the data produced by tcVISION. The first attribute to be selected is the type of cluster. Confluent offers three types: Basic, Standard, and Dedicated. For the purposes of this demonstration, Basic will be used. A Basic cluster does not incur charges for simply existing, but does for data transmission and data storage.

Select "Basic cluster" and begin configuration

Select Begin configuration.

Select a cloud provider

Here, a cloud provider can be chosen—AWS, Google Cloud, or Microsoft Azure. For this example, AWS is used. Select Continue and the characteristics of the new cluster are displayed, which we’ve named “tcVISION_cluster_0”:

Cluster characteristics

After entering your payment information (not shown), you can click on the cluster name to launch the cluster overview.

Cluster overview

In order to use Confluent with tcVISION, the user must provide tcVISION with information about the cluster they intend to use. Specifically, the user must supply the hostname and port of the Confluent AWS virtual machine, and the credentials needed to access the cluster.

Confluent refers to the hostname and port as a bootstrap server. There can be multiple bootstrap servers for the purpose of load balancing, but a single server is used for this demonstration.

To find bootstrap server information, click Cluster Settings on the left-hand side:

Cluster settings

The bootstrap server will be listed under “Identification,” and includes both the AWS hostname and the port.

Credentials in Confluent consist of an API Key and an API Secret. These are generated for the cluster and take the place of the Confluent user ID and password used to log in. To generate a key/secret pair, click API Access on the left:

API Keys page

Followed by Create Key:

Select API Key scope

For this example, we use “Global Access” here, so click Next:

API Key and secret

Pay particular attention to the tip about saving the key and secret somewhere safe, because once this panel is exited, there is no way to display the secret again. A descriptive string for this key/secret pair can be filled in. The key or secret text to be copied can be selected, or use the convenient icons at the end of the field to copy. Once the key/secret has been safely stored, check the box that says it has been done, and click Save. You will return to the “API Keys” panel, and the key is now displayed:

API Key displayed

Set up Confluent and define the topic

The last thing to do is define a topic within the cluster. Confluent producers have the capability to define their own topics within a cluster, but this capability can be disabled by a Confluent configuration and is disabled in the configuration used here.

Go back to the cluster Overview:

Cluster Overview

On the left sidebar, click Topics:

Topics

Then Create Topic:

Create a topic

The topic name is filled in (“CONFLUENT_CLOUD_TOPIC1”), overriding the number of partitions from 6 to 1, since that is what the Confluent demo uses. Click Create with defaults:

Cloud topic

A topic is now available, which can be populated with Db2 data.

Set up tcVISION and run a bulk load of Db2 data

tcVISION’s control board is a Windows graphical user interface (GUI) that allows users to configure the replication stream between various database platforms, including the IBM mainframe and Confluent. Using the control board and built-in wizards, users can define the metadata and the mappings between the mainframe and target.

The following sequence of screens shows the steps required to create the tcVISION metadata and scripts for replicating mainframe Db2 z/OS data to Confluent.

Access the tcVISION control board:

tcVISION control board

Log on to Db2 z/OS:

Db2 z/OS

Create metadata that is specific to the input (Db2) and output (Kafka) and the replication definition. In this example, the Db2 table is mapped to the Confluent Cloud Kafka topic using JSON:

Import of structure definitions

The tcVISION metadata wizard asks for the information required for the replication of the mainframe database to Confluent Cloud. For Db2 z/OS, it asks for the mainframe Db2 subsystem:

Source type for structure definition import

Db2 subsystem

tcVISION presents the tables contained in the Db2 z/OS catalog on the mainframe. Select the schemas and associated tables for replication:

Select the schemas and associated tables for replication

Once the required tcVISION wizard-based screens are completed, the tool automatically defines the mappings between the source and target. tcVISION’s metadata import wizard creates a default mapping that handles data type conversion issues, such as EBCDIC to ASCII, Endianness conversion, codepages, redefines data types, and more:

Default mapping

tcVISION data scripts are created through wizards. Data scripts control the replication of data from the source (Db2 z/OS) to the target (Confluent Cloud Kafka JSON). tcVISION bulk load scripts are a type of data script that performs the initial load of the Kafka topic. The following script shows data being accessed directly from the mainframe Db2 z/OS database. Another alternative to reduce MIPS consumption is to read the data from a Db2 image copy.

Data script

Bulk load script running:

Bulk load script running

After execution of the bulk load script, replication statistics of the Db2 bulk load into the Confluent Cloud Kafka topic can be viewed:

Replication statistics of the Db2 bulk load

Now that the topic has been loaded with data from Db2, it can be displayed in Confluent. To do this, navigate to the topics panel again:

Notice that there are now statistics indicating that the tcVISION producer uploaded some data to the topic. On the horizontal menu, switch from “Overview” to “Messages” to display the messages (data records) that the tcVISION bulk load placed in the topic. The display can be filtered in various ways, but for this example, the default is used: “Jump to Offset,” which says “start displaying sequentially from this offset.” Here, an offset of 0 (start at the beginning) is specified, since we just want to verify that the Db2 data uploaded by tcVISION was actually delivered:

Messages (data records) from tcVISION bulk load

Run a change script in tcVISION to show the changes in Confluent

To capture ongoing changes to Db2 in real time, a Db2 z/OS CDC replication script is created.

This script captures the changes on the Db2 z/OS side and applies them into the repository where the output target is Confluent Cloud topic.

Replication script

Replication script

Target database Confluent Cloud topic

The CDC replication is initiated from the tcVISION control board. The tcVISION control board shows a graphical representation of the replication:

Graphical representation of the replication

The CDC replication is now actively capturing and replicating data changes whenever they occur on the Db2 z/OS side. You can test it by making a change in the Db2 z/OS table:

 
********************************* Top of Data **********************************
---------+---------+---------+---------+---------+---------+---------+---------+
UPDATE SXE1.TVKFKATB                                                    00010004
SET DEPT = '696969'                                                     00040029
WHERE PERS_ID = 5;                                                      00050004
---------+---------+---------+---------+---------+---------+---------+---------+
DSNE615I NUMBER OF ROWS AFFECTED IS 1                                           
DSNE616I STATEMENT EXECUTION WAS SUCCESSFUL, SQLCODE IS 0                       
---------+---------+---------+---------+---------+---------+---------+---------+
--COMMIT;                                                               00060019
---------+---------+---------+---------+---------+---------+---------+---------+
DSNE617I COMMIT PERFORMED, SQLCODE IS 0                                         
DSNE616I STATEMENT EXECUTION WAS SUCCESSFUL, SQLCODE IS 0                       
---------+---------+---------+---------+---------+---------+---------+---------+
DSNE601I SQL STATEMENTS ASSUMED TO BE BETWEEN COLUMNS 1 AND 72                  
DSNE620I NUMBER OF SQL STATEMENTS PROCESSED IS 1                                
DSNE621I NUMBER OF INPUT RECORDS READ IS 4                                      
DSNE622I NUMBER OF OUTPUT RECORDS WRITTEN IS 17                                 
******************************** Bottom of Data ********************************

This change is processed and replicated by tcVISION. The tcVISION control board shows the statistics highlighting that one update was performed:

Display of extended statistics

Checking in Confluent, the Db2 z/OS change has successfully been propagated to the Confluent Cloud topic:

Db2 z/OS change successfully propagated to Confluent Cloud topic

tcVISION and Confluent are better together

With tcVISION’s groundbreaking Db2 CDC connector and Confluent’s ability to serve as the multi-tenant data hub, this combination creates a very powerful solution to aggregate data from multiple sources and have data published into various Kafka topics. Sourcing events from any kind of Db2 via a connector into Confluent will set data in motion for the entire organization. Simplicity and agility are key elements of the tcVISION and Confluent “better together” story.


__001_TSI_LOGO

Video: tcVISION Demonstration…

In this video, we show a tcVISION overview, then a demonstration of replication of mainframe data on AWS RDS for PostgreSQL:

Contact Treehouse Software for a tcVISION Demo Today!

No matter where you want your mainframe data to go – the Cloud, open systems, or any LUW target – tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

Treehouse Software Customer Success: ETS uses tcVISION for Real-Time Synchronization Between their Mainframe IDMS Data and AWS RDS for PostgreSQL

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

ETS_Graphic

This blog focuses on a current Treehouse Software customer – ETS. Headquartered in Princeton, New Jersey, ETS is a private, nonprofit organization with approximately 3,000 employees devoted to educational measurement and research. ETS develops and administer a broad range of educational products and services for government agencies, academic institutions and corporations, including the TOEFL® and TOEIC® tests, the GRE® General and Subject Tests, and the Praxis® assessments. At nonprofit ETS, our belief in the life-changing power of learning is at the root of everything we do — it’s behind the tools we develop to move learning forward, the research that inspires educational progress and the commitment we make to enable opportunity for learners everywhere. We’re with you on the journey to what’s possible.

Business Background

ETS products and services are available to institutions, businesses, organizations and governments in more than 180 countries around the world. The top industries served by ETS are K–12 Education, Higher Education, English-language Learning, Career Development, and Consulting Services.

Business Issue

Most of ETS’s high volume critical application data is stored on an IBM mainframe in IDMS databases.  The technology is very old, therefore it is difficult to recruit and retain qualified technical personnel to maintain applications.  ETS is moving to Cloud-based computing which will allow them to retire the mainframe environments and modernize the applications.  The data is used and shared across several applications.  ETS required a solution that would allow them to continue, uninterrupted, daily operations on their mainframe while replicating data to their AWS Cloud platform, where they could develop modern application features.  This solution enables ETS to maintain demanding daily processing while they modernize and develop innovative Cloud solutions to meet and exceed customer requirements.

The Technology Solution

ETS_Diagram

Treehouse Software and the ETS team developed a rigid testing plan to implement tcVISION and performed a Proof of Concept to measure the effectiveness of the data replication, considering the high volumes of data changes on the source databases.  We collaborated on architecture requirements and installation steps.  There were many considerations associated with this process, including monitoring, alarming, configuration options, high availability, measuring the impact to existing mainframe database performance, restart capability, and security.  Concurrently, a team of subject matter experts worked on data mappings and translation of database designs from the IDMS network databases to AWS PostgreSQL relational databases.  The goal was to be able to replicate two very large IBM mainframe IDMS databases real-time on two Cloud-based PostgreSQL databases. Implementation was done in phases, starting with one non-production database being replicated to the Cloud.  High-volume testing was performed on the source database to simulate peak processing, replicating millions of transactions to the target PostgreSQL databases.  Many technical challenges were encountered and resolved with outstanding technical assistance from the Treehouse Software support team.  Once in production, the tcVISION product was able to deliver real-time data to the Cloud platform with no interruptions to the customer’s daily processing. The customer was then able to develop modern application features and functions in the Cloud to achieve independence from the legacy mainframe systems.  Using new Cloud-based capabilities enabled the customer to be more agile with meeting new requirements.


__001_TSI_LOGO

Video: tcVISION Demonstration…

In this video, we show a tcVISION overview, then a demonstration of replication of mainframe data on AWS RDS for PostgreSQL:

Contact Treehouse Software for a tcVISION Demo Today!

No matter where you want your mainframe data to go – the Cloud, open systems, or any LUW target – tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software tcVISION Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

Download the AWS and Treehouse Software Mainframe Data Replication Solution Brief

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

___AWS_tcVISION_Solution_Brief

Treehouse Software’s tcVISION data replication product provides connectivity between your mainframe and AWS, putting some of the world’s most advanced technologies at your fingertips. tcVISION is an innovative technology that provides real-time and bi-directional data replication between the mainframe and many AWS targets, including Amazon RDS Aurora, Amazon RDS PostgreSQL, Amazon RDS MySQL/MariaDB, Amazon RDS Oracle, Amazon RDS SQL Server, Amazon S3, Amazon Kinesis, Amazon Redshift, and more. By working with Treehouse Software and using AWS solutions, tools, programs, and databases, you can save time and automate processes. View and download the AWS tcVISION Solution Brief here.


__tsi_logo_400x200

Contact Treehouse Software for a tcVISION Demo Today…

No matter where you want your mainframe data to go – the Cloud, open systems, or any LUW target –tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software Product Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

Let’s Go Shopping – Treehouse Software’s Grocery List for Customers Who are Planning Mainframe Data Replication on AWS

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

When we begin working with customers who have chosen Treehouse Software’s tcVISION mainframe data replication tool to synch data on AWS, many are just beginning to explore AWS. Some customers haven’t even chosen their desired target database(s) or service(s) yet. This blog focuses on helping customers identify the most popular tools and services needed when planning a Mainframe-to-AWS data replication project.

tcVISION‘s data replication connectivity to AWS puts some of the world’s most advanced technologies at customers’ fingertips. On the Mainframe side, the customers’ familiar on-premises environment with their mission critical data resides on various databases, such as IBM Db2, IBM VSAM, IBM IMS/DB, Software AG Adabas, CA IDMS, CA Datacom, or flat files. On the AWS side, there are over 200 fully featured Cloud-based products, including compute, storage, databases, analytics, networking, mobile, developer tools, management tools, IoT, security, and enterprise applications.  How can customers choose what’s right for them?

It’s like a big, virtual grocery store…

___AWS_Services_Grocery_Cart

As we all know, in a grocery store there are many items that appear to do the same thing, but closer inspection shows that certain products can meet specific needs. Here is a list of some of the most popular AWS products and services that we have seen our customers explore and implement for their mainframe data modernization projects:

Amazon EC2Amazon EC2’s simple web service interface allows you to obtain and configure capacity with minimal friction. It provides you with complete control of your computing resources and lets you run on Amazon’s proven computing environment. Amazon EC2 reduces the time required to obtain and boot new server instances to minutes, allowing you to quickly scale capacity, both up and down, as your computing requirements change. Amazon EC2 changes the economics of computing by allowing you to pay only for capacity that you actually use. Amazon EC2 provides developers the tools to build failure resilient applications and isolate them from common failure scenarios.

Amazon RDSAmazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the Cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching, and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need. Amazon RDS is available on several database instance types – optimized for memory, performance or I/O – and provides you with six familiar database engines from which to choose, including Amazon AuroraPostgreSQLMySQLMariaDBOracle Database, and SQL Server.

Amazon S3 – Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics.

Amazon S3 Bucket – An Amazon S3 Bucket is a resource that the user creates in an Amazon region, in which to upload data (photos, videos, documents, etc.).

Amazon CloudWatchAmazon CloudWatch is a powerful Cloud infrastructure monitoring service that gives developers, system operators, site reliability engineers (SRE), and IT managers actionable insights to monitor applications, understand and respond to system-wide performance changes, optimize resource utilization, and obtain a unified view of operational health. Given the dynamic nature of AWS resources, proactive measures including the dynamic re-sizing of infrastructure resources can be automatically initiated. Amazon CloudWatch alarms can be sent to the customer, such as a warning that CPU usage is too high, and as a result, an auto scale trigger can be set up to launch another EC2 instance to address the load. Additionally, customers can set alarms to recover, reboot, or shut down EC2 instances if something out of the ordinary happens.

AWS Direct ConnectAWS Direct Connect is a Cloud service solution that makes it easy to establish a dedicated, secure network connection from your on-premises computing environment to AWS. While in transit, your network traffic remains on the AWS global network and never touches the public internet.

Amazon KinesisAmazon Kinesis makes it easy to collect, process, and analyze real-time, streaming data so you can get timely insights and react quickly to new information. Amazon Kinesis offers key capabilities to cost-effectively process streaming data at any scale, along with the flexibility to choose the tools that best suit the requirements of your application. With Amazon Kinesis, you can ingest real-time data such as video, audio, application logs, website clickstreams, and IoT telemetry data for machine learning, analytics, and other applications. Amazon Kinesis enables you to process and analyze data as it arrives and respond instantly instead of having to wait until all your data is collected before the processing can begin.

AWS LambdaAWS Lambda is an event-driven, “serverless” computing platform provided by Amazon as a part of the Amazon Web Services. It is a computing service that runs code in response to events and automatically manages the computing resources required by that code.

AWS Machine LearningAWS Machine Learning Services offer the broadest and deepest set of machine learning services and supporting Cloud infrastructure, putting machine learning in the hands of every developer, data scientist and expert practitioner.

Amazon RedshiftAmazon Redshift is a popular and fastest growing Cloud data warehouse allowing customers to run and scale analytics in seconds on all of their data without having to manage a data warehouse infrastructure.


Video: tcVISION Mainframe-to-AWS data replication

tcVISION focuses on changed data capture (CDC) when transferring information between mainframe data sources and modern databases and applications. Through an innovative technology, changes occurring in any mainframe application data are tracked and captured, and then published on a variety of AWS targets. The product provides mainframe customers bi-directional, real-time data synchronization of changes on either platform to be reflected on the other platform (e.g., a change to a PostgreSQL table is reflected back on mainframe). The customer can then modernize their application on AWS without disrupting the existing critical work on the legacy system.

Treehouse Software was recently invited by AWS mainframe modernization technical teams to do a presentation and demonstration of tcVISION, our innovative Mainframe-to-Cloud data replication software product.

In this video, an introductory overview to Treehouse Software’s tcVISION is shown, along with a live demonstration of mainframe data replication on AWS RDS for PostgreSQL:


__tsi_logo_400x200

Contact Treehouse Software for a Demo Today…

No matter where you want your mainframe data to go – the Cloud, open systems, or any LUW target –tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software Product Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

Video: Mainframe-to-AWS Data Replication with tcVISION from Treehouse Software

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

___Video_Mainframe_To_Cloud

Treehouse Software was recently invited by AWS mainframe modernization technical teams to do a presentation and demonstration of tcVISION, our innovative Mainframe-to-Cloud data replication software product.

In this video, Chris Rudolph, Treehouse Software’s tcVISION Product Manager shows an overview of the product, then demonstrates replication of mainframe data on AWS RDS for PostgreSQL:


__tsi_logo_400x200

Contact Treehouse Software Today for a tcVISION Demonstration…

No matter where you want your mainframe data to go – the Cloud, Open Systems, or any LUW target – tcVISION from Treehouse Software is your answer.

If interested in more information, please contact customer sales at +1.724.759.7070, email us at sales@treehouse.com, or fill out the Treehouse Software Product Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

Mainframe VSAM Change Data Capture (CDC) to Cloud and Open Systems with tcVISION from Treehouse Software

by Joseph Brady, Director of Business Development and AWS and Cloud Alliance Leader at Treehouse Software, Inc.

tcVISION_Mainframe_VSAM

Treehouse Software is the worldwide distributor of tcVISION, the innovative software product that allows immediate data replication between an impressive array of Mainframe sources and Cloud and Open Systems targets. This blog focuses on tcVISION‘s support of VSAM mainframe data sources (batch and CICS on z/OS, and CICS on z/OS and z/VSE).

tcVISION performs VSAM Change Data Capture (CDC) either via its own “DBMS-Extensions”, or via IBM’s CICS VR product. tcVISION has separate DBMS-Extensions to capture changes from CICS (using the CICS External Interface) and batch (via a JCL wrapper). All captured changes, regardless of whether they are performed by tcVISION or CICS VR are written to the z/OS Logstream on the mainframe. tcVISION then reads the Logstream and transfers the transactions to a tcVISION server running in the Cloud or on-prem, which is responsible for queueing, transforming, and applying the captured changes to the specified target.

Additionally, when planning VSAM CDC there are a number of operational items to consider, such as volume of batch transactions, data changes that occur during periods of time while the VSAM file is offline, etc.

In this instructional video, tcVISION is shown capturing changes from VSAM on z/OS and writing them to SQL Server on Windows:

 


__tsi_logo_400x200

Contact Treehouse Software Today for a tcVISION Demonstration…

No matter where you want your mainframe data to go – the Cloud, Open Systems, or any LUW target – tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software Product Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

How to Replicate Mainframe Data to a Big Data Environment via Kafka with tcVISION

by Joseph Brady, Director of Business Development and AWS and Cloud Alliance Leader at Treehouse Software, Inc.

tcVISION from Treehouse Software allows enterprise customers to replicate data between mainframe, Cloud, or Hybrid Cloud while maintaining their legacy environments, and one of the more popular targets for mainframe modernization that we have been seeing is Apache Kafka®.

tcVISION_Mainframe_To_Kafka

What is Kafka? 

Kafka is an open-source distributed event streaming platform used by thousands of companies for high-performance data pipelines, streaming analytics, data integration, and mission-critical applications. A data pipeline processes and moves data from one system to another, and a streaming application is an application that consumes streams of data.

Kafka is reliable, stable, flexible, robust, and scales well with numerous consumers, working seamlessly with most popular data warehouses and data lakes like Hadoop, Redshift, S3, BigQuery, Azure, etc. Kafka can also be used for real-time analytics, as well as to process real-time streams to collect Big Data.

See how tcVISION easily connects mainframe systems to Kafka…

Kafka handles massive volumes of data and remains responsive, making Kafka a preferred platform when the volume of the data at the mainframe level –> BIG.

Kafka is a supported target in tcVISION, and in this instructional video, tcVISION is shown synchronizing data in real-time from Db2 on z/OS via Kafka to a Big Data environment:

Additional Reading: Treehouse Software is a Confluent technology partner and we recently co-authored a blog entitled, “Enterprise Mainframe Change Data Capture (CDC) to Apache Kafka with tcVISION and Confluent”.


__tsi_logo_400x200

Contact Treehouse Software Today…

No matter where you want your mainframe data to go – the Cloud, Open Systems, or any LUW target – tcVISION from Treehouse Software is your answer.

Just fill out the Treehouse Software Product Demonstration Request Form and a Treehouse representative will contact you to set up a time for your online tcVISION demonstration.

New Faces at Treehouse Software

by Joseph Brady, Director of Business Development and Cloud Alliance Leader at Treehouse Software, Inc.

__TSI_LOGO

Treehouse Software is growing and on the move! We are proud to have many staff members who have been here for 20+ years, and we have recently brought on several experienced business, mainframe, and Cloud experts. Meet our newest team members:

John Szakach, Chief Operating Officer

John joined Treehouse as a Business Strategy Consultant and in 2021 was promoted to Chief Operating Officer. While new to Treehouse, John brings over 40 years of relevant work experience to the organization and is an AWS Certified Cloud Practitioner as well as a Certified Project Management Professional. John has held a variety of management roles in different industries including VP of Organizational Effectiveness, VP of Quality Assurance, and VP of New Product Development. He has also held positions as Director of Flight Standards and Quality Control, and Director of Operations. In addition to over 51 years of total flight experience, including 20 years as a pilot for United Airlines, he has received numerous awards including the United Airlines Captain of the Year and the FAA Master Pilot Award, the FAA’s highest award for safety and compliance. John has a Bachelor’s Degree in aviation management.

Dan Miley, Product Support

Dan is a software engineer with deep experience and understanding of IBM Assembler, COBOL, JCL, IDMS, SAP ECC. He has worked with some of the world’s largest organizations, including president/consultant of his own company for over 10 years. Dan has already been instrumental in landing some major mainframe-to-Cloud data modernization customers for Treehouse Software.

Sasha Efron, Senior Technical Representative

Sasha is a mainframe technical specialist and DBA with over 25 years experience in in systems analysis, design, development, enhancement, testing, implementation and maintenance in insurance and banking systems with specialization in Software AG and IBM Mainframe technologies. He also has been involved in legacy modernization projects for several worldwide companies.

Joseph Rogan, Senior Technical Representative

Joseph is a Senior Technology Leader with 30+ years experience working in multiple industries, including transportation (specifically rail), logistics, education, financial services (banking, re-insurance, and trading systems), commercial insurance, and state government. His core competencies include database design and implementation, OLTP, OLAP, and data warehouse design, project planning, and project management. Joseph is also a highly trusted, conceptual, business partner and leader with excellent presentation, negotiating, management, mentoring, and strategic planning skills.

Daniel Vimont, Senior Technical Representative

Daniel brings 30+ years experience in multiple computer languages, databases, frameworks, and distributed processing for mainframe, Cloud, and open systems. He is very familiar with the principles of ETL and CDC in mainframe data transformation and migration. Dan is a Certified AWS Cloud Practitioner and has experience in designing and developing AWS/SDK (boto3) framework for on-premises invocation/monitoring of AWS services. Additionally, Dan’s versatile background as a data and software engineer, educator, and business advisor is a valuable asset to Treehouse’s vision of being a close partner in our customers’ planning and modernization efforts.

Treehouse Software Experts are Our Best Assets

management-team

When asked by prospective customers, “What are your primary differentiators?”, we immediately point to our people who have decades worth of experience in helping mainframe customers with innovative tools, services, and training. Our extensive experience, deep knowledge, and wide-ranging capabilities in mainframe technologies make the company a valued partner for third-party solution providers and a trusted advisor to customers.

We are fortunate to have a staff with a wealth of knowledge and skills that span not only Mainframe, but Cloud, LUW, and Open Systems technologies. Treehouse Software‘s technicians have installed products and trained end-users in some of the largest mainframe sites around the world, and our highly-rated 24X7 technical support is second to none.

The Treehouse Team Approach

Treehouse Software’s expert staff has proven its ability to work effectively as part of a larger team to meet clients’ complex business goals. AWS, Google, Microsoft, IBM, Oracle, Deloitte, Accenture, Confluent, and other large vendors have selected our expertise, technology, services, and training for their mainframe data modernization practices.


__tsi_logo_400x200

Contact Treehouse Software Today…

Treehouse Software has been helping enterprises mainframe customers since 1982, and in recent years, we have been developing a strong presence in the Cloud market space relating to mainframe data replication and modernization. As a result, Treehouse Software is currently working with technical and sales leaders from all popular Cloud platform companies and major systems integrators to take advantage of our deep mainframe skills and our tcVISION Mainframe-to-Cloud data replication solution.

No matter where you want your mainframe data to go – the Cloud, Open Systems, or any LUW target –Treehouse Software is here to help. Contact us to discuss your needs.