New installation of a Treehouse Software product? New employees? Need a refresher course? Want to explore untapped product features? Three-to-five-day training sessions are available for tRelational / DPS and other Treehouse products.
Treehouse Software training sessions are customized to meet your site’s unique environment and needs. As an example, the following three-day tRelational / DPS training class was conducted at a large University where they are using tRelational / DPS to create an archive of their legacy Adabas data to Oracle in preparation for their upcoming mainframe system retirement.
Over the course of the three-day session, the Treehouse instructor covered:
- tRelational AUTOGEN feature (auto-generation of complete RDBMS schemata – Tables, Columns, Primary Key, and Foreign Keys – based upon existing ADABAS file structures). AUTOGEN was quite attractive since it will be a very time efficient way to create the Adabas-to-Oracle mappings.
- The customer wanted to use Relational / DPS to automate as much of the process as possible, and take advantage of the batch functions of tRelational. Generally, the workshops of the class are all online, but in this case, the class ran through one online example after the File Implementation section of the training, and then jumped to the batch section of the training to show how that works. The class then returned to run additional File Implementations, Analysis, and Reporting in batch.
- The class continued through the Modeling and Mapping section, and ran more batch jobs with some additional Adabas files.
- Admin and Configuring DPS Parameters sections were next on the agenda. Because of the desire to automate as much as possible, the work focused on how to set up a single job stream that could be generated programmatically, where it was fed in the name of the file to be implemented and then do the AUTOGEN all in the same job. The tRelational AUTOGEN user exit was set up so the customer could identify and cater to the possibility of duplicate table names where multiple Adabas files have the same named MU (like “COMMENTS”); and handle the addition of an ETR (external transformation routine) being automatically added (there were some known HEX values in some fields) and then update the DPSCOLLGTH.
- The class finished up the Materialization (initial loading of data) training section and then set up the jobs to run Materialization, and installed DPSSPLIT (used to separate the control and data files into one control and one data file per table materialized) on one of their servers. Generally, most class attendees are not the ones responsible for running the Materialization and Propagation (ongoing synchronization of data) jobs, so the instructor doesn’t usually set them up, but in this case, it made sense. The class started out by running the Materialization from a full database ADASAV taken over the weekend. The output of the Materialization job was FTPed to their Server and DPSSPLIT run against it to create the individual loader files. GENDDL (generate the control statements that can be passed to the RDBMS to define the tables and columns) was also run and that output transferred. After setting up the SQL job to process the DDL, the Oracle SQL*LOADER was successfully run to load the data.
Finally, the class went through some additional tRelational / DPS features, and as a final exercise, set up a Propagation job and ran it against a current PLOG just to make sure it worked satisfactorily.
If you are interested in finding out more about Treehouse training classes for your product, contact Treehouse Software today for more information, or to schedule onsite training.