Book Image

Learning Pentaho Data Integration 8 CE - Third Edition

By : Carina Roldán
Book Image

Learning Pentaho Data Integration 8 CE - Third Edition

By: Carina Roldán

Overview of this book

Pentaho Data Integration(PDI) is an intuitive and graphical environment packed with drag-and-drop design and powerful Extract-Tranform-Load (ETL) capabilities. This book shows and explains the new interactive features of Spoon, the revamped look and feel, and the newest features of the tool including transformations and jobs Executors and the invaluable Metadata Injection capability. We begin with the installation of PDI software and then move on to cover all the key PDI concepts. Each of the chapter introduces new features, enabling you to gradually get practicing with the tool. First, you will learn to do all kind of data manipulation and work with simple plain files. Then, the book teaches you how you can work with relational databases inside PDI. Moreover, you will be given a primer on data warehouse concepts and you will learn how to load data in a data warehouse. During the course of this book, you will be familiarized with its intuitive, graphical and drag-and-drop design environment. By the end of this book, you will learn everything you need to know in order to meet your data manipulation requirements. Besides, your will be given best practices and advises for designing and deploying your projects.
Table of Contents (17 chapters)

Automating the execution

As you just learned, running Pan and Kitchen not only involves providing the name of the ktr or kjb file, but also typing several options, for example, parameters or names of log files. You can type the full command manually when you are developing or testing, but when your work is ready for production, you want to keep things simple and automated. The following tutorial explains how to embedd the execution of Kitchen inside a script. Once you have the script, you can schedule its execution using a system utility, for example, cron in Unix or scheduler in Windows.

Suppose that you have a Job named process_sales.kjb located in the c:\project\etl folder (Windows) or /home/project/etl folder (Unix), and you want to run it every day. You want to keep a history of logs, so the log of the execution will be written in a folder named C:\project\logs (Windows...