Book Image

Pentaho for Big Data Analytics

By : Manoj R Patil, Feris Thia
Book Image

Pentaho for Big Data Analytics

By: Manoj R Patil, Feris Thia

Overview of this book

<p>Pentaho accelerates the realization of value from big data with the most complete solution for big data analytics and data integration. The real power of big data analytics is the abstraction between data and analytics. Data can be distributed across the cluster in various formats, and the analytics platform should have the capability to talk to different heterogeneous data stores and fetch the filtered data to enrich its value.<br /><br />Pentaho Big Data Analytics is a practical, hands-on guide that provides you with clear, step-by-step exercises for using Pentaho to take advantage of big data systems, where data beats algorithm, and gives you a good grounding in using Pentaho Business Analytics’ capabilities.<br /><br />This book looks at the key ingredients of the Pentaho Business Analytics platform. We will see how to prepare the Pentaho BI environment, and get to grips with the big data ecosystem through. The book provides a clear guide to the essential tools of Pentaho Business Analytics, providing familiarity with both the various design tools for setting up reports, and the visualization tools necessary for complete data analysis.</p>
Table of Contents (14 chapters)
Pentaho for Big Data Analytics
Credits
About the Authors
About the Reviewers
www.PacktPub.com
Preface
Index

Transferring a file using secure FTP


The following steps will help you transfer a file using a secure FTP:

  1. Download the FileZilla installer from https://filezilla-project.org/. FileZilla is an open source FTP client that supports a secure FTP connection.

  2. Launch the installer and accept all the default options.

  3. Now, launch the FileZilla application.

  4. In the File menu, click on Site Manager.

  5. When the Site Manager dialog appears, click on the New Site button. This will create a new site entry; type in hortonworks as its name.

  6. In the Host textbox, type 192.168.1.122 as the destination host. Leave the Port textbox empty.

  7. In the Protocol listbox, select SFTP – SSH as the file transfer protocol.

  8. In the User textbox, type root, and in the Password textbox, type hadoop. Please note that all the entries are case sensitive.

  9. Click on the Connect button to close the dialog, which in turn starts an FTP session at the destination host.

  10. Once connected, you can transfer files between the localhost and the VM. In Chapter 3, Churning Big Data with Pentaho, we downloaded core-site.xml using this mechanism. We can also download the file from one of these locations: /usr/lib/hadoop/conf or /etc/hadoop/conf.empty /core-site.xml. The following screenshot shows a FileZilla SFTP session: