Book Image

Java Data Analysis

By : John R. Hubbard
Book Image

Java Data Analysis

By: John R. Hubbard

Overview of this book

Data analysis is a process of inspecting, cleansing, transforming, and modeling data with the aim of discovering useful information. Java is one of the most popular languages to perform your data analysis tasks. This book will help you learn the tools and techniques in Java to conduct data analysis without any hassle. After getting a quick overview of what data science is and the steps involved in the process, you’ll learn the statistical data analysis techniques and implement them using the popular Java APIs and libraries. Through practical examples, you will also learn the machine learning concepts such as classification and regression. In the process, you’ll familiarize yourself with tools such as Rapidminer and WEKA and see how these Java-based tools can be used effectively for analysis. You will also learn how to analyze text and other types of multimedia. Learn to work with relational, NoSQL, and time-series data. This book will also show you how you can utilize different Java-based libraries to create insightful and easy to understand plots and graphs. By the end of this book, you will have a solid understanding of the various data analysis techniques, and how to implement them using Java.
Table of Contents (20 chapters)
Java Data Analysis
Credits
About the Author
About the Reviewers
www.PacktPub.com
Customer Feedback
Preface
Index

Scalability


The great benefit of the MapReduce framework is that it is scalable. The WordCount program in Example1.java was run on 80 files containing fewer than 10,000 words. With little modification, it could be run on 80,000 files with 10,000,000 words. That flexibility in software is called scalability.

To manage that thousand-fold increase in input, the hash table might have to be replaced. Even if we had enough memory to load a table that large, the Java processing would probably fail because of the proliferation of objects. Object-oriented programming is certainly the best way to implement an algorithm. But if you want clarity, speed, and flexibility it is not so efficient at handling large datasets.

We don't really need the hash table, which is instantiated at line 24 in Listing 11-1. We can implement the same idea by hashing the data into a set of files instead. This is illustrated in Figure 11-3.

Replacing the hash table with file chunks would require modifying the code at lines 34...