Book Image

Apache Spark 2.x Cookbook

By : Rishi Yadav
Book Image

Apache Spark 2.x Cookbook

By: Rishi Yadav

Overview of this book

While Apache Spark 1.x gained a lot of traction and adoption in the early years, Spark 2.x delivers notable improvements in the areas of API, schema awareness, Performance, Structured Streaming, and simplifying building blocks to build better, faster, smarter, and more accessible big data applications. This book uncovers all these features in the form of structured recipes to analyze and mature large and complex sets of data. Starting with installing and configuring Apache Spark with various cluster managers, you will learn to set up development environments. Further on, you will be introduced to working with RDDs, DataFrames and Datasets to operate on schema aware data, and real-time streaming with various sources such as Twitter Stream and Apache Kafka. You will also work through recipes on machine learning, including supervised learning, unsupervised learning & recommendation engines in Spark. Last but not least, the final few chapters delve deeper into the concepts of graph processing using GraphX, securing your implementations, cluster optimization, and troubleshooting.
Table of Contents (19 chapters)
Title Page
Credits
About the Author
About the Reviewer
www.PacktPub.com
Customer Feedback
Preface

Developing a Spark applications in Eclipse with SBT


SBT is a build tool made especially for Scala-based development. SBT follows Maven-based naming conventions and declarative dependency management.

SBT provides the following enhancements over Maven:

  • Dependencies are in the form of key-value pairs in the build.sbt file, as opposed to the pom.xml file in Maven
  • It provides a shell that makes it very handy to perform build operations
  • For simple projects without dependencies, you do not even need the build.sbt file

In the build.sbt file, the first line is the project definition:

lazy val root = (project in file("."))

Each project has an immutable map of key-value pairs. This map is changed by the settings in SBT, as follows:

lazy val root = (project in file(".")). 
  settings( 
    name := "wordcount" 
  ) 

Every change in the settings field leads to a new map, as it's an immutable map.

How to do it...

Here's how we go about adding the sbteclipse plugin:

  1. Add this to the global plugin file:
$ mkdir /home...