Book Image

Microsoft SQL Server 2012 with Hadoop

By : Debarchan Sarkar
Book Image

Microsoft SQL Server 2012 with Hadoop

By: Debarchan Sarkar

Overview of this book

With the explosion of data, the open source Apache Hadoop ecosystem is gaining traction, thanks to its huge ecosystem that has arisen around the core functionalities of its distributed file system (HDFS) and Map Reduce. As of today, being able to have SQL Server talking to Hadoop has become increasingly important because the two are indeed complementary. While petabytes of unstructured data can be stored in Hadoop taking hours to be queried, terabytes of structured data can be stored in SQL Server 2012 and queried in seconds. This leads to the need to transfer and integrate data between Hadoop and SQL Server. Microsoft SQL Server 2012 with Hadoop is aimed at SQL Server developers. It will quickly show you how to get Hadoop activated on SQL Server 2012 (it ships with this version). Once this is done, the book will focus on how to manage big data with Hadoop and use Hadoop Hive to query the data. It will also cover topics such as using in-memory functions by SQL Server and using tools for BI with big data. Microsoft SQL Server 2012 with Hadoop focuses on data integration techniques between relational (SQL Server 2012) and non-relational (Hadoop) worlds. It will walk you through different tools for the bi-directional movement of data with practical examples. You will learn to use open source connectors like SQOOP to import and export data between SQL Server 2012 and Hadoop, and to work with leading in-memory BI tools to create ETL solutions using the Hive ODBC driver for developing your data movement projects. Finally, this book will give you a glimpse of the present day self-service BI tools such as Excel and PowerView to consume Hadoop data and provide powerful insights on the data.
Table of Contents (12 chapters)

The Sqoop import tool


You're now ready to use SQL Server-Hadoop Connector and import data from SQL Server 2012 to HDFS. The input to the import process is a SQL Server table, which will be read row-by-row into HDFS by Sqoop. The output of this import process is a set of files containing a copy of the imported table. Since the import process is performed in parallel, the output will be in multiple files.

When using the sqoop import command, you must specify the following mandatory arguments:

  • --connect argument specifying the connection string to the SQL Server database

  • --username and --password arguments to provide valid credentials to connect to the SQL Server database

  • --table or --query argument to import an entire table or results of a custom query execution

The following command imports data from ErrorLog table in SQL Server Adventureworks2012 database to delimited text files in /data/ErrorLogs directory on HDFS.

Note

Sqoop 1.4.2 does not recognize SQL Server tables, which do not belong to...