Book Image

MySQL 8 for Big Data

By : Shabbir Challawala, Chintan Mehta, Kandarp Patel, Jaydip Lakhatariya
Book Image

MySQL 8 for Big Data

By: Shabbir Challawala, Chintan Mehta, Kandarp Patel, Jaydip Lakhatariya

Overview of this book

With organizations handling large amounts of data on a regular basis, MySQL has become a popular solution to handle this structured Big Data. In this book, you will see how DBAs can use MySQL 8 to handle billions of records, and load and retrieve data with performance comparable or superior to commercial DB solutions with higher costs. Many organizations today depend on MySQL for their websites and a Big Data solution for their data archiving, storage, and analysis needs. However, integrating them can be challenging. This book will show you how to implement a successful Big Data strategy with Apache Hadoop and MySQL 8. It will cover real-time use case scenario to explain integration and achieve Big Data solutions using technologies such as Apache Hadoop, Apache Sqoop, and MySQL Applier. Also, the book includes case studies on Apache Sqoop and real-time event processing. By the end of this book, you will know how to efficiently use MySQL 8 to manage data for your Big Data applications.
Table of Contents (17 chapters)
Title Page
Credits
About the Authors
About the Reviewers
www.PacktPub.com
Customer Feedback
Preface

Real-time integration with MySQL Applier


There are many MySQL applier packages available on GitHub. We can use any of them which provides framework for replication and an example of real-time replication:

  • Flipkart/MySQL-replication-listener
  • SponsorPay/MySQL-replication-listener
  • bullsoft/MySQL-replication-listener

For our configuration, let's use Flipkart/MySQL-replication-listener. You can clone the Git library using the following command:

$ git clone https://github.com/Flipkart/MySQL-replication-listener.git

Here are some environment variables required by the package. Make sure that all are set properly.

  • HADOOP_HOME: The Hadoop root directory path
  • CMAKE_MODULE_PATH: The path of the root directory where FindHDFS.cmake and FindJNI.cmake files are located in HDFS
  • HDFS_LIB_PATHS: The path of the libhdfs.so file available in HADOOP
  • JAVA_HOME: You need to set the Java home path for this variable

Now build and compile all the libraries using the following command:

$ cd src
$ cmake . -DCMAKE_MODULE_PATH:String...