Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
MySQL 8 for Big Data

You're reading from   MySQL 8 for Big Data Effective data processing with MySQL 8, Hadoop, NoSQL APIs, and other Big Data tools

Arrow left icon
Product type Paperback
Published in Oct 2017
Publisher Packt
ISBN-13 9781788397186
Length 296 pages
Edition 1st Edition
Languages
Tools
Concepts
Arrow right icon
Authors (4):
Arrow left icon
Chintan Mehta Chintan Mehta
Author Profile Icon Chintan Mehta
Chintan Mehta
Shabbir Challawala Shabbir Challawala
Author Profile Icon Shabbir Challawala
Shabbir Challawala
Jaydip Lakhatariya Jaydip Lakhatariya
Author Profile Icon Jaydip Lakhatariya
Jaydip Lakhatariya
Kandarp Patel Kandarp Patel
Author Profile Icon Kandarp Patel
Kandarp Patel
Arrow right icon
View More author details
Toc

Table of Contents (11) Chapters Close

Preface 1. Introduction to Big Data and MySQL 8 FREE CHAPTER 2. Data Query Techniques in MySQL 8 3. Indexing your data for High-Performing Queries 4. Using Memcached with MySQL 8 5. Partitioning High Volume Data 6. Replication for building highly available solutions 7. MySQL 8 Best Practices 8. NoSQL API for Integrating with Big Data Solutions 9. Case study: Part I - Apache Sqoop for exchanging data between MySQL and Hadoop 10. Case study: Part II - Real time event processing using MySQL applier

Integrating Apache Sqoop with MySQL and Hadoop


Apache Sqoop can only work if Hadoop is installed on the server. Apache Sqoop requires Linux based operating system to work . ForHadoop and Sqoop to work on the Linux server, Java must be installed on the server. Once Sqoop is installed on the server, we will need to download Sqoop's MySQL connector which will allow JDBC driver to connect with MySQL database for transferring data with Hadoop.

Hadoop

is an open source, Big Data framework to process and analyze large amount of data sets quickly by using a cluster of environment. Because of Hadoop's multiple slave nodes environment, it's easy to avoid system failure or data loss if one or more nodes go off. Hadoop basically works with multiple modules such as Yet Another Resource Negotiator (YARN), Hadoop distributed file system (HDFS), and MapReduce. Hadoop's MapReduce algorithm is used for parallel processing of the data. MapReduce is used to convert unstructured data to a structured format using...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image