Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Analytics for the Internet of Things (IoT)

You're reading from   Analytics for the Internet of Things (IoT) Intelligent analytics for your intelligent devices

Arrow left icon
Product type Paperback
Published in Jul 2017
Publisher Packt
ISBN-13 9781787120730
Length 378 pages
Edition 1st Edition
Languages
Arrow right icon
Author (1):
Arrow left icon
Andrew Minteer Andrew Minteer
Author Profile Icon Andrew Minteer
Andrew Minteer
Arrow right icon
View More author details
Toc

Table of Contents (14) Chapters Close

Preface 1. Defining IoT Analytics and Challenges FREE CHAPTER 2. IoT Devices and Networking Protocols 3. IoT Analytics for the Cloud 4. Creating an AWS Cloud Analytics Environment 5. Collecting All That Data - Strategies and Techniques 6. Getting to Know Your Data - Exploring IoT Data 7. Decorating Your Data - Adding External Datasets to Innovate 8. Communicating with Others - Visualization and Dashboarding 9. Applying Geospatial Analytics to IoT Data 10. Data Science for IoT Analytics 11. Strategies to Organize Data for Analytics 12. The Economics of IoT Analytics 13. Bringing It All Together

Apache Spark for data processing


Apache Spark is a new-ish project (at least in the world of big data, which moves at warp speed) that integrates well with Hadoop but does not necessarily require Hadoop components to operate. It is a

fast and general engine for large-scale data processing

as described on the Spark project team welcome page. The tagline of

lightning fast cluster computing

is a little catchier: we like that one better.

Apache Spark logo

What is Apache Spark?

Good question, glad you asked. Spark was built for distributed cluster computing, so everything scales nicely without any code changes. The word general in the general engine description is very appropriate for Spark. It refers to the many and varied ways you can use it.

You can use it for ETL data processing, machine learning modeling, graph processing, stream data processing, and SQL and structure data processing. It is a boon for analytics in a distributed computing world.

It has APIs for multiple programming languages such...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image