Taming Big Data with Apache Spark and Python – Hands On!
Sundog Education by Frank Kane – Training the World in Big Data and Machine Learning
Taming Big Data with Apache Spark and Python – Hands On!: Dive right in with 15+ hands-on examples of analyzing large data sets with Apache Spark, on your desktop or on Hadoop!
Taming Big Data with Apache Spark and Python – Hands On! Price
Frame big data analysis problems as Spark problems
Use Amazon’s Elastic MapReduce service to run your job on a cluster with Hadoop YARN
Install and run Apache Spark on a desktop computer or on a cluster
Use Spark’s Resilient Distributed Datasets to process and analyze large data sets across many CPU’s
Implement iterative algorithms such as breadth-first-search using Spark
Use the MLLib machine learning library to answer common data mining questions
Understand how Spark SQL lets you work with structured data
Understand how Spark Streaming lets your process continuous streams of data in real time
Tune and troubleshoot large jobs running on a cluster
Share information between nodes on a Spark cluster using broadcast variables and accumulators
Understand how the GraphX library helps with network analysis problems
- Access to a personal computer. This course uses Windows, but the sample code will work fine on Linux as well.
- Some prior programming or scripting experience. Python experience will help a lot, but you can pick it up as we go.
Taming Big Data with Apache Spark and Python – Hands On!: Description
New! Updated for Spark 2.0.0
“Big data” analysis is a hot and highly valuable skill – and this course will teach you the hottest technology in big data: Apache Spark. Employers including Amazon, EBay, NASA JPL, and Yahoo all use Spark to quickly extract meaning from massive data sets across a fault-tolerant Hadoop cluster. You’ll learn those same techniques, using your own Windows system right at home. It’s easier than you might think.
Learn and master the art of framing data analysis problems as Spark problems through over 15 hands-on examples, and then scale them up to run on cloud computing services in this course. You’ll be learning from an ex-engineer and senior manager from Amazon and IMDb.
Learn the concepts of Spark’s Resilient Distributed DatastoresDevelop and run Spark jobs quickly using PythonTranslate complex analysis problems into iterative or multi-stage Spark scriptsScale up to larger data sets using Amazon’s Elastic MapReduce serviceUnderstand how Hadoop YARN distributes Spark across computing clustersLearn about other Spark technologies, like Spark SQL, Spark Streaming, and GraphX
By the end of this course, you’ll be running code that analyzes gigabytes worth of information – in the cloud – in a matter of minutes.
This course uses the familiar Python programming language; if you’d rather use Scala to get the best performance out of Spark, see my “Apache Spark with Scala – Hands On with Big Data” course instead.
We’ll have some fun along the way. You’ll get warmed up with some simple examples of using Spark to analyze movie ratings data and text in a book. Once you’ve got the basics under your belt, we’ll move to some more complex and interesting tasks. We’ll use a million movie ratings to find movies that are similar to each other, and you might even discover some new movies you might like in the process! We’ll analyze a social graph of superheroes, and learn who the most “popular” superhero is – and develop a system to find “degrees of separation” between superheroes. Are all Marvel superheroes within a few degrees of being connected to The Incredible Hulk? You’ll find the answer.
This course is very hands-on; you’ll spend most of your time following along with the instructor as we write, analyze, and run real code together – both on your own system, and in the cloud using Amazon’s Elastic MapReduce service. 5 hours of video content is included, with over 15 real examples of increasing complexity you can build, run and study yourself. Move through them at your own pace, on your own schedule. The course wraps up with an overview of other Spark-based technologies, including Spark SQL, Spark Streaming, and GraphX.
Enjoy the course!
Who is the target audience?
People with some software development background who want to learn the hottest technology in big data analysis will want to check this out. This course focuses on Spark from a software development standpoint; we introduce some machine learning and data mining concepts along the way, but that’s not the focus. If you want to learn how to use Spark to carve up huge datasets and extract meaning from them, then this course is for you.
If you’ve never written a computer program or a script before, this course isn’t for you – yet. I suggest starting with a Python course first, if programming is new to you.
If your software development job involves, or will involve, processing large amounts of data, you need to know about Spark.
If you’re training for a new career in data science or big data, Spark is an important part of it.
The program targets All Levels and has a duration of 5 hours divided on 45 sessions and has 1 supplementary learning materials. This course has been initially developed by Sundog Education by Frank Kane / Inst. Title, as part of a partnership with Udemy ®.
Description. This course is just one of a series about ‘PHP ’, which introduces you to the complete skills and tools you need to study effectively about Taming Big Data with Apache Spark and Python – Hands On! online.
Become an effective online learner and develop your online communication skills, explore new interests and career opportunities and learn to code or develop your programming skills by having this online development course, from beginner to advanced level.