In this course, you’ll learn how to use Spark to work with big data and build machine learning models at scale, including how to wrangle and model massive datasets with PySpark, the Python library for interacting with Spark. In the first lesson, you will learn about big data and how Spark fits into the big data ecosystem. In lesson two, you will be practicing processing and cleaning datasets to get comfortable with Spark’s SQL and dataframe APIs. In the third lesson, you will debug and optimize your Spark code when running on a cluster. In lesson four, you will use Spark’s Machine Learning Library to train machine learning models at scale.Read more.
This resource is offered by an affiliate partner. If you pay for training, we may earn a commission to support this site.
The techniques and tools covered in Spark are most similar to the requirements found in Data Engineer job advertisements.
Spark is a part of three structured learning paths.