Big Data Analytics Using Spark

In data science, data is called “big” if it cannot fit into the memory of a single standard laptop or workstation. The analysis of big datasets requires using a cluster of tens, hundreds or thousands of computers. Effectively using such clusters requires the use of distributed files systems, such as the Hadoop Distributed File System (HDFS) and corresponding computational models, such as Hadoop, MapReduce and Spark. In this course, part of the Data Science MicroMasters program, you will learn what the bottlenecks are in massive parallel computation and how to use spark to minimize these bottlenecks. You will learn how to perform supervised an unsupervised machine learning on massive datasets using the Machine Learning Library (MLlib). In this course, as in the other ones in this MicroMasters program, you will gain hands-on experience using PySpark within the Jupyter notebooks environment.

Created by: The University of California, San Diego

Level: Advanced

Find Out More
Share
Facebook
Twitter
Pinterest
Reddit
StumbleUpon
LinkedIn
Email

Online Courses

Back to Top

Log In

Contact Us

Upload An Image

Please select an image to upload
Note: must be in .png, .gif or .jpg format
OR
Provide URL where image can be downloaded
Note: must be in .png, .gif or .jpg format

By clicking this button,
you agree to the terms of use

By clicking "Create Alert" I agree to the Uloop Terms of Use.

Image not available.

Add a Photo

Please select a photo to upload
Note: must be in .png, .gif or .jpg format