Distributed Machine Learning With PySpark

Time to read
less than
1 minute
Read so far

Distributed Machine Learning With PySpark

Fri, 10/05/2019 - 10:00
scout

Spark is known as a fast general-purpose cluster-computing framework for processing big data. In this post, we’re going to cover how Spark works under the hood and the things you need to know to be able to effectively perform distributing machine learning using PySpark. The post assumes basic familiarity with Python and the concepts of machine learning like regression, gradient descent, etc.

https://scoutapm.com/blog/distributed-machine-learning-with-pyspark