Curt Beck
1 min readFeb 24, 2020

--

Hi Susan,

I am just starting out with Machine Learning in Spark and this guide was a great introduction. Really great walk through!

Just wanted to add something I came across on the web: In Spark 2.0, you can create a SparkContext in a more concise way:

from pyspark.sql import SparkSession

spark = SparkSession.builder.getOrCreate()

SparkSession essentially condenses SparkConf, SparkContext, sqlContext all into one unified API

Thanks again for the article!

--

--

Curt Beck
Curt Beck

Written by Curt Beck

Stumbled into a data-centric role several years ago and have not looked back! Passionate about leveraging technology to uncover answers and improve the world.

No responses yet