0 votes

While trying to run the sample code provided in the Jupyter Python Spark Notebook, I get an error "no module named pyspark.sql" :

Do I need to configure something in order to use pyspark ?
I'm running DSS community on an EC2 AMI.

by

1 Answer

0 votes
Best answer
Hi,

You need to setup the DSS / Spark integration.

* For DSS 3.1: https://doc.dataiku.com/dss/3.1/installation/spark.html

* For DSS 4: https://doc.dataiku.com/dss/latest/spark/installation.html
by
1,078 questions
1,123 answers
1,246 comments
10,669 users

©Dataiku 2012-2018 - Privacy Policy