Unable to import pyspark module


I am using Itversity lab for pyspark(CCA 175 Spark and Hadoop Developer - Python(pyspark).

I was trying to install pyspark from python3.6. But Unable to install pyspark because of of permission issue.

Then i tried to import pyspark module from python 3.5 and python3.6 assuming that pyspark is already installed.

But I am getting error while importing pyspark module.

Could you please let me know how to import pyspark module from python3 in itversity lab?

All the errors are given in the attached screen shot.

Learn Spark 1.6.x or Spark 2.x on our state of the art big data labs

  • Click here for access to state of the art 13 node Hadoop and Spark Cluster

@itversity…Could you please help me with this query?

Tapas Kabi