Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

how to properly build spark 2.0 from source, to include pyspark?

I've just built spark 2.0 on Ubuntu host, using "sbt assembly". Everything finished fine, but, when I attempted to submit a pyspark job:

bin/spark-submit --master spark://localhost:7077 examples/src/main/python/pi.py 1000

I got this error:

Failed to find Spark jars directory (/home/ubuntu/spark/spark-2.0.0/assembly/target/scala-2.10/jars).
You need to build Spark with the target "package" before running this program.

What should I do in order to re-build spark 2.0 to include pyspark?

like image 611
Eugene Goldberg Avatar asked Jan 23 '26 17:01

Eugene Goldberg


1 Answers

Try:

  1. Install sbt

  2. Build:

    https://github.com/apache/spark.git
    cd spark
    git checkout v2.0.0
    sbt package
    
like image 140
user6022341 Avatar answered Jan 27 '26 01:01

user6022341



Donate For Us

If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!