I have a 'user' table in MySQL. I want to read it to my Spark SQL program. How can I read the table from MySQL to the Apache Spark's SparkSQL module using Python? Is there a connector I can use for this task? Thanks.
There is a similar question answered. Start pyspark
like this
./bin/pyspark --packages mysql:mysql-connector-java:5.1.38
Then just run
sqlContext.read.format("jdbc").options(
url ="jdbc:mysql://localhost/mysql",
driver="com.mysql.jdbc.Driver",
dbtable="user",
user="root",
password=""
).load().take(10)
This would most likely just work. But this depends on your mysql set-up, so if it doesn't try changing password, username, db-url and other settings.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With