i have spark instance , i'm trying connect existing netezza datawarehouse applicance retrieve data.
using sparksql's sqlcontext
, , according spark sql programming guide, achievable read
method. i've determined need provide jdbc driver using --jars
flag, rather spark_classpath
in documentation. operation looks
// pyspark df = sqlcontext.read.format('jdbc').options( ... ).load() // spark-shell val df = sqlcontext.read.format("jdbc").options( ... ).load()
i can find documentation connecting netezza using jdbc, not how correctly pass username , password. 'options' need pass here?
in pyspark
df = sqlcontext.read.format('jdbc').options(url='jdbc:netezza://server1:5480/database', \ user='kirk', password='****', dbtable='schema.mytable', \ driver='org.netezza.driver').load()
and in spark-shell
val df = sqlcontext.read.format("jdbc").options(map( "url" -> "jdbc:netezza://server1:5480/database", "user" -> "kirk", "password" -> "****", "dbtable" -> "schema.mytable", "driver" -> "org.netezza.driver")).load()
note netezza likes things in caps. don't know if necessary, doesn't hurt.
Comments
Post a Comment