Is it possible to broadcast Tensorflow libraries using spark-submit -package -
i using cluster, not managed myself. tensorflow libraries not installed on cluster nodes. run spark programs using tensorflow package. not sure if possible use spark-submit --packages broadcast tensorflow packages across cluster nodes.
i not sure tensorflow itself, can pass local jars using --jars
, files using --files
job. below example:
spark-submit --master yarn-cluster --num-executors 5 --driver-memory 640m --executor-memory 640m --conf spark.yarn.maxappattempts=1000 \ --jars /usr/hdp/current/spark-client-1.6.1/lib/datanucleus-api-jdo-3.2.6.jar,/usr/hdp/current/spark-client-1.6.1/lib/datanucleus-core-3.2.10.jar,/usr/hdp/current/spark-client-1.6.1/lib/datanucleus-rdbms-3.2.9.jar \ --files /usr/hdp/current/spark-client-1.6.1/conf/hive-site.xml \ --class com.foobar.main
this example of how start spark streaming job , application master , driver run on cluster spark not installed. need pass long jars , configs run.
Comments
Post a Comment