Is it possible to broadcast Tensorflow libraries using spark-submit -package -


i using cluster, not managed myself. tensorflow libraries not installed on cluster nodes. run spark programs using tensorflow package. not sure if possible use spark-submit --packages broadcast tensorflow packages across cluster nodes.

i not sure tensorflow itself, can pass local jars using --jars , files using --files job. below example:

spark-submit --master yarn-cluster --num-executors 5 --driver-memory 640m --executor-memory 640m --conf spark.yarn.maxappattempts=1000 \     --jars /usr/hdp/current/spark-client-1.6.1/lib/datanucleus-api-jdo-3.2.6.jar,/usr/hdp/current/spark-client-1.6.1/lib/datanucleus-core-3.2.10.jar,/usr/hdp/current/spark-client-1.6.1/lib/datanucleus-rdbms-3.2.9.jar  \     --files /usr/hdp/current/spark-client-1.6.1/conf/hive-site.xml \     --class com.foobar.main 

this example of how start spark streaming job , application master , driver run on cluster spark not installed. need pass long jars , configs run.


Comments

Popular posts from this blog

java - SSE Emitter : Manage timeouts and complete() -

jquery - uncaught exception: DataTables Editor - remote hosting of code not allowed -

java - How to resolve error - package com.squareup.okhttp3 doesn't exist? -