-
Notifications
You must be signed in to change notification settings - Fork 118
Error: Only local python files are supported: gs://... #527
Comments
I should note that for the GCS connector, I also had to add some runtime config files (notably I also had to add |
cc @liyinan926 |
This should be fixed by adding |
BTW: @paulreimer I found that instead of baking in |
Interesting, it took me so long to figure out to add One nice thing about baking it in though, is the |
(I was using only GCE resources, and so allowing "application default credentials" to Just Work, instead of manually specifying service accounts.) |
@paulreimer Yes, you can use |
Sounds good, I will need something like that for the non-GCE clusters. I was unable to build a working distribution with the My build also fails for Scala jobs that worked before with my image with the GCS container added (using the 0.4.0 release jars), so something must be wrong with my build environment (I have never built before). I used I would be happy to test updated binaries from a working build, with the |
Try this build command |
Right on, that command worked for me, and the suggested change also worked! I was able to successfully submit my python job, using Note, I only had to replace the Thanks so much, I really appreciate your help, @liyinan926 ! |
Cool! Can you submit a PR with the change? Thanks! |
… files) when isKubernetes is set (apache-spark-on-k8s#527)
I extended the docker image using the recent
spark-2.2.0-k8s-0.4.0-bin-2.7.3
release to add the GCS (Google Cloud Storage) connector.Observed:
It works great for scala jobs / jars with a
gs://<bucket>/
prefix - I see it creates the init container and does populate thespark-files
from what was already in GCS. However, when I try to submit a python job (or use--py-files
), thespark-submit
client does not allow thegs://
prefix and refuses the job.Expected:
The job to be allowed by
spark-submit
, the relevant files populated in an initcontainer, and available for thespark-driver-py
andspark-executor-py
to use successfully.(FYI To add the GCS connector, I added these lines to
spark-base
Dockerfile:)The text was updated successfully, but these errors were encountered: