-
Notifications
You must be signed in to change notification settings - Fork 244
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
[BUG] ZSTD version mismatch in integration tests #10589
Comments
I dug into this a bit, and unexpectedly found that the RAPIDS Accelerator is not using ZSTD during these tests. Dataproc 2.0 is running Spark 3.1.x, so the tests avoid trying to use the ZSTD codec in that case. However Spark itself is trying to use ZSTD for the map statistics during shuffle, and that's what's failing during decode,. The RAPIDS Accelerator shouldn't be involved in that code path at all, especially since the RAPIDS shuffle is not configured for these tests. I tried rolling back to a couple of plugin snapshot versions that were known to pass (one each from 3/10 and 2/28) and they both fail in the same way. I ssh'd to the worker nodes to manually verify the classpath was using the intended jar version and not the new one. |
Looks like this is related to SPARK-35199. The workaround provided is to set |
Tried on 2.0 version,
|
@NvTimLiu as discussed let's update the Dataproc 2.0 integration tests only to use |
Multiple integration tests
test_parquet_append_with_downcast, test_parquet_write_column_name_with_dots
etc failed with the following error:The text was updated successfully, but these errors were encountered: