-
Notifications
You must be signed in to change notification settings - Fork 240
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use Spark's Utils.getContextOrSparkClassLoader
to load Shims [databricks]
#5646
Use Spark's Utils.getContextOrSparkClassLoader
to load Shims [databricks]
#5646
Conversation
Signed-off-by: Gera Shegalov <gera@apache.org>
build |
…Utils.getContextOrSparkClassLoader
current premerge status:
|
…Utils.getContextOrSparkClassLoader
Signed-off-by: Gera Shegalov <gera@apache.org>
build |
Many more tests passed.
|
…Utils.getContextOrSparkClassLoader
Signed-off-by: Gera Shegalov <gera@apache.org>
build |
…Utils.getContextOrSparkClassLoader
Signed-off-by: Gera Shegalov <gera@apache.org>
build |
retrigger workflows |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Loving this.
@@ -1402,14 +1402,6 @@ object RapidsConf { | |||
.booleanConf | |||
.createWithDefault(true) | |||
|
|||
val FORCE_SHIMCALLER_CLASSLOADER = conf("spark.rapids.force.caller.classloader") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Does anything else in spark2 use this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is merely due to spark2 cloning RapidsConf source. We verified there is no actual use of spark.rapids.force.caller.classloader in the spark 2 explain support.
…Utils.getContextOrSparkClassLoader
…Utils.getContextOrSparkClassLoader
Upmerged and removed newer references to force.caller.classloader. Ran iceberg tests locally without issues using the local and standalone modes. |
build |
Utils.getContextOrSparkClassLoader
to load Shims [databricks]
As of #5646 (22.10) specifying `allowConventionalDistJar` is not necessary for the newer JDK's sake. The only legitimate use case is if the user does not want side-effects of dealing with a [multi-shim jar](https://github.com/NVIDIA/spark-rapids/blob/branch-23.08/CONTRIBUTING.md#building-a-distribution-for-a-single-spark-release) ```bash JAVA_HOME=/usr/lib/jvm/java-17-openjdk-amd64 mvn clean install -Dbuildver=341 -DskipTests JAVA_HOME=/usr/lib/jvm/java-17-openjdk-amd64 TEST_PARALLEL=0 SPARK_HOME=~/dist/spark-3.4.1-bin-hadoop3 ./integration_tests/run_pyspark_from_build.sh -s -k array_exists ``` Signed-off-by: Gera Shegalov <gera@apache.org>
Fixes #3851
Spark loads external datasources using
Utils.getContextOrSparkClassLoader
Trampoline to
Utils.getContextOrSparkClassLoader
to make our current code work with external sources, and to unblock JDK9+Signed-off-by: Gera Shegalov gera@apache.org