-
Notifications
You must be signed in to change notification settings - Fork 242
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] DatabricksShimVersion must carry runtime version info #3532
Comments
note that may work if running an actual notebook but we may need to figure out something for testing scala> spark.conf.get("spark.databricks.clusterUsageTags.sparkVersion") |
We can grab it at build time and store as db-version.properties in under db shim dir:
For shims to work correctly, we have to have shims per runtime instead of per spark version to the tune of: At run time. |
How does grabbing I'm also not a fan of putting the Spark version in the Databricks shim version. They could change it at a whim (and have in the past, from 3.1.0 to 3.1.1). All that really matters is the Databricks runtime version, and that's what the Databricks user is more familiar with. |
I can be more explicit about the algorithm but hopefully it was clear where I am going with it:
This is the point of this bug, the databricks ShimServiceProviders should compare the build-time and run-time |
Describe the bug
Shim layer in the Plugin currently identifies Databricks runtime versions using the exposed Spark version only. However, this is not a 1:1 mapping https://docs.databricks.com/release-notes/runtime/releases.html.
Reviewing Spark WebUI in the runtime under "Environment" one can see a config key that we can use to retrieve the runtime version info without creating a binary dependency in Scala using
spark.conf.get("spark.databricks.clusterUsageTags.sparkVersion")
Steps/Code to reproduce bug
https://nvidia.github.io/spark-rapids/docs/get-started/getting-started-databricks.html
Expected behavior
Shim layer should correctly identify the runtime version to handle Spark differences in runtime versions.
Environment details (please complete the following information)
The text was updated successfully, but these errors were encountered: