-
Notifications
You must be signed in to change notification settings - Fork 14.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding Support for Yarn queue and other extras in SparkSubmit Operator and Hook #35911
Comments
Please assign this to me, this is a quick fix |
@pateash assigned you. |
So do we have a task to complete here? |
@eladkal, |
We normally allow to override connection extra from the operator level. For example with Snowflake you can set different warehouse than the default one set in the connection. |
Yes, it's done using args AFAIK, |
I think the update caused the new issue since SparkSubmitOperator accidentally override the Base Operator's parameter.
https://github.com/apache/airflow/blob/main/airflow/models/baseoperator.py#L587
So if you are using CeleryKubernetesExecutor, you cannot decide the Operator run on KubernetesExecutor by setting the |
Since 4.6.0 where this was introduced, onwards, using SparkSubmitOperator is broken for us, as the queue can no longer be taken into account due to this change. Before, this |
Description
Spark-submit
--queue thequeue option specifies the YARN queue to which the application should be submitted.
more - https://spark.apache.org/docs/3.2.0/running-on-yarn.html
Use case/motivation
The --queue option is particularly useful in a multi-tenant environment where different users or groups have allocated resources in specific YARN queues.
Related issues
No response
Are you willing to submit a PR?
Code of Conduct
The text was updated successfully, but these errors were encountered: