You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When working with a workflow using python_wheel_task (built with poetry), modifications of the python code of the package wheel are not deployed when re-deploying the bundle with databricks bundle deploy.
Please list the steps required to reproduce the issue, for example:
Run databricks bundle deploy ...
Run databricks bundle run ...
Modify the source code of the package, like adding a log or a sys.exit()
Run databricks bundle deploy ...
Run databricks bundle run ...
Observe that your modifications are not deployed
Expected Behavior
Since my package is built in the bundle deploy step, the modifications should be included and deployed on the cluster
Actual Behavior
Modifications are not deployed on the existing cluster.
OS and CLI version
WSL Ubuntu 20.04.6 LTS -- Databricks CLI v0.209.1
Is this a regression?
I don't think so
When we are using an existing cluster ID in our bundle is that we are in a development phase, and want to iterate fastly, not waiting at each deploy that a new cluster for our job is deployed, so upgrading the version of our python package is not really an option.
The text was updated successfully, but these errors were encountered:
Hi @FrancoisLem ! This is a limitation on cluster libraries side which requires cluster restart when the wheel is updated for changes to pick it up.
You have 2 options to work this around:
Describe the issue
When working with a workflow using python_wheel_task (built with poetry), modifications of the python code of the package wheel are not deployed when re-deploying the bundle with databricks bundle deploy.
Configuration
`bundle:
name: my_bundle
include:
artifacts:
my-wheel:
type: whl
build: poetry build
and the job
resources:jobs:
my_job:
name: my_workflow
tasks:
################## ETL TASK ##########################
- task_key: "etl_task"
# job_cluster_key: basic-spark-cluster
existing_cluster_id: XXXX-YYYYY-2jtbhpqj
max_retries: 0
python_wheel_task:
package_name: my_package
entry_point: my_entrypoint
parameters:
[
"--config-file-path",
"/Workspace/${workspace.file_path}/conf/tasks/databricks/main_dbx_config.yml",
"--mode", "train"
]
libraries:
- whl: ./dist/my_wheel-.whl
`
Steps to reproduce the behavior
Please list the steps required to reproduce the issue, for example:
databricks bundle deploy ...
databricks bundle run ...
databricks bundle deploy ...
databricks bundle run ...
Expected Behavior
Since my package is built in the bundle deploy step, the modifications should be included and deployed on the cluster
Actual Behavior
Modifications are not deployed on the existing cluster.
OS and CLI version
WSL Ubuntu 20.04.6 LTS -- Databricks CLI v0.209.1
Is this a regression?
I don't think so
When we are using an existing cluster ID in our bundle is that we are in a development phase, and want to iterate fastly, not waiting at each deploy that a new cluster for our job is deployed, so upgrading the version of our python package is not really an option.
The text was updated successfully, but these errors were encountered: