diff --git a/acceptance/bundle/templates/default-python/classic/output/my_default_python/tests/conftest.py b/acceptance/bundle/templates/default-python/classic/output/my_default_python/tests/conftest.py index 72ebfeb566..4ad66de888 100644 --- a/acceptance/bundle/templates/default-python/classic/output/my_default_python/tests/conftest.py +++ b/acceptance/bundle/templates/default-python/classic/output/my_default_python/tests/conftest.py @@ -59,7 +59,11 @@ def _loader(filename: str): def _enable_fallback_compute(): """Enable serverless compute if no compute is specified.""" conf = WorkspaceClient().config - if conf.serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): + # Guard the access with a hasattr for compatibility with older databricks SDK versions. + has_serverless_compute_id = ( + hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + ) + if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): return url = "https://docs.databricks.com/dev-tools/databricks-connect/cluster-config" diff --git a/acceptance/bundle/templates/default-python/serverless-customcatalog/output.txt b/acceptance/bundle/templates/default-python/serverless-customcatalog/output.txt index daa2aba833..0b47631758 100644 --- a/acceptance/bundle/templates/default-python/serverless-customcatalog/output.txt +++ b/acceptance/bundle/templates/default-python/serverless-customcatalog/output.txt @@ -69,3 +69,16 @@ To get started, refer to the project README.md file and the documentation at htt + "defaultValue": "customcatalog", "label": "Catalog", "name": "catalog", +--- [TESTROOT]/bundle/templates/default-python/serverless-customcatalog/../serverless/output/my_default_python/tests/conftest.py ++++ output/my_default_python/tests/conftest.py +@@ -60,5 +60,9 @@ + """Enable serverless compute if no compute is specified.""" + conf = WorkspaceClient().config +- if conf.serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): ++ # Guard the access with a hasattr for compatibility with older databricks SDK versions. ++ has_serverless_compute_id = ( ++ hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id ++ ) ++ if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): + return + diff --git a/acceptance/bundle/templates/default-python/serverless/output/my_default_python/tests/conftest.py b/acceptance/bundle/templates/default-python/serverless/output/my_default_python/tests/conftest.py index 72ebfeb566..4ad66de888 100644 --- a/acceptance/bundle/templates/default-python/serverless/output/my_default_python/tests/conftest.py +++ b/acceptance/bundle/templates/default-python/serverless/output/my_default_python/tests/conftest.py @@ -59,7 +59,11 @@ def _loader(filename: str): def _enable_fallback_compute(): """Enable serverless compute if no compute is specified.""" conf = WorkspaceClient().config - if conf.serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): + # Guard the access with a hasattr for compatibility with older databricks SDK versions. + has_serverless_compute_id = ( + hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + ) + if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): return url = "https://docs.databricks.com/dev-tools/databricks-connect/cluster-config" diff --git a/acceptance/bundle/templates/pydabs/check-formatting/output.txt b/acceptance/bundle/templates/pydabs/check-formatting/output.txt index c9b4edad76..4b3ff774d3 100644 --- a/acceptance/bundle/templates/pydabs/check-formatting/output.txt +++ b/acceptance/bundle/templates/pydabs/check-formatting/output.txt @@ -17,5 +17,19 @@ Checking code formatting with ruff format --line-length 88 Checking code formatting with ruff format --line-length 120 >>> ruff format --isolated --line-length 120 --quiet --diff --check my_pydabs - ->>> yamlcheck.py +--- my_pydabs/tests/conftest.py ++++ my_pydabs/tests/conftest.py +@@ -60,9 +60,7 @@ + """Enable serverless compute if no compute is specified.""" + conf = WorkspaceClient().config + # Guard the access with a hasattr for compatibility with older databricks SDK versions. +- has_serverless_compute_id = ( +- hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id +- ) ++ has_serverless_compute_id = hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): + return + + + +Exit code: 1 diff --git a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py index 72ebfeb566..4ad66de888 100644 --- a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py +++ b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py @@ -59,7 +59,11 @@ def _loader(filename: str): def _enable_fallback_compute(): """Enable serverless compute if no compute is specified.""" conf = WorkspaceClient().config - if conf.serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): + # Guard the access with a hasattr for compatibility with older databricks SDK versions. + has_serverless_compute_id = ( + hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + ) + if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): return url = "https://docs.databricks.com/dev-tools/databricks-connect/cluster-config"