From a6e041a1fe24150c7f2b66e4f5ae2a2185877092 Mon Sep 17 00:00:00 2001 From: timothytamm-db <169097992+timothytamm-db@users.noreply.github.com> Date: Fri, 16 Jan 2026 10:03:03 -0800 Subject: [PATCH 1/5] Update conftest.py --- .../default/template/{{.project_name}}/tests/conftest.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py index 72ebfeb566..23502fb28e 100644 --- a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py +++ b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py @@ -59,7 +59,8 @@ def _loader(filename: str): def _enable_fallback_compute(): """Enable serverless compute if no compute is specified.""" conf = WorkspaceClient().config - if conf.serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): + has_serverles_compute_id = hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + if has_serverles_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): return url = "https://docs.databricks.com/dev-tools/databricks-connect/cluster-config" @@ -83,6 +84,8 @@ def _allow_stderr_output(config: pytest.Config): def pytest_configure(config: pytest.Config): """Configure pytest session.""" with _allow_stderr_output(config): + src_path = pathlib.Path(__file__).parent.parent / "src" + sys.path.insert(0, str(src_path)) _enable_fallback_compute() # Initialize Spark session eagerly, so it is available even when From b9d0fc297bf036a75ae54b6917ae5e781742dcc4 Mon Sep 17 00:00:00 2001 From: timothytamm-db <169097992+timothytamm-db@users.noreply.github.com> Date: Tue, 20 Jan 2026 16:57:59 -0800 Subject: [PATCH 2/5] Update libs/template/templates/default/template/{{.project_name}}/tests/conftest.py Co-authored-by: Lennart Kats (databricks) --- .../default/template/{{.project_name}}/tests/conftest.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py index 23502fb28e..55a84514c5 100644 --- a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py +++ b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py @@ -59,8 +59,8 @@ def _loader(filename: str): def _enable_fallback_compute(): """Enable serverless compute if no compute is specified.""" conf = WorkspaceClient().config - has_serverles_compute_id = hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id - if has_serverles_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): + has_serverless_compute_id = hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): return url = "https://docs.databricks.com/dev-tools/databricks-connect/cluster-config" From 90de0bab611ab9d78c145292619b42ce66a9ead9 Mon Sep 17 00:00:00 2001 From: timothytamm-db <169097992+timothytamm-db@users.noreply.github.com> Date: Tue, 20 Jan 2026 17:00:18 -0800 Subject: [PATCH 3/5] Update conftest.py --- .../default/template/{{.project_name}}/tests/conftest.py | 1 + 1 file changed, 1 insertion(+) diff --git a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py index 55a84514c5..17b0c4a4a1 100644 --- a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py +++ b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py @@ -59,6 +59,7 @@ def _loader(filename: str): def _enable_fallback_compute(): """Enable serverless compute if no compute is specified.""" conf = WorkspaceClient().config + # Guard the access with a hasattr for compatibility with older databricks SDK versions. has_serverless_compute_id = hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): return From 90de912243ce78f66bc5d9490e56dd2f096153c1 Mon Sep 17 00:00:00 2001 From: timothytamm-db <169097992+timothytamm-db@users.noreply.github.com> Date: Tue, 20 Jan 2026 17:43:51 -0800 Subject: [PATCH 4/5] Update conftest.py --- .../default/template/{{.project_name}}/tests/conftest.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py index 17b0c4a4a1..48845c75b3 100644 --- a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py +++ b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py @@ -85,8 +85,6 @@ def _allow_stderr_output(config: pytest.Config): def pytest_configure(config: pytest.Config): """Configure pytest session.""" with _allow_stderr_output(config): - src_path = pathlib.Path(__file__).parent.parent / "src" - sys.path.insert(0, str(src_path)) _enable_fallback_compute() # Initialize Spark session eagerly, so it is available even when From 68132e7109a2e9d6df5efde4d599155731dc087d Mon Sep 17 00:00:00 2001 From: timothytamm-db <169097992+timothytamm-db@users.noreply.github.com> Date: Wed, 21 Jan 2026 09:08:23 -0800 Subject: [PATCH 5/5] Update conftest.py --- .../default/template/{{.project_name}}/tests/conftest.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py index 48845c75b3..4ad66de888 100644 --- a/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py +++ b/libs/template/templates/default/template/{{.project_name}}/tests/conftest.py @@ -60,7 +60,9 @@ def _enable_fallback_compute(): """Enable serverless compute if no compute is specified.""" conf = WorkspaceClient().config # Guard the access with a hasattr for compatibility with older databricks SDK versions. - has_serverless_compute_id = hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + has_serverless_compute_id = ( + hasattr(conf, "serverless_compute_id") and conf.serverless_compute_id + ) if has_serverless_compute_id or conf.cluster_id or os.environ.get("SPARK_REMOTE"): return