Skip to content

Commit 29b7966

Browse files
authored
Move testing to externally configured test envs (#1739)
1 parent c719de2 commit 29b7966

File tree

7 files changed

+17
-30
lines changed

7 files changed

+17
-30
lines changed

clusters/acceptance/cluster_test.go

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ func TestAccClusterResource_CreateClusterWithLibraries(t *testing.T) {
1616
resource "databricks_cluster" "this" {
1717
cluster_name = "libs-{var.RANDOM}"
1818
spark_version = data.databricks_spark_version.latest.id
19-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
19+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
2020
autotermination_minutes = 10
2121
num_workers = 1
2222
spark_conf = {
@@ -64,7 +64,7 @@ func TestAccClusterResource_CreateSingleNodeCluster(t *testing.T) {
6464
resource "databricks_cluster" "this" {
6565
cluster_name = "singlenode-{var.RANDOM}"
6666
spark_version = data.databricks_spark_version.latest.id
67-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
67+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
6868
num_workers = 0
6969
autotermination_minutes = 10
7070
spark_conf = {

clusters/acceptance/clusters_api_test.go

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,6 @@ import (
77

88
"github.com/databricks/terraform-provider-databricks/clusters"
99
"github.com/databricks/terraform-provider-databricks/common"
10-
"github.com/databricks/terraform-provider-databricks/internal/compute"
1110
"github.com/databricks/terraform-provider-databricks/qa"
1211
"github.com/stretchr/testify/assert"
1312
"github.com/stretchr/testify/require"
@@ -27,9 +26,8 @@ func TestAccListClustersIntegration(t *testing.T) {
2726
SparkVersion: clustersAPI.LatestSparkVersionOrDefault(
2827
clusters.SparkVersionRequest{
2928
Latest: true,
30-
LongTermSupport: true,
3129
}),
32-
InstancePoolID: compute.CommonInstancePoolID(),
30+
InstancePoolID: qa.GetEnvOrSkipTest(t, "TEST_INSTANCE_POOL_ID"),
3331
IdempotencyToken: "acc-list-" + randomName,
3432
AutoterminationMinutes: 15,
3533
}
@@ -79,9 +77,8 @@ func TestAccListClustersResizeIntegrationTest(t *testing.T) {
7977
SparkVersion: clustersAPI.LatestSparkVersionOrDefault(
8078
clusters.SparkVersionRequest{
8179
Latest: true,
82-
LongTermSupport: true,
8380
}),
84-
InstancePoolID: compute.CommonInstancePoolID(),
81+
InstancePoolID: qa.GetEnvOrSkipTest(t, "TEST_INSTANCE_POOL_ID"),
8582
IdempotencyToken: "acc-list-" + randomName,
8683
AutoterminationMinutes: 15,
8784
}

internal/compute/common_instances.go

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,10 @@ func CommonInstancePoolID() string {
2424
if commonInstancePool != nil {
2525
return commonInstancePool.InstancePoolID
2626
}
27+
configured := os.Getenv("TEST_INSTANCE_POOL_ID")
28+
if configured != "" {
29+
return configured
30+
}
2731
client := common.CommonEnvironmentClient()
2832
oncePool.Do(func() { // atomic
2933
log.Printf("[INFO] Initializing common instance pool")
@@ -92,7 +96,6 @@ func NewTinyClusterInCommonPool() (c clusters.ClusterInfo, err error) {
9296
ClusterName: "Terraform " + randomName,
9397
SparkVersion: clustersAPI.LatestSparkVersionOrDefault(clusters.SparkVersionRequest{
9498
Latest: true,
95-
LongTermSupport: true,
9699
}),
97100
InstancePoolID: CommonInstancePoolID(),
98101
IdempotencyToken: "tf-" + randomName,
@@ -112,7 +115,6 @@ func NewTinyClusterInCommonPoolPossiblyReused() (c clusters.ClusterInfo) {
112115
ClusterName: currentCluster,
113116
SparkVersion: clustersAPI.LatestSparkVersionOrDefault(clusters.SparkVersionRequest{
114117
Latest: true,
115-
LongTermSupport: true,
116118
}),
117119
InstancePoolID: CommonInstancePoolID(),
118120
IdempotencyToken: "tf-" + randomName,

pipelines/acceptance/pipeline_test.go

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ func TestAccPipelineResource_CreatePipeline(t *testing.T) {
6969
}
7070
7171
cluster {
72-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
72+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
7373
label = "default"
7474
num_workers = 2
7575
custom_tags = {
@@ -78,7 +78,7 @@ func TestAccPipelineResource_CreatePipeline(t *testing.T) {
7878
}
7979
8080
cluster {
81-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
81+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
8282
label = "maintenance"
8383
num_workers = 1
8484
custom_tags = {
@@ -117,7 +117,7 @@ func TestAccAwsPipelineResource_CreatePipeline(t *testing.T) {
117117
}
118118
119119
cluster {
120-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
120+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
121121
label = "default"
122122
num_workers = 2
123123
custom_tags = {
@@ -128,7 +128,7 @@ func TestAccAwsPipelineResource_CreatePipeline(t *testing.T) {
128128
}
129129
}
130130
cluster {
131-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
131+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
132132
label = "maintenance"
133133
num_workers = 1
134134
custom_tags = {
@@ -158,7 +158,7 @@ func TestAccAwsPipelineResource_CreatePipeline(t *testing.T) {
158158
}
159159
160160
cluster {
161-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
161+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
162162
label = "default"
163163
num_workers = 3
164164
custom_tags = {
@@ -169,7 +169,7 @@ func TestAccAwsPipelineResource_CreatePipeline(t *testing.T) {
169169
}
170170
}
171171
cluster {
172-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
172+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
173173
label = "maintenance"
174174
num_workers = 1
175175
custom_tags = {

sql/acceptance/sql_dashboard_test.go

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ func TestAccDashboard(t *testing.T) {
4646
}
4747
4848
resource "databricks_sql_query" "q1" {
49-
data_source_id = databricks_sql_endpoint.this.data_source_id
49+
data_source_id = "{env.TEST_DEFAULT_WAREHOUSE_DATASOURCE_ID}"
5050
name = "tf-{var.RANDOM}-query"
5151
query = "SELECT 1"
5252
}
@@ -58,12 +58,6 @@ func TestAccDashboard(t *testing.T) {
5858
5959
options = jsonencode({})
6060
}
61-
62-
resource "databricks_sql_endpoint" "this" {
63-
name = "tf-{var.RANDOM}-endpoint"
64-
cluster_size = "Small"
65-
max_num_clusters = 1
66-
}
6761
`,
6862
},
6963
})

sql/acceptance/sql_query_test.go

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ func TestAccQuery(t *testing.T) {
1212
{
1313
Template: `
1414
resource "databricks_sql_query" "q1" {
15-
data_source_id = databricks_sql_endpoint.this.data_source_id
15+
data_source_id = "{env.TEST_DEFAULT_WAREHOUSE_DATASOURCE_ID}"
1616
name = "tf-{var.RANDOM}"
1717
query = "SELECT {{ p1 }} AS p1, 2 as p2"
1818
@@ -43,12 +43,6 @@ func TestAccQuery(t *testing.T) {
4343
}
4444
})
4545
}
46-
47-
resource "databricks_sql_endpoint" "this" {
48-
name = "tf-{var.RANDOM}"
49-
cluster_size = "Small"
50-
max_num_clusters = 1
51-
}
5246
`,
5347
},
5448
})

storage/acceptance/aws_s3_mount_test.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ func TestAccAwsS3IamMount_WithCluster(t *testing.T) {
5252
resource "databricks_cluster" "this" {
5353
cluster_name = "ready-{var.RANDOM}"
5454
spark_version = data.databricks_spark_version.latest.id
55-
instance_pool_id = "{var.COMMON_INSTANCE_POOL_ID}"
55+
instance_pool_id = "{env.TEST_INSTANCE_POOL_ID}"
5656
autotermination_minutes = 5
5757
num_workers = 1
5858
aws_attributes {

0 commit comments

Comments
 (0)