From ecade681792f26880370772b0966b3f63e07b425 Mon Sep 17 00:00:00 2001 From: Pulumi Bot <30351955+pulumi-bot@users.noreply.github.com> Date: Fri, 31 May 2024 10:12:35 -0700 Subject: [PATCH] Upgrade terraform-provider-databricks to v1.46.0 (#484) This PR was generated via `$ upgrade-provider pulumi/pulumi-databricks --kind=all --target-bridge-version=latest`. --- - Upgrading terraform-provider-databricks from 1.45.0 to 1.46.0. Fixes #483 --- .../pulumi-resource-databricks/schema.json | 131 ++- provider/go.mod | 2 +- provider/go.sum | 4 +- .../GetJobJobSettingsSettingsHealthRule.cs | 12 +- ...GetJobJobSettingsSettingsHealthRuleArgs.cs | 12 +- .../GetJobJobSettingsSettingsJobCluster.cs | 8 +- ...GetJobJobSettingsSettingsJobClusterArgs.cs | 8 +- .../Inputs/GetJobJobSettingsSettingsTask.cs | 4 +- .../GetJobJobSettingsSettingsTaskArgs.cs | 4 +- ...bJobSettingsSettingsTaskForEachTaskTask.cs | 4 +- ...SettingsSettingsTaskForEachTaskTaskArgs.cs | 4 +- ...gsSettingsTaskForEachTaskTaskHealthRule.cs | 12 +- ...ttingsTaskForEachTaskTaskHealthRuleArgs.cs | 12 +- ...tingsSettingsTaskForEachTaskTaskSqlTask.cs | 4 +- ...sSettingsTaskForEachTaskTaskSqlTaskArgs.cs | 4 +- ...GetJobJobSettingsSettingsTaskHealthRule.cs | 12 +- ...obJobSettingsSettingsTaskHealthRuleArgs.cs | 12 +- .../GetJobJobSettingsSettingsTaskSqlTask.cs | 4 +- ...etJobJobSettingsSettingsTaskSqlTaskArgs.cs | 4 +- sdk/dotnet/Inputs/JobHealthRuleArgs.cs | 12 +- sdk/dotnet/Inputs/JobHealthRuleGetArgs.cs | 12 +- sdk/dotnet/Inputs/JobJobClusterArgs.cs | 8 +- sdk/dotnet/Inputs/JobJobClusterGetArgs.cs | 8 +- .../Inputs/JobJobClusterNewClusterArgs.cs | 3 - .../Inputs/JobJobClusterNewClusterGetArgs.cs | 3 - sdk/dotnet/Inputs/JobNewClusterArgs.cs | 3 - sdk/dotnet/Inputs/JobNewClusterGetArgs.cs | 3 - sdk/dotnet/Inputs/JobTaskArgs.cs | 4 +- .../Inputs/JobTaskForEachTaskTaskArgs.cs | 4 +- .../Inputs/JobTaskForEachTaskTaskGetArgs.cs | 4 +- .../JobTaskForEachTaskTaskHealthRuleArgs.cs | 12 +- ...JobTaskForEachTaskTaskHealthRuleGetArgs.cs | 12 +- .../JobTaskForEachTaskTaskNewClusterArgs.cs | 3 - ...JobTaskForEachTaskTaskNewClusterGetArgs.cs | 3 - .../JobTaskForEachTaskTaskSqlTaskArgs.cs | 4 +- .../JobTaskForEachTaskTaskSqlTaskGetArgs.cs | 4 +- sdk/dotnet/Inputs/JobTaskGetArgs.cs | 4 +- sdk/dotnet/Inputs/JobTaskHealthRuleArgs.cs | 12 +- sdk/dotnet/Inputs/JobTaskHealthRuleGetArgs.cs | 12 +- sdk/dotnet/Inputs/JobTaskNewClusterArgs.cs | 3 - sdk/dotnet/Inputs/JobTaskNewClusterGetArgs.cs | 3 - sdk/dotnet/Inputs/JobTaskSqlTaskArgs.cs | 4 +- sdk/dotnet/Inputs/JobTaskSqlTaskGetArgs.cs | 4 +- sdk/dotnet/Inputs/JobTriggerArgs.cs | 3 - sdk/dotnet/Inputs/JobTriggerGetArgs.cs | 3 - .../Inputs/JobTriggerTableUpdateArgs.cs | 13 - .../Inputs/JobTriggerTableUpdateGetArgs.cs | 13 - ...tJobJobSettingsSettingsHealthRuleResult.cs | 12 +- ...tJobJobSettingsSettingsJobClusterResult.cs | 8 +- ...ingsTaskForEachTaskTaskHealthRuleResult.cs | 12 +- ...ttingsSettingsTaskForEachTaskTaskResult.cs | 4 +- ...ettingsTaskForEachTaskTaskSqlTaskResult.cs | 4 +- ...JobSettingsSettingsTaskHealthRuleResult.cs | 12 +- .../GetJobJobSettingsSettingsTaskResult.cs | 4 +- ...JobJobSettingsSettingsTaskSqlTaskResult.cs | 4 +- sdk/dotnet/Outputs/JobHealthRule.cs | 12 +- sdk/dotnet/Outputs/JobJobCluster.cs | 8 +- sdk/dotnet/Outputs/JobJobClusterNewCluster.cs | 4 - sdk/dotnet/Outputs/JobNewCluster.cs | 4 - sdk/dotnet/Outputs/JobTask.cs | 4 +- sdk/dotnet/Outputs/JobTaskForEachTaskTask.cs | 4 +- .../JobTaskForEachTaskTaskHealthRule.cs | 12 +- .../JobTaskForEachTaskTaskNewCluster.cs | 4 - .../Outputs/JobTaskForEachTaskTaskSqlTask.cs | 4 +- sdk/dotnet/Outputs/JobTaskHealthRule.cs | 12 +- sdk/dotnet/Outputs/JobTaskNewCluster.cs | 4 - sdk/dotnet/Outputs/JobTaskSqlTask.cs | 4 +- sdk/dotnet/Outputs/JobTrigger.cs | 3 - sdk/dotnet/Outputs/JobTriggerTableUpdate.cs | 12 - sdk/go/databricks/pulumiTypes.go | 967 ++---------------- sdk/go/databricks/pulumiTypes1.go | 86 +- .../GetJobJobSettingsSettingsHealthRule.java | 42 +- ...tJobJobSettingsSettingsHealthRuleArgs.java | 42 +- .../GetJobJobSettingsSettingsJobCluster.java | 29 +- ...tJobJobSettingsSettingsJobClusterArgs.java | 29 +- .../inputs/GetJobJobSettingsSettingsTask.java | 13 +- .../GetJobJobSettingsSettingsTaskArgs.java | 13 +- ...obSettingsSettingsTaskForEachTaskTask.java | 13 +- ...ttingsSettingsTaskForEachTaskTaskArgs.java | 13 +- ...SettingsTaskForEachTaskTaskHealthRule.java | 42 +- ...ingsTaskForEachTaskTaskHealthRuleArgs.java | 42 +- ...ngsSettingsTaskForEachTaskTaskSqlTask.java | 14 +- ...ettingsTaskForEachTaskTaskSqlTaskArgs.java | 14 +- ...tJobJobSettingsSettingsTaskHealthRule.java | 42 +- ...JobSettingsSettingsTaskHealthRuleArgs.java | 42 +- .../GetJobJobSettingsSettingsTaskSqlTask.java | 14 +- ...JobJobSettingsSettingsTaskSqlTaskArgs.java | 14 +- .../databricks/inputs/JobHealthRuleArgs.java | 42 +- .../databricks/inputs/JobJobClusterArgs.java | 29 +- .../inputs/JobJobClusterNewClusterArgs.java | 17 - .../databricks/inputs/JobNewClusterArgs.java | 17 - .../pulumi/databricks/inputs/JobTaskArgs.java | 14 +- .../inputs/JobTaskForEachTaskTaskArgs.java | 14 +- .../JobTaskForEachTaskTaskHealthRuleArgs.java | 42 +- .../JobTaskForEachTaskTaskNewClusterArgs.java | 17 - .../JobTaskForEachTaskTaskSqlTaskArgs.java | 14 +- .../inputs/JobTaskHealthRuleArgs.java | 42 +- .../inputs/JobTaskNewClusterArgs.java | 17 - .../databricks/inputs/JobTaskSqlTaskArgs.java | 14 +- .../databricks/inputs/JobTriggerArgs.java | 20 - .../inputs/JobTriggerTableUpdateArgs.java | 86 -- .../GetJobJobSettingsSettingsHealthRule.java | 45 +- .../GetJobJobSettingsSettingsJobCluster.java | 31 +- .../GetJobJobSettingsSettingsTask.java | 14 +- ...obSettingsSettingsTaskForEachTaskTask.java | 14 +- ...SettingsTaskForEachTaskTaskHealthRule.java | 45 +- ...ngsSettingsTaskForEachTaskTaskSqlTask.java | 15 +- ...tJobJobSettingsSettingsTaskHealthRule.java | 45 +- .../GetJobJobSettingsSettingsTaskSqlTask.java | 15 +- .../databricks/outputs/JobHealthRule.java | 45 +- .../databricks/outputs/JobJobCluster.java | 31 +- .../outputs/JobJobClusterNewCluster.java | 13 - .../databricks/outputs/JobNewCluster.java | 13 - .../pulumi/databricks/outputs/JobTask.java | 15 +- .../outputs/JobTaskForEachTaskTask.java | 15 +- .../JobTaskForEachTaskTaskHealthRule.java | 45 +- .../JobTaskForEachTaskTaskNewCluster.java | 13 - .../JobTaskForEachTaskTaskSqlTask.java | 15 +- .../databricks/outputs/JobTaskHealthRule.java | 45 +- .../databricks/outputs/JobTaskNewCluster.java | 13 - .../databricks/outputs/JobTaskSqlTask.java | 15 +- .../pulumi/databricks/outputs/JobTrigger.java | 8 - .../outputs/JobTriggerTableUpdate.java | 32 - sdk/nodejs/types/input.ts | 109 +- sdk/nodejs/types/output.ts | 83 +- sdk/python/pulumi_databricks/_inputs.py | 500 ++++----- sdk/python/pulumi_databricks/outputs.py | 396 +++---- 127 files changed, 1475 insertions(+), 2595 deletions(-) diff --git a/provider/cmd/pulumi-resource-databricks/schema.json b/provider/cmd/pulumi-resource-databricks/schema.json index 2955d91e..622294f9 100644 --- a/provider/cmd/pulumi-resource-databricks/schema.json +++ b/provider/cmd/pulumi-resource-databricks/schema.json @@ -1285,7 +1285,12 @@ "description": "integer value used to compare to the given metric.\n" } }, - "type": "object" + "type": "object", + "required": [ + "metric", + "op", + "value" + ] }, "databricks:index/JobJobCluster:JobJobCluster": { "properties": { @@ -1298,7 +1303,11 @@ "description": "Same set of parameters as for databricks.Cluster resource.\n" } }, - "type": "object" + "type": "object", + "required": [ + "jobClusterKey", + "newCluster" + ] }, "databricks:index/JobJobClusterNewCluster:JobJobClusterNewCluster": { "properties": { @@ -1308,9 +1317,6 @@ "autoscale": { "$ref": "#/types/databricks:index/JobJobClusterNewClusterAutoscale:JobJobClusterNewClusterAutoscale" }, - "autoterminationMinutes": { - "type": "integer" - }, "awsAttributes": { "$ref": "#/types/databricks:index/JobJobClusterNewClusterAwsAttributes:JobJobClusterNewClusterAwsAttributes" }, @@ -1947,9 +1953,6 @@ "autoscale": { "$ref": "#/types/databricks:index/JobNewClusterAutoscale:JobNewClusterAutoscale" }, - "autoterminationMinutes": { - "type": "integer" - }, "awsAttributes": { "$ref": "#/types/databricks:index/JobNewClusterAwsAttributes:JobNewClusterAwsAttributes" }, @@ -2836,10 +2839,14 @@ } }, "type": "object", + "required": [ + "taskKey" + ], "language": { "nodejs": { "requiredOutputs": [ - "retryOnTimeout" + "retryOnTimeout", + "taskKey" ] } } @@ -3088,10 +3095,14 @@ } }, "type": "object", + "required": [ + "taskKey" + ], "language": { "nodejs": { "requiredOutputs": [ - "retryOnTimeout" + "retryOnTimeout", + "taskKey" ] } } @@ -3239,7 +3250,12 @@ "description": "integer value used to compare to the given metric.\n" } }, - "type": "object" + "type": "object", + "required": [ + "metric", + "op", + "value" + ] }, "databricks:index/JobTaskForEachTaskTaskLibrary:JobTaskForEachTaskTaskLibrary": { "properties": { @@ -3323,9 +3339,6 @@ "autoscale": { "$ref": "#/types/databricks:index/JobTaskForEachTaskTaskNewClusterAutoscale:JobTaskForEachTaskTaskNewClusterAutoscale" }, - "autoterminationMinutes": { - "type": "integer" - }, "awsAttributes": { "$ref": "#/types/databricks:index/JobTaskForEachTaskTaskNewClusterAwsAttributes:JobTaskForEachTaskTaskNewClusterAwsAttributes" }, @@ -4123,7 +4136,10 @@ "description": "ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless \u0026 Pro warehouses are supported right now.\n" } }, - "type": "object" + "type": "object", + "required": [ + "warehouseId" + ] }, "databricks:index/JobTaskForEachTaskTaskSqlTaskAlert:JobTaskForEachTaskTaskSqlTaskAlert": { "properties": { @@ -4336,7 +4352,12 @@ "description": "integer value used to compare to the given metric.\n" } }, - "type": "object" + "type": "object", + "required": [ + "metric", + "op", + "value" + ] }, "databricks:index/JobTaskLibrary:JobTaskLibrary": { "properties": { @@ -4420,9 +4441,6 @@ "autoscale": { "$ref": "#/types/databricks:index/JobTaskNewClusterAutoscale:JobTaskNewClusterAutoscale" }, - "autoterminationMinutes": { - "type": "integer" - }, "awsAttributes": { "$ref": "#/types/databricks:index/JobTaskNewClusterAwsAttributes:JobTaskNewClusterAwsAttributes" }, @@ -5220,7 +5238,10 @@ "description": "ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless \u0026 Pro warehouses are supported right now.\n" } }, - "type": "object" + "type": "object", + "required": [ + "warehouseId" + ] }, "databricks:index/JobTaskSqlTaskAlert:JobTaskSqlTaskAlert": { "properties": { @@ -5417,8 +5438,7 @@ "$ref": "#/types/databricks:index/JobTriggerTable:JobTriggerTable" }, "tableUpdate": { - "$ref": "#/types/databricks:index/JobTriggerTableUpdate:JobTriggerTableUpdate", - "description": "configuration block to define a trigger for Table Update events consisting of following attributes:\n" + "$ref": "#/types/databricks:index/JobTriggerTableUpdate:JobTriggerTableUpdate" } }, "type": "object" @@ -5466,23 +5486,19 @@ "databricks:index/JobTriggerTableUpdate:JobTriggerTableUpdate": { "properties": { "condition": { - "type": "string", - "description": "The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`.\n" + "type": "string" }, "minTimeBetweenTriggersSeconds": { - "type": "integer", - "description": "If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds.\n" + "type": "integer" }, "tableNames": { "type": "array", "items": { "type": "string" - }, - "description": "A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`.\n" + } }, "waitAfterLastChangeSeconds": { - "type": "integer", - "description": "If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds.\n" + "type": "integer" } }, "type": "object", @@ -6078,7 +6094,9 @@ "language": { "nodejs": { "requiredOutputs": [ - "name" + "name", + "workloadSize", + "workloadType" ] } } @@ -10371,7 +10389,12 @@ "type": "integer" } }, - "type": "object" + "type": "object", + "required": [ + "metric", + "op", + "value" + ] }, "databricks:index/getJobJobSettingsSettingsJobCluster:getJobJobSettingsSettingsJobCluster": { "properties": { @@ -10382,7 +10405,11 @@ "$ref": "#/types/databricks:index/getJobJobSettingsSettingsJobClusterNewCluster:getJobJobSettingsSettingsJobClusterNewCluster" } }, - "type": "object" + "type": "object", + "required": [ + "jobClusterKey", + "newCluster" + ] }, "databricks:index/getJobJobSettingsSettingsJobClusterNewCluster:getJobJobSettingsSettingsJobClusterNewCluster": { "properties": { @@ -11671,11 +11698,14 @@ }, "type": "object", "required": [ - "retryOnTimeout" + "retryOnTimeout", + "taskKey" ], "language": { "nodejs": { - "requiredInputs": [] + "requiredInputs": [ + "taskKey" + ] } } }, @@ -11886,11 +11916,14 @@ }, "type": "object", "required": [ - "retryOnTimeout" + "retryOnTimeout", + "taskKey" ], "language": { "nodejs": { - "requiredInputs": [] + "requiredInputs": [ + "taskKey" + ] } } }, @@ -12017,7 +12050,12 @@ "type": "integer" } }, - "type": "object" + "type": "object", + "required": [ + "metric", + "op", + "value" + ] }, "databricks:index/getJobJobSettingsSettingsTaskForEachTaskTaskLibrary:getJobJobSettingsSettingsTaskForEachTaskTaskLibrary": { "properties": { @@ -12719,7 +12757,10 @@ "type": "string" } }, - "type": "object" + "type": "object", + "required": [ + "warehouseId" + ] }, "databricks:index/getJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlert:getJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlert": { "properties": { @@ -12915,7 +12956,12 @@ "type": "integer" } }, - "type": "object" + "type": "object", + "required": [ + "metric", + "op", + "value" + ] }, "databricks:index/getJobJobSettingsSettingsTaskLibrary:getJobJobSettingsSettingsTaskLibrary": { "properties": { @@ -13617,7 +13663,10 @@ "type": "string" } }, - "type": "object" + "type": "object", + "required": [ + "warehouseId" + ] }, "databricks:index/getJobJobSettingsSettingsTaskSqlTaskAlert:getJobJobSettingsSettingsTaskSqlTaskAlert": { "properties": { diff --git a/provider/go.mod b/provider/go.mod index cc359d44..cba28981 100644 --- a/provider/go.mod +++ b/provider/go.mod @@ -6,7 +6,7 @@ replace github.com/hashicorp/terraform-plugin-sdk/v2 => github.com/pulumi/terraf require ( github.com/databricks/databricks-sdk-go v0.41.0 - github.com/databricks/terraform-provider-databricks v1.45.0 + github.com/databricks/terraform-provider-databricks v1.46.0 github.com/pulumi/pulumi-terraform-bridge/v3 v3.83.0 ) diff --git a/provider/go.sum b/provider/go.sum index 242aea06..3dc2ddaa 100644 --- a/provider/go.sum +++ b/provider/go.sum @@ -1658,8 +1658,8 @@ github.com/d2g/dhcp4server v0.0.0-20181031114812-7d4a0a7f59a5/go.mod h1:Eo87+Kg/ github.com/d2g/hardwareaddr v0.0.0-20190221164911-e7d9fbe030e4/go.mod h1:bMl4RjIciD2oAxI7DmWRx6gbeqrkoLqv3MV0vzNad+I= github.com/databricks/databricks-sdk-go v0.41.0 h1:OyhYY+Q6+gqkWeXmpGEiacoU2RStTeWPF0x4vmqbQdc= github.com/databricks/databricks-sdk-go v0.41.0/go.mod h1:rLIhh7DvifVLmf2QxMr/vMRGqdrTZazn8VYo4LilfCo= -github.com/databricks/terraform-provider-databricks v1.45.0 h1:W5+8n0sDdrfXs884jIEaYf8TOPw0BoDiL4AC457ClSo= -github.com/databricks/terraform-provider-databricks v1.45.0/go.mod h1:D1A1+fQQRGWGcM+a5GvPGm0PWXGRzStsFZXyWITznVc= +github.com/databricks/terraform-provider-databricks v1.46.0 h1:nT23H6OR4fXnSGIGi67bId2TKuXPaa9qnB9I/U6d/tc= +github.com/databricks/terraform-provider-databricks v1.46.0/go.mod h1:D1A1+fQQRGWGcM+a5GvPGm0PWXGRzStsFZXyWITznVc= github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c= github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRule.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRule.cs index a8ebbc80..b09cc86d 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRule.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRule.cs @@ -12,14 +12,14 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsHealthRuleArgs : global::Pulumi.InvokeArgs { - [Input("metric")] - public string? Metric { get; set; } + [Input("metric", required: true)] + public string Metric { get; set; } = null!; - [Input("op")] - public string? Op { get; set; } + [Input("op", required: true)] + public string Op { get; set; } = null!; - [Input("value")] - public int? Value { get; set; } + [Input("value", required: true)] + public int Value { get; set; } public GetJobJobSettingsSettingsHealthRuleArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRuleArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRuleArgs.cs index e947c61e..f0fe4683 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRuleArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsHealthRuleArgs.cs @@ -12,14 +12,14 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsHealthRuleInputArgs : global::Pulumi.ResourceArgs { - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public GetJobJobSettingsSettingsHealthRuleInputArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobCluster.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobCluster.cs index 0fb6759d..be8e3246 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobCluster.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobCluster.cs @@ -12,11 +12,11 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsJobClusterArgs : global::Pulumi.InvokeArgs { - [Input("jobClusterKey")] - public string? JobClusterKey { get; set; } + [Input("jobClusterKey", required: true)] + public string JobClusterKey { get; set; } = null!; - [Input("newCluster")] - public Inputs.GetJobJobSettingsSettingsJobClusterNewClusterArgs? NewCluster { get; set; } + [Input("newCluster", required: true)] + public Inputs.GetJobJobSettingsSettingsJobClusterNewClusterArgs NewCluster { get; set; } = null!; public GetJobJobSettingsSettingsJobClusterArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobClusterArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobClusterArgs.cs index e0d3222e..ea0771af 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobClusterArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsJobClusterArgs.cs @@ -12,11 +12,11 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsJobClusterInputArgs : global::Pulumi.ResourceArgs { - [Input("jobClusterKey")] - public Input? JobClusterKey { get; set; } + [Input("jobClusterKey", required: true)] + public Input JobClusterKey { get; set; } = null!; - [Input("newCluster")] - public Input? NewCluster { get; set; } + [Input("newCluster", required: true)] + public Input NewCluster { get; set; } = null!; public GetJobJobSettingsSettingsJobClusterInputArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTask.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTask.cs index 49a49419..dbedd3e8 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTask.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTask.cs @@ -97,8 +97,8 @@ public List Libraries [Input("sqlTask")] public Inputs.GetJobJobSettingsSettingsTaskSqlTaskArgs? SqlTask { get; set; } - [Input("taskKey")] - public string? TaskKey { get; set; } + [Input("taskKey", required: true)] + public string TaskKey { get; set; } = null!; [Input("timeoutSeconds")] public int? TimeoutSeconds { get; set; } diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskArgs.cs index e27ffdfd..554fe28e 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskArgs.cs @@ -97,8 +97,8 @@ public InputList Libraries [Input("sqlTask")] public Input? SqlTask { get; set; } - [Input("taskKey")] - public Input? TaskKey { get; set; } + [Input("taskKey", required: true)] + public Input TaskKey { get; set; } = null!; [Input("timeoutSeconds")] public Input? TimeoutSeconds { get; set; } diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.cs index b09f03de..74cbe331 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.cs @@ -94,8 +94,8 @@ public List Libr [Input("sqlTask")] public Inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs? SqlTask { get; set; } - [Input("taskKey")] - public string? TaskKey { get; set; } + [Input("taskKey", required: true)] + public string TaskKey { get; set; } = null!; [Input("timeoutSeconds")] public int? TimeoutSeconds { get; set; } diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.cs index 990f6261..a797f710 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.cs @@ -94,8 +94,8 @@ public InputList? SqlTask { get; set; } - [Input("taskKey")] - public Input? TaskKey { get; set; } + [Input("taskKey", required: true)] + public Input TaskKey { get; set; } = null!; [Input("timeoutSeconds")] public Input? TimeoutSeconds { get; set; } diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.cs index d88f956b..e40d6401 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.cs @@ -12,14 +12,14 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs : global::Pulumi.InvokeArgs { - [Input("metric")] - public string? Metric { get; set; } + [Input("metric", required: true)] + public string Metric { get; set; } = null!; - [Input("op")] - public string? Op { get; set; } + [Input("op", required: true)] + public string Op { get; set; } = null!; - [Input("value")] - public int? Value { get; set; } + [Input("value", required: true)] + public int Value { get; set; } public GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.cs index f4925d2d..2fd49f15 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.cs @@ -12,14 +12,14 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleInputArgs : global::Pulumi.ResourceArgs { - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleInputArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.cs index 5252de86..eacd37e2 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.cs @@ -32,8 +32,8 @@ public Dictionary Parameters [Input("query")] public Inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryArgs? Query { get; set; } - [Input("warehouseId")] - public string? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public string WarehouseId { get; set; } = null!; public GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.cs index 98f3a6e7..eb39a7ab 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.cs @@ -32,8 +32,8 @@ public InputMap Parameters [Input("query")] public Input? Query { get; set; } - [Input("warehouseId")] - public Input? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public Input WarehouseId { get; set; } = null!; public GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskInputArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRule.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRule.cs index ab27ee59..63709d83 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRule.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRule.cs @@ -12,14 +12,14 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsTaskHealthRuleArgs : global::Pulumi.InvokeArgs { - [Input("metric")] - public string? Metric { get; set; } + [Input("metric", required: true)] + public string Metric { get; set; } = null!; - [Input("op")] - public string? Op { get; set; } + [Input("op", required: true)] + public string Op { get; set; } = null!; - [Input("value")] - public int? Value { get; set; } + [Input("value", required: true)] + public int Value { get; set; } public GetJobJobSettingsSettingsTaskHealthRuleArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.cs index f128f590..5870ae0a 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.cs @@ -12,14 +12,14 @@ namespace Pulumi.Databricks.Inputs public sealed class GetJobJobSettingsSettingsTaskHealthRuleInputArgs : global::Pulumi.ResourceArgs { - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public GetJobJobSettingsSettingsTaskHealthRuleInputArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTask.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTask.cs index cd845f54..c7c67c7a 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTask.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTask.cs @@ -32,8 +32,8 @@ public Dictionary Parameters [Input("query")] public Inputs.GetJobJobSettingsSettingsTaskSqlTaskQueryArgs? Query { get; set; } - [Input("warehouseId")] - public string? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public string WarehouseId { get; set; } = null!; public GetJobJobSettingsSettingsTaskSqlTaskArgs() { diff --git a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.cs b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.cs index f3b7ab31..b757ea2f 100644 --- a/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.cs +++ b/sdk/dotnet/Inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.cs @@ -32,8 +32,8 @@ public InputMap Parameters [Input("query")] public Input? Query { get; set; } - [Input("warehouseId")] - public Input? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public Input WarehouseId { get; set; } = null!; public GetJobJobSettingsSettingsTaskSqlTaskInputArgs() { diff --git a/sdk/dotnet/Inputs/JobHealthRuleArgs.cs b/sdk/dotnet/Inputs/JobHealthRuleArgs.cs index cb472d85..0abd925d 100644 --- a/sdk/dotnet/Inputs/JobHealthRuleArgs.cs +++ b/sdk/dotnet/Inputs/JobHealthRuleArgs.cs @@ -15,20 +15,20 @@ public sealed class JobHealthRuleArgs : global::Pulumi.ResourceArgs /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; /// /// integer value used to compare to the given metric. /// - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public JobHealthRuleArgs() { diff --git a/sdk/dotnet/Inputs/JobHealthRuleGetArgs.cs b/sdk/dotnet/Inputs/JobHealthRuleGetArgs.cs index 8dee139c..b773040f 100644 --- a/sdk/dotnet/Inputs/JobHealthRuleGetArgs.cs +++ b/sdk/dotnet/Inputs/JobHealthRuleGetArgs.cs @@ -15,20 +15,20 @@ public sealed class JobHealthRuleGetArgs : global::Pulumi.ResourceArgs /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; /// /// integer value used to compare to the given metric. /// - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public JobHealthRuleGetArgs() { diff --git a/sdk/dotnet/Inputs/JobJobClusterArgs.cs b/sdk/dotnet/Inputs/JobJobClusterArgs.cs index cb189ef4..fb02a439 100644 --- a/sdk/dotnet/Inputs/JobJobClusterArgs.cs +++ b/sdk/dotnet/Inputs/JobJobClusterArgs.cs @@ -15,14 +15,14 @@ public sealed class JobJobClusterArgs : global::Pulumi.ResourceArgs /// /// Identifier that can be referenced in `task` block, so that cluster is shared between tasks /// - [Input("jobClusterKey")] - public Input? JobClusterKey { get; set; } + [Input("jobClusterKey", required: true)] + public Input JobClusterKey { get; set; } = null!; /// /// Same set of parameters as for databricks.Cluster resource. /// - [Input("newCluster")] - public Input? NewCluster { get; set; } + [Input("newCluster", required: true)] + public Input NewCluster { get; set; } = null!; public JobJobClusterArgs() { diff --git a/sdk/dotnet/Inputs/JobJobClusterGetArgs.cs b/sdk/dotnet/Inputs/JobJobClusterGetArgs.cs index 0b3fd699..59cdfe4a 100644 --- a/sdk/dotnet/Inputs/JobJobClusterGetArgs.cs +++ b/sdk/dotnet/Inputs/JobJobClusterGetArgs.cs @@ -15,14 +15,14 @@ public sealed class JobJobClusterGetArgs : global::Pulumi.ResourceArgs /// /// Identifier that can be referenced in `task` block, so that cluster is shared between tasks /// - [Input("jobClusterKey")] - public Input? JobClusterKey { get; set; } + [Input("jobClusterKey", required: true)] + public Input JobClusterKey { get; set; } = null!; /// /// Same set of parameters as for databricks.Cluster resource. /// - [Input("newCluster")] - public Input? NewCluster { get; set; } + [Input("newCluster", required: true)] + public Input NewCluster { get; set; } = null!; public JobJobClusterGetArgs() { diff --git a/sdk/dotnet/Inputs/JobJobClusterNewClusterArgs.cs b/sdk/dotnet/Inputs/JobJobClusterNewClusterArgs.cs index 8239da1b..204155ff 100644 --- a/sdk/dotnet/Inputs/JobJobClusterNewClusterArgs.cs +++ b/sdk/dotnet/Inputs/JobJobClusterNewClusterArgs.cs @@ -18,9 +18,6 @@ public sealed class JobJobClusterNewClusterArgs : global::Pulumi.ResourceArgs [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobJobClusterNewClusterGetArgs.cs b/sdk/dotnet/Inputs/JobJobClusterNewClusterGetArgs.cs index 198699b1..eb535b64 100644 --- a/sdk/dotnet/Inputs/JobJobClusterNewClusterGetArgs.cs +++ b/sdk/dotnet/Inputs/JobJobClusterNewClusterGetArgs.cs @@ -18,9 +18,6 @@ public sealed class JobJobClusterNewClusterGetArgs : global::Pulumi.ResourceArgs [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobNewClusterArgs.cs b/sdk/dotnet/Inputs/JobNewClusterArgs.cs index d52e953c..e7dec293 100644 --- a/sdk/dotnet/Inputs/JobNewClusterArgs.cs +++ b/sdk/dotnet/Inputs/JobNewClusterArgs.cs @@ -18,9 +18,6 @@ public sealed class JobNewClusterArgs : global::Pulumi.ResourceArgs [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobNewClusterGetArgs.cs b/sdk/dotnet/Inputs/JobNewClusterGetArgs.cs index 4e811bff..2916ee23 100644 --- a/sdk/dotnet/Inputs/JobNewClusterGetArgs.cs +++ b/sdk/dotnet/Inputs/JobNewClusterGetArgs.cs @@ -18,9 +18,6 @@ public sealed class JobNewClusterGetArgs : global::Pulumi.ResourceArgs [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobTaskArgs.cs b/sdk/dotnet/Inputs/JobTaskArgs.cs index a6e10e25..a03d0645 100644 --- a/sdk/dotnet/Inputs/JobTaskArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskArgs.cs @@ -147,8 +147,8 @@ public InputList Libraries /// string specifying an unique key for a given task. /// * `*_task` - (Required) one of the specific task blocks described below: /// - [Input("taskKey")] - public Input? TaskKey { get; set; } + [Input("taskKey", required: true)] + public Input TaskKey { get; set; } = null!; /// /// (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskArgs.cs index 03d36d6e..f0c9f2d9 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskArgs.cs @@ -144,8 +144,8 @@ public InputList Libraries /// string specifying an unique key for a given task. /// * `*_task` - (Required) one of the specific task blocks described below: /// - [Input("taskKey")] - public Input? TaskKey { get; set; } + [Input("taskKey", required: true)] + public Input TaskKey { get; set; } = null!; /// /// (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskGetArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskGetArgs.cs index 32b71e1d..bfd2f501 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskGetArgs.cs @@ -144,8 +144,8 @@ public InputList Libraries /// string specifying an unique key for a given task. /// * `*_task` - (Required) one of the specific task blocks described below: /// - [Input("taskKey")] - public Input? TaskKey { get; set; } + [Input("taskKey", required: true)] + public Input TaskKey { get; set; } = null!; /// /// (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleArgs.cs index 9420f46d..81539026 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleArgs.cs @@ -15,20 +15,20 @@ public sealed class JobTaskForEachTaskTaskHealthRuleArgs : global::Pulumi.Resour /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; /// /// integer value used to compare to the given metric. /// - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public JobTaskForEachTaskTaskHealthRuleArgs() { diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleGetArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleGetArgs.cs index 1ff2486f..e1f18299 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskHealthRuleGetArgs.cs @@ -15,20 +15,20 @@ public sealed class JobTaskForEachTaskTaskHealthRuleGetArgs : global::Pulumi.Res /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; /// /// integer value used to compare to the given metric. /// - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public JobTaskForEachTaskTaskHealthRuleGetArgs() { diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterArgs.cs index 114126e1..ce79cde5 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterArgs.cs @@ -18,9 +18,6 @@ public sealed class JobTaskForEachTaskTaskNewClusterArgs : global::Pulumi.Resour [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterGetArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterGetArgs.cs index a7530384..80a978d3 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskNewClusterGetArgs.cs @@ -18,9 +18,6 @@ public sealed class JobTaskForEachTaskTaskNewClusterGetArgs : global::Pulumi.Res [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskArgs.cs index 2c1e6213..258d4f05 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskArgs.cs @@ -51,8 +51,8 @@ public InputMap Parameters /// /// ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. /// - [Input("warehouseId")] - public Input? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public Input WarehouseId { get; set; } = null!; public JobTaskForEachTaskTaskSqlTaskArgs() { diff --git a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskGetArgs.cs b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskGetArgs.cs index 9475ca44..c0cdb44a 100644 --- a/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskForEachTaskTaskSqlTaskGetArgs.cs @@ -51,8 +51,8 @@ public InputMap Parameters /// /// ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. /// - [Input("warehouseId")] - public Input? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public Input WarehouseId { get; set; } = null!; public JobTaskForEachTaskTaskSqlTaskGetArgs() { diff --git a/sdk/dotnet/Inputs/JobTaskGetArgs.cs b/sdk/dotnet/Inputs/JobTaskGetArgs.cs index 7eebc028..c47ea414 100644 --- a/sdk/dotnet/Inputs/JobTaskGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskGetArgs.cs @@ -147,8 +147,8 @@ public InputList Libraries /// string specifying an unique key for a given task. /// * `*_task` - (Required) one of the specific task blocks described below: /// - [Input("taskKey")] - public Input? TaskKey { get; set; } + [Input("taskKey", required: true)] + public Input TaskKey { get; set; } = null!; /// /// (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. diff --git a/sdk/dotnet/Inputs/JobTaskHealthRuleArgs.cs b/sdk/dotnet/Inputs/JobTaskHealthRuleArgs.cs index c04f180d..a77bb5d6 100644 --- a/sdk/dotnet/Inputs/JobTaskHealthRuleArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskHealthRuleArgs.cs @@ -15,20 +15,20 @@ public sealed class JobTaskHealthRuleArgs : global::Pulumi.ResourceArgs /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; /// /// integer value used to compare to the given metric. /// - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public JobTaskHealthRuleArgs() { diff --git a/sdk/dotnet/Inputs/JobTaskHealthRuleGetArgs.cs b/sdk/dotnet/Inputs/JobTaskHealthRuleGetArgs.cs index f3825ccb..ffe36afd 100644 --- a/sdk/dotnet/Inputs/JobTaskHealthRuleGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskHealthRuleGetArgs.cs @@ -15,20 +15,20 @@ public sealed class JobTaskHealthRuleGetArgs : global::Pulumi.ResourceArgs /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - [Input("metric")] - public Input? Metric { get; set; } + [Input("metric", required: true)] + public Input Metric { get; set; } = null!; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - [Input("op")] - public Input? Op { get; set; } + [Input("op", required: true)] + public Input Op { get; set; } = null!; /// /// integer value used to compare to the given metric. /// - [Input("value")] - public Input? Value { get; set; } + [Input("value", required: true)] + public Input Value { get; set; } = null!; public JobTaskHealthRuleGetArgs() { diff --git a/sdk/dotnet/Inputs/JobTaskNewClusterArgs.cs b/sdk/dotnet/Inputs/JobTaskNewClusterArgs.cs index 99602740..d99ca62d 100644 --- a/sdk/dotnet/Inputs/JobTaskNewClusterArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskNewClusterArgs.cs @@ -18,9 +18,6 @@ public sealed class JobTaskNewClusterArgs : global::Pulumi.ResourceArgs [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobTaskNewClusterGetArgs.cs b/sdk/dotnet/Inputs/JobTaskNewClusterGetArgs.cs index 259d8aee..1d2b8bfa 100644 --- a/sdk/dotnet/Inputs/JobTaskNewClusterGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskNewClusterGetArgs.cs @@ -18,9 +18,6 @@ public sealed class JobTaskNewClusterGetArgs : global::Pulumi.ResourceArgs [Input("autoscale")] public Input? Autoscale { get; set; } - [Input("autoterminationMinutes")] - public Input? AutoterminationMinutes { get; set; } - [Input("awsAttributes")] public Input? AwsAttributes { get; set; } diff --git a/sdk/dotnet/Inputs/JobTaskSqlTaskArgs.cs b/sdk/dotnet/Inputs/JobTaskSqlTaskArgs.cs index 4721516f..75b00a74 100644 --- a/sdk/dotnet/Inputs/JobTaskSqlTaskArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskSqlTaskArgs.cs @@ -51,8 +51,8 @@ public InputMap Parameters /// /// ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. /// - [Input("warehouseId")] - public Input? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public Input WarehouseId { get; set; } = null!; public JobTaskSqlTaskArgs() { diff --git a/sdk/dotnet/Inputs/JobTaskSqlTaskGetArgs.cs b/sdk/dotnet/Inputs/JobTaskSqlTaskGetArgs.cs index fc26ac46..3e681422 100644 --- a/sdk/dotnet/Inputs/JobTaskSqlTaskGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskSqlTaskGetArgs.cs @@ -51,8 +51,8 @@ public InputMap Parameters /// /// ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. /// - [Input("warehouseId")] - public Input? WarehouseId { get; set; } + [Input("warehouseId", required: true)] + public Input WarehouseId { get; set; } = null!; public JobTaskSqlTaskGetArgs() { diff --git a/sdk/dotnet/Inputs/JobTriggerArgs.cs b/sdk/dotnet/Inputs/JobTriggerArgs.cs index b9c00ea4..98ca7bab 100644 --- a/sdk/dotnet/Inputs/JobTriggerArgs.cs +++ b/sdk/dotnet/Inputs/JobTriggerArgs.cs @@ -27,9 +27,6 @@ public sealed class JobTriggerArgs : global::Pulumi.ResourceArgs [Input("table")] public Input? Table { get; set; } - /// - /// configuration block to define a trigger for Table Update events consisting of following attributes: - /// [Input("tableUpdate")] public Input? TableUpdate { get; set; } diff --git a/sdk/dotnet/Inputs/JobTriggerGetArgs.cs b/sdk/dotnet/Inputs/JobTriggerGetArgs.cs index 37d496f3..62d9de6d 100644 --- a/sdk/dotnet/Inputs/JobTriggerGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTriggerGetArgs.cs @@ -27,9 +27,6 @@ public sealed class JobTriggerGetArgs : global::Pulumi.ResourceArgs [Input("table")] public Input? Table { get; set; } - /// - /// configuration block to define a trigger for Table Update events consisting of following attributes: - /// [Input("tableUpdate")] public Input? TableUpdate { get; set; } diff --git a/sdk/dotnet/Inputs/JobTriggerTableUpdateArgs.cs b/sdk/dotnet/Inputs/JobTriggerTableUpdateArgs.cs index 880dae24..5b39b9ce 100644 --- a/sdk/dotnet/Inputs/JobTriggerTableUpdateArgs.cs +++ b/sdk/dotnet/Inputs/JobTriggerTableUpdateArgs.cs @@ -12,33 +12,20 @@ namespace Pulumi.Databricks.Inputs public sealed class JobTriggerTableUpdateArgs : global::Pulumi.ResourceArgs { - /// - /// The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - /// [Input("condition")] public Input? Condition { get; set; } - /// - /// If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - /// [Input("minTimeBetweenTriggersSeconds")] public Input? MinTimeBetweenTriggersSeconds { get; set; } [Input("tableNames", required: true)] private InputList? _tableNames; - - /// - /// A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - /// public InputList TableNames { get => _tableNames ?? (_tableNames = new InputList()); set => _tableNames = value; } - /// - /// If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - /// [Input("waitAfterLastChangeSeconds")] public Input? WaitAfterLastChangeSeconds { get; set; } diff --git a/sdk/dotnet/Inputs/JobTriggerTableUpdateGetArgs.cs b/sdk/dotnet/Inputs/JobTriggerTableUpdateGetArgs.cs index dff5431c..c0302a66 100644 --- a/sdk/dotnet/Inputs/JobTriggerTableUpdateGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTriggerTableUpdateGetArgs.cs @@ -12,33 +12,20 @@ namespace Pulumi.Databricks.Inputs public sealed class JobTriggerTableUpdateGetArgs : global::Pulumi.ResourceArgs { - /// - /// The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - /// [Input("condition")] public Input? Condition { get; set; } - /// - /// If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - /// [Input("minTimeBetweenTriggersSeconds")] public Input? MinTimeBetweenTriggersSeconds { get; set; } [Input("tableNames", required: true)] private InputList? _tableNames; - - /// - /// A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - /// public InputList TableNames { get => _tableNames ?? (_tableNames = new InputList()); set => _tableNames = value; } - /// - /// If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - /// [Input("waitAfterLastChangeSeconds")] public Input? WaitAfterLastChangeSeconds { get; set; } diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsHealthRuleResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsHealthRuleResult.cs index f2ec6b17..c12f2bb3 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsHealthRuleResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsHealthRuleResult.cs @@ -13,17 +13,17 @@ namespace Pulumi.Databricks.Outputs [OutputType] public sealed class GetJobJobSettingsSettingsHealthRuleResult { - public readonly string? Metric; - public readonly string? Op; - public readonly int? Value; + public readonly string Metric; + public readonly string Op; + public readonly int Value; [OutputConstructor] private GetJobJobSettingsSettingsHealthRuleResult( - string? metric, + string metric, - string? op, + string op, - int? value) + int value) { Metric = metric; Op = op; diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsJobClusterResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsJobClusterResult.cs index 63e6cc9c..1f91137e 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsJobClusterResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsJobClusterResult.cs @@ -13,14 +13,14 @@ namespace Pulumi.Databricks.Outputs [OutputType] public sealed class GetJobJobSettingsSettingsJobClusterResult { - public readonly string? JobClusterKey; - public readonly Outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult? NewCluster; + public readonly string JobClusterKey; + public readonly Outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult NewCluster; [OutputConstructor] private GetJobJobSettingsSettingsJobClusterResult( - string? jobClusterKey, + string jobClusterKey, - Outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult? newCluster) + Outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult newCluster) { JobClusterKey = jobClusterKey; NewCluster = newCluster; diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleResult.cs index 6d3b7345..9a6343a5 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleResult.cs @@ -13,17 +13,17 @@ namespace Pulumi.Databricks.Outputs [OutputType] public sealed class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleResult { - public readonly string? Metric; - public readonly string? Op; - public readonly int? Value; + public readonly string Metric; + public readonly string Op; + public readonly int Value; [OutputConstructor] private GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleResult( - string? metric, + string metric, - string? op, + string op, - int? value) + int value) { Metric = metric; Op = op; diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskResult.cs index dd9b0afd..f7c89e4d 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskResult.cs @@ -37,7 +37,7 @@ public sealed class GetJobJobSettingsSettingsTaskForEachTaskTaskResult public readonly Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTaskResult? SparkPythonTask; public readonly Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTaskResult? SparkSubmitTask; public readonly Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult? SqlTask; - public readonly string? TaskKey; + public readonly string TaskKey; public readonly int? TimeoutSeconds; public readonly Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotificationsResult? WebhookNotifications; @@ -91,7 +91,7 @@ private GetJobJobSettingsSettingsTaskForEachTaskTaskResult( Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult? sqlTask, - string? taskKey, + string taskKey, int? timeoutSeconds, diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult.cs index de61556d..dff2a1d0 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult.cs @@ -18,7 +18,7 @@ public sealed class GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult public readonly Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFileResult? File; public readonly ImmutableDictionary? Parameters; public readonly Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryResult? Query; - public readonly string? WarehouseId; + public readonly string WarehouseId; [OutputConstructor] private GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult( @@ -32,7 +32,7 @@ private GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult( Outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryResult? query, - string? warehouseId) + string warehouseId) { Alert = alert; Dashboard = dashboard; diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskHealthRuleResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskHealthRuleResult.cs index 4785f48d..e26aa7b8 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskHealthRuleResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskHealthRuleResult.cs @@ -13,17 +13,17 @@ namespace Pulumi.Databricks.Outputs [OutputType] public sealed class GetJobJobSettingsSettingsTaskHealthRuleResult { - public readonly string? Metric; - public readonly string? Op; - public readonly int? Value; + public readonly string Metric; + public readonly string Op; + public readonly int Value; [OutputConstructor] private GetJobJobSettingsSettingsTaskHealthRuleResult( - string? metric, + string metric, - string? op, + string op, - int? value) + int value) { Metric = metric; Op = op; diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskResult.cs index 9a5b2edc..8635342e 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskResult.cs @@ -38,7 +38,7 @@ public sealed class GetJobJobSettingsSettingsTaskResult public readonly Outputs.GetJobJobSettingsSettingsTaskSparkPythonTaskResult? SparkPythonTask; public readonly Outputs.GetJobJobSettingsSettingsTaskSparkSubmitTaskResult? SparkSubmitTask; public readonly Outputs.GetJobJobSettingsSettingsTaskSqlTaskResult? SqlTask; - public readonly string? TaskKey; + public readonly string TaskKey; public readonly int? TimeoutSeconds; public readonly Outputs.GetJobJobSettingsSettingsTaskWebhookNotificationsResult? WebhookNotifications; @@ -94,7 +94,7 @@ private GetJobJobSettingsSettingsTaskResult( Outputs.GetJobJobSettingsSettingsTaskSqlTaskResult? sqlTask, - string? taskKey, + string taskKey, int? timeoutSeconds, diff --git a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskSqlTaskResult.cs b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskSqlTaskResult.cs index f001d119..f249b56f 100644 --- a/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskSqlTaskResult.cs +++ b/sdk/dotnet/Outputs/GetJobJobSettingsSettingsTaskSqlTaskResult.cs @@ -18,7 +18,7 @@ public sealed class GetJobJobSettingsSettingsTaskSqlTaskResult public readonly Outputs.GetJobJobSettingsSettingsTaskSqlTaskFileResult? File; public readonly ImmutableDictionary? Parameters; public readonly Outputs.GetJobJobSettingsSettingsTaskSqlTaskQueryResult? Query; - public readonly string? WarehouseId; + public readonly string WarehouseId; [OutputConstructor] private GetJobJobSettingsSettingsTaskSqlTaskResult( @@ -32,7 +32,7 @@ private GetJobJobSettingsSettingsTaskSqlTaskResult( Outputs.GetJobJobSettingsSettingsTaskSqlTaskQueryResult? query, - string? warehouseId) + string warehouseId) { Alert = alert; Dashboard = dashboard; diff --git a/sdk/dotnet/Outputs/JobHealthRule.cs b/sdk/dotnet/Outputs/JobHealthRule.cs index 781ff337..0e2ca4e0 100644 --- a/sdk/dotnet/Outputs/JobHealthRule.cs +++ b/sdk/dotnet/Outputs/JobHealthRule.cs @@ -16,23 +16,23 @@ public sealed class JobHealthRule /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - public readonly string? Metric; + public readonly string Metric; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - public readonly string? Op; + public readonly string Op; /// /// integer value used to compare to the given metric. /// - public readonly int? Value; + public readonly int Value; [OutputConstructor] private JobHealthRule( - string? metric, + string metric, - string? op, + string op, - int? value) + int value) { Metric = metric; Op = op; diff --git a/sdk/dotnet/Outputs/JobJobCluster.cs b/sdk/dotnet/Outputs/JobJobCluster.cs index f759a876..5b68a195 100644 --- a/sdk/dotnet/Outputs/JobJobCluster.cs +++ b/sdk/dotnet/Outputs/JobJobCluster.cs @@ -16,17 +16,17 @@ public sealed class JobJobCluster /// /// Identifier that can be referenced in `task` block, so that cluster is shared between tasks /// - public readonly string? JobClusterKey; + public readonly string JobClusterKey; /// /// Same set of parameters as for databricks.Cluster resource. /// - public readonly Outputs.JobJobClusterNewCluster? NewCluster; + public readonly Outputs.JobJobClusterNewCluster NewCluster; [OutputConstructor] private JobJobCluster( - string? jobClusterKey, + string jobClusterKey, - Outputs.JobJobClusterNewCluster? newCluster) + Outputs.JobJobClusterNewCluster newCluster) { JobClusterKey = jobClusterKey; NewCluster = newCluster; diff --git a/sdk/dotnet/Outputs/JobJobClusterNewCluster.cs b/sdk/dotnet/Outputs/JobJobClusterNewCluster.cs index b75fa7fb..28a67357 100644 --- a/sdk/dotnet/Outputs/JobJobClusterNewCluster.cs +++ b/sdk/dotnet/Outputs/JobJobClusterNewCluster.cs @@ -15,7 +15,6 @@ public sealed class JobJobClusterNewCluster { public readonly bool? ApplyPolicyDefaultValues; public readonly Outputs.JobJobClusterNewClusterAutoscale? Autoscale; - public readonly int? AutoterminationMinutes; public readonly Outputs.JobJobClusterNewClusterAwsAttributes? AwsAttributes; public readonly Outputs.JobJobClusterNewClusterAzureAttributes? AzureAttributes; public readonly string? ClusterId; @@ -54,8 +53,6 @@ private JobJobClusterNewCluster( Outputs.JobJobClusterNewClusterAutoscale? autoscale, - int? autoterminationMinutes, - Outputs.JobJobClusterNewClusterAwsAttributes? awsAttributes, Outputs.JobJobClusterNewClusterAzureAttributes? azureAttributes, @@ -114,7 +111,6 @@ private JobJobClusterNewCluster( { ApplyPolicyDefaultValues = applyPolicyDefaultValues; Autoscale = autoscale; - AutoterminationMinutes = autoterminationMinutes; AwsAttributes = awsAttributes; AzureAttributes = azureAttributes; ClusterId = clusterId; diff --git a/sdk/dotnet/Outputs/JobNewCluster.cs b/sdk/dotnet/Outputs/JobNewCluster.cs index 7f848bcd..e8558aec 100644 --- a/sdk/dotnet/Outputs/JobNewCluster.cs +++ b/sdk/dotnet/Outputs/JobNewCluster.cs @@ -15,7 +15,6 @@ public sealed class JobNewCluster { public readonly bool? ApplyPolicyDefaultValues; public readonly Outputs.JobNewClusterAutoscale? Autoscale; - public readonly int? AutoterminationMinutes; public readonly Outputs.JobNewClusterAwsAttributes? AwsAttributes; public readonly Outputs.JobNewClusterAzureAttributes? AzureAttributes; public readonly string? ClusterId; @@ -54,8 +53,6 @@ private JobNewCluster( Outputs.JobNewClusterAutoscale? autoscale, - int? autoterminationMinutes, - Outputs.JobNewClusterAwsAttributes? awsAttributes, Outputs.JobNewClusterAzureAttributes? azureAttributes, @@ -114,7 +111,6 @@ private JobNewCluster( { ApplyPolicyDefaultValues = applyPolicyDefaultValues; Autoscale = autoscale; - AutoterminationMinutes = autoterminationMinutes; AwsAttributes = awsAttributes; AzureAttributes = azureAttributes; ClusterId = clusterId; diff --git a/sdk/dotnet/Outputs/JobTask.cs b/sdk/dotnet/Outputs/JobTask.cs index 2c1f59b5..09aefec3 100644 --- a/sdk/dotnet/Outputs/JobTask.cs +++ b/sdk/dotnet/Outputs/JobTask.cs @@ -84,7 +84,7 @@ public sealed class JobTask /// string specifying an unique key for a given task. /// * `*_task` - (Required) one of the specific task blocks described below: /// - public readonly string? TaskKey; + public readonly string TaskKey; /// /// (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. /// @@ -148,7 +148,7 @@ private JobTask( Outputs.JobTaskSqlTask? sqlTask, - string? taskKey, + string taskKey, int? timeoutSeconds, diff --git a/sdk/dotnet/Outputs/JobTaskForEachTaskTask.cs b/sdk/dotnet/Outputs/JobTaskForEachTaskTask.cs index 886f936c..27f88f10 100644 --- a/sdk/dotnet/Outputs/JobTaskForEachTaskTask.cs +++ b/sdk/dotnet/Outputs/JobTaskForEachTaskTask.cs @@ -83,7 +83,7 @@ public sealed class JobTaskForEachTaskTask /// string specifying an unique key for a given task. /// * `*_task` - (Required) one of the specific task blocks described below: /// - public readonly string? TaskKey; + public readonly string TaskKey; /// /// (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. /// @@ -145,7 +145,7 @@ private JobTaskForEachTaskTask( Outputs.JobTaskForEachTaskTaskSqlTask? sqlTask, - string? taskKey, + string taskKey, int? timeoutSeconds, diff --git a/sdk/dotnet/Outputs/JobTaskForEachTaskTaskHealthRule.cs b/sdk/dotnet/Outputs/JobTaskForEachTaskTaskHealthRule.cs index d21abae9..22b3d552 100644 --- a/sdk/dotnet/Outputs/JobTaskForEachTaskTaskHealthRule.cs +++ b/sdk/dotnet/Outputs/JobTaskForEachTaskTaskHealthRule.cs @@ -16,23 +16,23 @@ public sealed class JobTaskForEachTaskTaskHealthRule /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - public readonly string? Metric; + public readonly string Metric; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - public readonly string? Op; + public readonly string Op; /// /// integer value used to compare to the given metric. /// - public readonly int? Value; + public readonly int Value; [OutputConstructor] private JobTaskForEachTaskTaskHealthRule( - string? metric, + string metric, - string? op, + string op, - int? value) + int value) { Metric = metric; Op = op; diff --git a/sdk/dotnet/Outputs/JobTaskForEachTaskTaskNewCluster.cs b/sdk/dotnet/Outputs/JobTaskForEachTaskTaskNewCluster.cs index 4183cfd3..b2926dbb 100644 --- a/sdk/dotnet/Outputs/JobTaskForEachTaskTaskNewCluster.cs +++ b/sdk/dotnet/Outputs/JobTaskForEachTaskTaskNewCluster.cs @@ -15,7 +15,6 @@ public sealed class JobTaskForEachTaskTaskNewCluster { public readonly bool? ApplyPolicyDefaultValues; public readonly Outputs.JobTaskForEachTaskTaskNewClusterAutoscale? Autoscale; - public readonly int? AutoterminationMinutes; public readonly Outputs.JobTaskForEachTaskTaskNewClusterAwsAttributes? AwsAttributes; public readonly Outputs.JobTaskForEachTaskTaskNewClusterAzureAttributes? AzureAttributes; public readonly string? ClusterId; @@ -54,8 +53,6 @@ private JobTaskForEachTaskTaskNewCluster( Outputs.JobTaskForEachTaskTaskNewClusterAutoscale? autoscale, - int? autoterminationMinutes, - Outputs.JobTaskForEachTaskTaskNewClusterAwsAttributes? awsAttributes, Outputs.JobTaskForEachTaskTaskNewClusterAzureAttributes? azureAttributes, @@ -114,7 +111,6 @@ private JobTaskForEachTaskTaskNewCluster( { ApplyPolicyDefaultValues = applyPolicyDefaultValues; Autoscale = autoscale; - AutoterminationMinutes = autoterminationMinutes; AwsAttributes = awsAttributes; AzureAttributes = azureAttributes; ClusterId = clusterId; diff --git a/sdk/dotnet/Outputs/JobTaskForEachTaskTaskSqlTask.cs b/sdk/dotnet/Outputs/JobTaskForEachTaskTaskSqlTask.cs index 5a916916..79724a91 100644 --- a/sdk/dotnet/Outputs/JobTaskForEachTaskTaskSqlTask.cs +++ b/sdk/dotnet/Outputs/JobTaskForEachTaskTaskSqlTask.cs @@ -36,7 +36,7 @@ public sealed class JobTaskForEachTaskTaskSqlTask /// /// ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. /// - public readonly string? WarehouseId; + public readonly string WarehouseId; [OutputConstructor] private JobTaskForEachTaskTaskSqlTask( @@ -50,7 +50,7 @@ private JobTaskForEachTaskTaskSqlTask( Outputs.JobTaskForEachTaskTaskSqlTaskQuery? query, - string? warehouseId) + string warehouseId) { Alert = alert; Dashboard = dashboard; diff --git a/sdk/dotnet/Outputs/JobTaskHealthRule.cs b/sdk/dotnet/Outputs/JobTaskHealthRule.cs index 0cc299b6..7d497228 100644 --- a/sdk/dotnet/Outputs/JobTaskHealthRule.cs +++ b/sdk/dotnet/Outputs/JobTaskHealthRule.cs @@ -16,23 +16,23 @@ public sealed class JobTaskHealthRule /// /// string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). /// - public readonly string? Metric; + public readonly string Metric; /// /// string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. /// - public readonly string? Op; + public readonly string Op; /// /// integer value used to compare to the given metric. /// - public readonly int? Value; + public readonly int Value; [OutputConstructor] private JobTaskHealthRule( - string? metric, + string metric, - string? op, + string op, - int? value) + int value) { Metric = metric; Op = op; diff --git a/sdk/dotnet/Outputs/JobTaskNewCluster.cs b/sdk/dotnet/Outputs/JobTaskNewCluster.cs index ccf634ed..a309789b 100644 --- a/sdk/dotnet/Outputs/JobTaskNewCluster.cs +++ b/sdk/dotnet/Outputs/JobTaskNewCluster.cs @@ -15,7 +15,6 @@ public sealed class JobTaskNewCluster { public readonly bool? ApplyPolicyDefaultValues; public readonly Outputs.JobTaskNewClusterAutoscale? Autoscale; - public readonly int? AutoterminationMinutes; public readonly Outputs.JobTaskNewClusterAwsAttributes? AwsAttributes; public readonly Outputs.JobTaskNewClusterAzureAttributes? AzureAttributes; public readonly string? ClusterId; @@ -54,8 +53,6 @@ private JobTaskNewCluster( Outputs.JobTaskNewClusterAutoscale? autoscale, - int? autoterminationMinutes, - Outputs.JobTaskNewClusterAwsAttributes? awsAttributes, Outputs.JobTaskNewClusterAzureAttributes? azureAttributes, @@ -114,7 +111,6 @@ private JobTaskNewCluster( { ApplyPolicyDefaultValues = applyPolicyDefaultValues; Autoscale = autoscale; - AutoterminationMinutes = autoterminationMinutes; AwsAttributes = awsAttributes; AzureAttributes = azureAttributes; ClusterId = clusterId; diff --git a/sdk/dotnet/Outputs/JobTaskSqlTask.cs b/sdk/dotnet/Outputs/JobTaskSqlTask.cs index a9c0d932..135582e4 100644 --- a/sdk/dotnet/Outputs/JobTaskSqlTask.cs +++ b/sdk/dotnet/Outputs/JobTaskSqlTask.cs @@ -36,7 +36,7 @@ public sealed class JobTaskSqlTask /// /// ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. /// - public readonly string? WarehouseId; + public readonly string WarehouseId; [OutputConstructor] private JobTaskSqlTask( @@ -50,7 +50,7 @@ private JobTaskSqlTask( Outputs.JobTaskSqlTaskQuery? query, - string? warehouseId) + string warehouseId) { Alert = alert; Dashboard = dashboard; diff --git a/sdk/dotnet/Outputs/JobTrigger.cs b/sdk/dotnet/Outputs/JobTrigger.cs index 00ca8499..e94f3689 100644 --- a/sdk/dotnet/Outputs/JobTrigger.cs +++ b/sdk/dotnet/Outputs/JobTrigger.cs @@ -22,9 +22,6 @@ public sealed class JobTrigger /// public readonly string? PauseStatus; public readonly Outputs.JobTriggerTable? Table; - /// - /// configuration block to define a trigger for Table Update events consisting of following attributes: - /// public readonly Outputs.JobTriggerTableUpdate? TableUpdate; [OutputConstructor] diff --git a/sdk/dotnet/Outputs/JobTriggerTableUpdate.cs b/sdk/dotnet/Outputs/JobTriggerTableUpdate.cs index 82615404..0b9c08d0 100644 --- a/sdk/dotnet/Outputs/JobTriggerTableUpdate.cs +++ b/sdk/dotnet/Outputs/JobTriggerTableUpdate.cs @@ -13,21 +13,9 @@ namespace Pulumi.Databricks.Outputs [OutputType] public sealed class JobTriggerTableUpdate { - /// - /// The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - /// public readonly string? Condition; - /// - /// If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - /// public readonly int? MinTimeBetweenTriggersSeconds; - /// - /// A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - /// public readonly ImmutableArray TableNames; - /// - /// If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - /// public readonly int? WaitAfterLastChangeSeconds; [OutputConstructor] diff --git a/sdk/go/databricks/pulumiTypes.go b/sdk/go/databricks/pulumiTypes.go index d62e3409..85b7a4e3 100644 --- a/sdk/go/databricks/pulumiTypes.go +++ b/sdk/go/databricks/pulumiTypes.go @@ -10614,11 +10614,11 @@ func (o JobHealthPtrOutput) Rules() JobHealthRuleArrayOutput { type JobHealthRule struct { // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). - Metric *string `pulumi:"metric"` + Metric string `pulumi:"metric"` // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. - Op *string `pulumi:"op"` + Op string `pulumi:"op"` // integer value used to compare to the given metric. - Value *int `pulumi:"value"` + Value int `pulumi:"value"` } // JobHealthRuleInput is an input type that accepts JobHealthRuleArgs and JobHealthRuleOutput values. @@ -10634,11 +10634,11 @@ type JobHealthRuleInput interface { type JobHealthRuleArgs struct { // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). - Metric pulumi.StringPtrInput `pulumi:"metric"` + Metric pulumi.StringInput `pulumi:"metric"` // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. - Op pulumi.StringPtrInput `pulumi:"op"` + Op pulumi.StringInput `pulumi:"op"` // integer value used to compare to the given metric. - Value pulumi.IntPtrInput `pulumi:"value"` + Value pulumi.IntInput `pulumi:"value"` } func (JobHealthRuleArgs) ElementType() reflect.Type { @@ -10693,18 +10693,18 @@ func (o JobHealthRuleOutput) ToJobHealthRuleOutputWithContext(ctx context.Contex } // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). -func (o JobHealthRuleOutput) Metric() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobHealthRule) *string { return v.Metric }).(pulumi.StringPtrOutput) +func (o JobHealthRuleOutput) Metric() pulumi.StringOutput { + return o.ApplyT(func(v JobHealthRule) string { return v.Metric }).(pulumi.StringOutput) } // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. -func (o JobHealthRuleOutput) Op() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobHealthRule) *string { return v.Op }).(pulumi.StringPtrOutput) +func (o JobHealthRuleOutput) Op() pulumi.StringOutput { + return o.ApplyT(func(v JobHealthRule) string { return v.Op }).(pulumi.StringOutput) } // integer value used to compare to the given metric. -func (o JobHealthRuleOutput) Value() pulumi.IntPtrOutput { - return o.ApplyT(func(v JobHealthRule) *int { return v.Value }).(pulumi.IntPtrOutput) +func (o JobHealthRuleOutput) Value() pulumi.IntOutput { + return o.ApplyT(func(v JobHealthRule) int { return v.Value }).(pulumi.IntOutput) } type JobHealthRuleArrayOutput struct{ *pulumi.OutputState } @@ -10729,9 +10729,9 @@ func (o JobHealthRuleArrayOutput) Index(i pulumi.IntInput) JobHealthRuleOutput { type JobJobCluster struct { // Identifier that can be referenced in `task` block, so that cluster is shared between tasks - JobClusterKey *string `pulumi:"jobClusterKey"` + JobClusterKey string `pulumi:"jobClusterKey"` // Same set of parameters as for Cluster resource. - NewCluster *JobJobClusterNewCluster `pulumi:"newCluster"` + NewCluster JobJobClusterNewCluster `pulumi:"newCluster"` } // JobJobClusterInput is an input type that accepts JobJobClusterArgs and JobJobClusterOutput values. @@ -10747,9 +10747,9 @@ type JobJobClusterInput interface { type JobJobClusterArgs struct { // Identifier that can be referenced in `task` block, so that cluster is shared between tasks - JobClusterKey pulumi.StringPtrInput `pulumi:"jobClusterKey"` + JobClusterKey pulumi.StringInput `pulumi:"jobClusterKey"` // Same set of parameters as for Cluster resource. - NewCluster JobJobClusterNewClusterPtrInput `pulumi:"newCluster"` + NewCluster JobJobClusterNewClusterInput `pulumi:"newCluster"` } func (JobJobClusterArgs) ElementType() reflect.Type { @@ -10804,13 +10804,13 @@ func (o JobJobClusterOutput) ToJobJobClusterOutputWithContext(ctx context.Contex } // Identifier that can be referenced in `task` block, so that cluster is shared between tasks -func (o JobJobClusterOutput) JobClusterKey() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobJobCluster) *string { return v.JobClusterKey }).(pulumi.StringPtrOutput) +func (o JobJobClusterOutput) JobClusterKey() pulumi.StringOutput { + return o.ApplyT(func(v JobJobCluster) string { return v.JobClusterKey }).(pulumi.StringOutput) } // Same set of parameters as for Cluster resource. -func (o JobJobClusterOutput) NewCluster() JobJobClusterNewClusterPtrOutput { - return o.ApplyT(func(v JobJobCluster) *JobJobClusterNewCluster { return v.NewCluster }).(JobJobClusterNewClusterPtrOutput) +func (o JobJobClusterOutput) NewCluster() JobJobClusterNewClusterOutput { + return o.ApplyT(func(v JobJobCluster) JobJobClusterNewCluster { return v.NewCluster }).(JobJobClusterNewClusterOutput) } type JobJobClusterArrayOutput struct{ *pulumi.OutputState } @@ -10836,7 +10836,6 @@ func (o JobJobClusterArrayOutput) Index(i pulumi.IntInput) JobJobClusterOutput { type JobJobClusterNewCluster struct { ApplyPolicyDefaultValues *bool `pulumi:"applyPolicyDefaultValues"` Autoscale *JobJobClusterNewClusterAutoscale `pulumi:"autoscale"` - AutoterminationMinutes *int `pulumi:"autoterminationMinutes"` AwsAttributes *JobJobClusterNewClusterAwsAttributes `pulumi:"awsAttributes"` AzureAttributes *JobJobClusterNewClusterAzureAttributes `pulumi:"azureAttributes"` ClusterId *string `pulumi:"clusterId"` @@ -10882,7 +10881,6 @@ type JobJobClusterNewClusterInput interface { type JobJobClusterNewClusterArgs struct { ApplyPolicyDefaultValues pulumi.BoolPtrInput `pulumi:"applyPolicyDefaultValues"` Autoscale JobJobClusterNewClusterAutoscalePtrInput `pulumi:"autoscale"` - AutoterminationMinutes pulumi.IntPtrInput `pulumi:"autoterminationMinutes"` AwsAttributes JobJobClusterNewClusterAwsAttributesPtrInput `pulumi:"awsAttributes"` AzureAttributes JobJobClusterNewClusterAzureAttributesPtrInput `pulumi:"azureAttributes"` ClusterId pulumi.StringPtrInput `pulumi:"clusterId"` @@ -10926,47 +10924,6 @@ func (i JobJobClusterNewClusterArgs) ToJobJobClusterNewClusterOutputWithContext( return pulumi.ToOutputWithContext(ctx, i).(JobJobClusterNewClusterOutput) } -func (i JobJobClusterNewClusterArgs) ToJobJobClusterNewClusterPtrOutput() JobJobClusterNewClusterPtrOutput { - return i.ToJobJobClusterNewClusterPtrOutputWithContext(context.Background()) -} - -func (i JobJobClusterNewClusterArgs) ToJobJobClusterNewClusterPtrOutputWithContext(ctx context.Context) JobJobClusterNewClusterPtrOutput { - return pulumi.ToOutputWithContext(ctx, i).(JobJobClusterNewClusterOutput).ToJobJobClusterNewClusterPtrOutputWithContext(ctx) -} - -// JobJobClusterNewClusterPtrInput is an input type that accepts JobJobClusterNewClusterArgs, JobJobClusterNewClusterPtr and JobJobClusterNewClusterPtrOutput values. -// You can construct a concrete instance of `JobJobClusterNewClusterPtrInput` via: -// -// JobJobClusterNewClusterArgs{...} -// -// or: -// -// nil -type JobJobClusterNewClusterPtrInput interface { - pulumi.Input - - ToJobJobClusterNewClusterPtrOutput() JobJobClusterNewClusterPtrOutput - ToJobJobClusterNewClusterPtrOutputWithContext(context.Context) JobJobClusterNewClusterPtrOutput -} - -type jobJobClusterNewClusterPtrType JobJobClusterNewClusterArgs - -func JobJobClusterNewClusterPtr(v *JobJobClusterNewClusterArgs) JobJobClusterNewClusterPtrInput { - return (*jobJobClusterNewClusterPtrType)(v) -} - -func (*jobJobClusterNewClusterPtrType) ElementType() reflect.Type { - return reflect.TypeOf((**JobJobClusterNewCluster)(nil)).Elem() -} - -func (i *jobJobClusterNewClusterPtrType) ToJobJobClusterNewClusterPtrOutput() JobJobClusterNewClusterPtrOutput { - return i.ToJobJobClusterNewClusterPtrOutputWithContext(context.Background()) -} - -func (i *jobJobClusterNewClusterPtrType) ToJobJobClusterNewClusterPtrOutputWithContext(ctx context.Context) JobJobClusterNewClusterPtrOutput { - return pulumi.ToOutputWithContext(ctx, i).(JobJobClusterNewClusterPtrOutput) -} - type JobJobClusterNewClusterOutput struct{ *pulumi.OutputState } func (JobJobClusterNewClusterOutput) ElementType() reflect.Type { @@ -10981,16 +10938,6 @@ func (o JobJobClusterNewClusterOutput) ToJobJobClusterNewClusterOutputWithContex return o } -func (o JobJobClusterNewClusterOutput) ToJobJobClusterNewClusterPtrOutput() JobJobClusterNewClusterPtrOutput { - return o.ToJobJobClusterNewClusterPtrOutputWithContext(context.Background()) -} - -func (o JobJobClusterNewClusterOutput) ToJobJobClusterNewClusterPtrOutputWithContext(ctx context.Context) JobJobClusterNewClusterPtrOutput { - return o.ApplyTWithContext(ctx, func(_ context.Context, v JobJobClusterNewCluster) *JobJobClusterNewCluster { - return &v - }).(JobJobClusterNewClusterPtrOutput) -} - func (o JobJobClusterNewClusterOutput) ApplyPolicyDefaultValues() pulumi.BoolPtrOutput { return o.ApplyT(func(v JobJobClusterNewCluster) *bool { return v.ApplyPolicyDefaultValues }).(pulumi.BoolPtrOutput) } @@ -10999,10 +10946,6 @@ func (o JobJobClusterNewClusterOutput) Autoscale() JobJobClusterNewClusterAutosc return o.ApplyT(func(v JobJobClusterNewCluster) *JobJobClusterNewClusterAutoscale { return v.Autoscale }).(JobJobClusterNewClusterAutoscalePtrOutput) } -func (o JobJobClusterNewClusterOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v JobJobClusterNewCluster) *int { return v.AutoterminationMinutes }).(pulumi.IntPtrOutput) -} - func (o JobJobClusterNewClusterOutput) AwsAttributes() JobJobClusterNewClusterAwsAttributesPtrOutput { return o.ApplyT(func(v JobJobClusterNewCluster) *JobJobClusterNewClusterAwsAttributes { return v.AwsAttributes }).(JobJobClusterNewClusterAwsAttributesPtrOutput) } @@ -11116,310 +11059,6 @@ func (o JobJobClusterNewClusterOutput) WorkloadType() JobJobClusterNewClusterWor return o.ApplyT(func(v JobJobClusterNewCluster) *JobJobClusterNewClusterWorkloadType { return v.WorkloadType }).(JobJobClusterNewClusterWorkloadTypePtrOutput) } -type JobJobClusterNewClusterPtrOutput struct{ *pulumi.OutputState } - -func (JobJobClusterNewClusterPtrOutput) ElementType() reflect.Type { - return reflect.TypeOf((**JobJobClusterNewCluster)(nil)).Elem() -} - -func (o JobJobClusterNewClusterPtrOutput) ToJobJobClusterNewClusterPtrOutput() JobJobClusterNewClusterPtrOutput { - return o -} - -func (o JobJobClusterNewClusterPtrOutput) ToJobJobClusterNewClusterPtrOutputWithContext(ctx context.Context) JobJobClusterNewClusterPtrOutput { - return o -} - -func (o JobJobClusterNewClusterPtrOutput) Elem() JobJobClusterNewClusterOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) JobJobClusterNewCluster { - if v != nil { - return *v - } - var ret JobJobClusterNewCluster - return ret - }).(JobJobClusterNewClusterOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) ApplyPolicyDefaultValues() pulumi.BoolPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *bool { - if v == nil { - return nil - } - return v.ApplyPolicyDefaultValues - }).(pulumi.BoolPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) Autoscale() JobJobClusterNewClusterAutoscalePtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *JobJobClusterNewClusterAutoscale { - if v == nil { - return nil - } - return v.Autoscale - }).(JobJobClusterNewClusterAutoscalePtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *int { - if v == nil { - return nil - } - return v.AutoterminationMinutes - }).(pulumi.IntPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) AwsAttributes() JobJobClusterNewClusterAwsAttributesPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *JobJobClusterNewClusterAwsAttributes { - if v == nil { - return nil - } - return v.AwsAttributes - }).(JobJobClusterNewClusterAwsAttributesPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) AzureAttributes() JobJobClusterNewClusterAzureAttributesPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *JobJobClusterNewClusterAzureAttributes { - if v == nil { - return nil - } - return v.AzureAttributes - }).(JobJobClusterNewClusterAzureAttributesPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) ClusterId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.ClusterId - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) ClusterLogConf() JobJobClusterNewClusterClusterLogConfPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *JobJobClusterNewClusterClusterLogConf { - if v == nil { - return nil - } - return v.ClusterLogConf - }).(JobJobClusterNewClusterClusterLogConfPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) ClusterMountInfos() JobJobClusterNewClusterClusterMountInfoArrayOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) []JobJobClusterNewClusterClusterMountInfo { - if v == nil { - return nil - } - return v.ClusterMountInfos - }).(JobJobClusterNewClusterClusterMountInfoArrayOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) ClusterName() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.ClusterName - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) CustomTags() pulumi.MapOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) map[string]interface{} { - if v == nil { - return nil - } - return v.CustomTags - }).(pulumi.MapOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) DataSecurityMode() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.DataSecurityMode - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) DockerImage() JobJobClusterNewClusterDockerImagePtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *JobJobClusterNewClusterDockerImage { - if v == nil { - return nil - } - return v.DockerImage - }).(JobJobClusterNewClusterDockerImagePtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) DriverInstancePoolId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.DriverInstancePoolId - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) DriverNodeTypeId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.DriverNodeTypeId - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) EnableElasticDisk() pulumi.BoolPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *bool { - if v == nil { - return nil - } - return v.EnableElasticDisk - }).(pulumi.BoolPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) EnableLocalDiskEncryption() pulumi.BoolPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *bool { - if v == nil { - return nil - } - return v.EnableLocalDiskEncryption - }).(pulumi.BoolPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) GcpAttributes() JobJobClusterNewClusterGcpAttributesPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *JobJobClusterNewClusterGcpAttributes { - if v == nil { - return nil - } - return v.GcpAttributes - }).(JobJobClusterNewClusterGcpAttributesPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) IdempotencyToken() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.IdempotencyToken - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) InitScripts() JobJobClusterNewClusterInitScriptArrayOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) []JobJobClusterNewClusterInitScript { - if v == nil { - return nil - } - return v.InitScripts - }).(JobJobClusterNewClusterInitScriptArrayOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) InstancePoolId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.InstancePoolId - }).(pulumi.StringPtrOutput) -} - -// (List) An optional list of libraries to be installed on the cluster that will execute the job. See library Configuration Block below. -func (o JobJobClusterNewClusterPtrOutput) Libraries() JobJobClusterNewClusterLibraryArrayOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) []JobJobClusterNewClusterLibrary { - if v == nil { - return nil - } - return v.Libraries - }).(JobJobClusterNewClusterLibraryArrayOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) NodeTypeId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.NodeTypeId - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) NumWorkers() pulumi.IntPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *int { - if v == nil { - return nil - } - return v.NumWorkers - }).(pulumi.IntPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) PolicyId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.PolicyId - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) RuntimeEngine() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.RuntimeEngine - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) SingleUserName() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.SingleUserName - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) SparkConf() pulumi.MapOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) map[string]interface{} { - if v == nil { - return nil - } - return v.SparkConf - }).(pulumi.MapOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) SparkEnvVars() pulumi.MapOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) map[string]interface{} { - if v == nil { - return nil - } - return v.SparkEnvVars - }).(pulumi.MapOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) SparkVersion() pulumi.StringPtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *string { - if v == nil { - return nil - } - return &v.SparkVersion - }).(pulumi.StringPtrOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) SshPublicKeys() pulumi.StringArrayOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) []string { - if v == nil { - return nil - } - return v.SshPublicKeys - }).(pulumi.StringArrayOutput) -} - -func (o JobJobClusterNewClusterPtrOutput) WorkloadType() JobJobClusterNewClusterWorkloadTypePtrOutput { - return o.ApplyT(func(v *JobJobClusterNewCluster) *JobJobClusterNewClusterWorkloadType { - if v == nil { - return nil - } - return v.WorkloadType - }).(JobJobClusterNewClusterWorkloadTypePtrOutput) -} - type JobJobClusterNewClusterAutoscale struct { MaxWorkers *int `pulumi:"maxWorkers"` MinWorkers *int `pulumi:"minWorkers"` @@ -15967,7 +15606,6 @@ func (o JobLibraryPypiPtrOutput) Repo() pulumi.StringPtrOutput { type JobNewCluster struct { ApplyPolicyDefaultValues *bool `pulumi:"applyPolicyDefaultValues"` Autoscale *JobNewClusterAutoscale `pulumi:"autoscale"` - AutoterminationMinutes *int `pulumi:"autoterminationMinutes"` AwsAttributes *JobNewClusterAwsAttributes `pulumi:"awsAttributes"` AzureAttributes *JobNewClusterAzureAttributes `pulumi:"azureAttributes"` ClusterId *string `pulumi:"clusterId"` @@ -16013,7 +15651,6 @@ type JobNewClusterInput interface { type JobNewClusterArgs struct { ApplyPolicyDefaultValues pulumi.BoolPtrInput `pulumi:"applyPolicyDefaultValues"` Autoscale JobNewClusterAutoscalePtrInput `pulumi:"autoscale"` - AutoterminationMinutes pulumi.IntPtrInput `pulumi:"autoterminationMinutes"` AwsAttributes JobNewClusterAwsAttributesPtrInput `pulumi:"awsAttributes"` AzureAttributes JobNewClusterAzureAttributesPtrInput `pulumi:"azureAttributes"` ClusterId pulumi.StringPtrInput `pulumi:"clusterId"` @@ -16130,10 +15767,6 @@ func (o JobNewClusterOutput) Autoscale() JobNewClusterAutoscalePtrOutput { return o.ApplyT(func(v JobNewCluster) *JobNewClusterAutoscale { return v.Autoscale }).(JobNewClusterAutoscalePtrOutput) } -func (o JobNewClusterOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v JobNewCluster) *int { return v.AutoterminationMinutes }).(pulumi.IntPtrOutput) -} - func (o JobNewClusterOutput) AwsAttributes() JobNewClusterAwsAttributesPtrOutput { return o.ApplyT(func(v JobNewCluster) *JobNewClusterAwsAttributes { return v.AwsAttributes }).(JobNewClusterAwsAttributesPtrOutput) } @@ -16289,15 +15922,6 @@ func (o JobNewClusterPtrOutput) Autoscale() JobNewClusterAutoscalePtrOutput { }).(JobNewClusterAutoscalePtrOutput) } -func (o JobNewClusterPtrOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v *JobNewCluster) *int { - if v == nil { - return nil - } - return v.AutoterminationMinutes - }).(pulumi.IntPtrOutput) -} - func (o JobNewClusterPtrOutput) AwsAttributes() JobNewClusterAwsAttributesPtrOutput { return o.ApplyT(func(v *JobNewCluster) *JobNewClusterAwsAttributes { if v == nil { @@ -22579,7 +22203,7 @@ type JobTask struct { SqlTask *JobTaskSqlTask `pulumi:"sqlTask"` // string specifying an unique key for a given task. // * `*_task` - (Required) one of the specific task blocks described below: - TaskKey *string `pulumi:"taskKey"` + TaskKey string `pulumi:"taskKey"` // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. TimeoutSeconds *int `pulumi:"timeoutSeconds"` // (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. @@ -22641,7 +22265,7 @@ type JobTaskArgs struct { SqlTask JobTaskSqlTaskPtrInput `pulumi:"sqlTask"` // string specifying an unique key for a given task. // * `*_task` - (Required) one of the specific task blocks described below: - TaskKey pulumi.StringPtrInput `pulumi:"taskKey"` + TaskKey pulumi.StringInput `pulumi:"taskKey"` // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. TimeoutSeconds pulumi.IntPtrInput `pulumi:"timeoutSeconds"` // (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. @@ -22820,8 +22444,8 @@ func (o JobTaskOutput) SqlTask() JobTaskSqlTaskPtrOutput { // string specifying an unique key for a given task. // * `*_task` - (Required) one of the specific task blocks described below: -func (o JobTaskOutput) TaskKey() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTask) *string { return v.TaskKey }).(pulumi.StringPtrOutput) +func (o JobTaskOutput) TaskKey() pulumi.StringOutput { + return o.ApplyT(func(v JobTask) string { return v.TaskKey }).(pulumi.StringOutput) } // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. @@ -23846,7 +23470,7 @@ type JobTaskForEachTaskTask struct { SqlTask *JobTaskForEachTaskTaskSqlTask `pulumi:"sqlTask"` // string specifying an unique key for a given task. // * `*_task` - (Required) one of the specific task blocks described below: - TaskKey *string `pulumi:"taskKey"` + TaskKey string `pulumi:"taskKey"` // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. TimeoutSeconds *int `pulumi:"timeoutSeconds"` // (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. @@ -23907,7 +23531,7 @@ type JobTaskForEachTaskTaskArgs struct { SqlTask JobTaskForEachTaskTaskSqlTaskPtrInput `pulumi:"sqlTask"` // string specifying an unique key for a given task. // * `*_task` - (Required) one of the specific task blocks described below: - TaskKey pulumi.StringPtrInput `pulumi:"taskKey"` + TaskKey pulumi.StringInput `pulumi:"taskKey"` // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. TimeoutSeconds pulumi.IntPtrInput `pulumi:"timeoutSeconds"` // (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. @@ -24110,8 +23734,8 @@ func (o JobTaskForEachTaskTaskOutput) SqlTask() JobTaskForEachTaskTaskSqlTaskPtr // string specifying an unique key for a given task. // * `*_task` - (Required) one of the specific task blocks described below: -func (o JobTaskForEachTaskTaskOutput) TaskKey() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTaskForEachTaskTask) *string { return v.TaskKey }).(pulumi.StringPtrOutput) +func (o JobTaskForEachTaskTaskOutput) TaskKey() pulumi.StringOutput { + return o.ApplyT(func(v JobTaskForEachTaskTask) string { return v.TaskKey }).(pulumi.StringOutput) } // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. @@ -24397,7 +24021,7 @@ func (o JobTaskForEachTaskTaskPtrOutput) TaskKey() pulumi.StringPtrOutput { if v == nil { return nil } - return v.TaskKey + return &v.TaskKey }).(pulumi.StringPtrOutput) } @@ -25336,11 +24960,11 @@ func (o JobTaskForEachTaskTaskHealthPtrOutput) Rules() JobTaskForEachTaskTaskHea type JobTaskForEachTaskTaskHealthRule struct { // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). - Metric *string `pulumi:"metric"` + Metric string `pulumi:"metric"` // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. - Op *string `pulumi:"op"` + Op string `pulumi:"op"` // integer value used to compare to the given metric. - Value *int `pulumi:"value"` + Value int `pulumi:"value"` } // JobTaskForEachTaskTaskHealthRuleInput is an input type that accepts JobTaskForEachTaskTaskHealthRuleArgs and JobTaskForEachTaskTaskHealthRuleOutput values. @@ -25356,11 +24980,11 @@ type JobTaskForEachTaskTaskHealthRuleInput interface { type JobTaskForEachTaskTaskHealthRuleArgs struct { // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). - Metric pulumi.StringPtrInput `pulumi:"metric"` + Metric pulumi.StringInput `pulumi:"metric"` // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. - Op pulumi.StringPtrInput `pulumi:"op"` + Op pulumi.StringInput `pulumi:"op"` // integer value used to compare to the given metric. - Value pulumi.IntPtrInput `pulumi:"value"` + Value pulumi.IntInput `pulumi:"value"` } func (JobTaskForEachTaskTaskHealthRuleArgs) ElementType() reflect.Type { @@ -25415,18 +25039,18 @@ func (o JobTaskForEachTaskTaskHealthRuleOutput) ToJobTaskForEachTaskTaskHealthRu } // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). -func (o JobTaskForEachTaskTaskHealthRuleOutput) Metric() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTaskForEachTaskTaskHealthRule) *string { return v.Metric }).(pulumi.StringPtrOutput) +func (o JobTaskForEachTaskTaskHealthRuleOutput) Metric() pulumi.StringOutput { + return o.ApplyT(func(v JobTaskForEachTaskTaskHealthRule) string { return v.Metric }).(pulumi.StringOutput) } // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. -func (o JobTaskForEachTaskTaskHealthRuleOutput) Op() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTaskForEachTaskTaskHealthRule) *string { return v.Op }).(pulumi.StringPtrOutput) +func (o JobTaskForEachTaskTaskHealthRuleOutput) Op() pulumi.StringOutput { + return o.ApplyT(func(v JobTaskForEachTaskTaskHealthRule) string { return v.Op }).(pulumi.StringOutput) } // integer value used to compare to the given metric. -func (o JobTaskForEachTaskTaskHealthRuleOutput) Value() pulumi.IntPtrOutput { - return o.ApplyT(func(v JobTaskForEachTaskTaskHealthRule) *int { return v.Value }).(pulumi.IntPtrOutput) +func (o JobTaskForEachTaskTaskHealthRuleOutput) Value() pulumi.IntOutput { + return o.ApplyT(func(v JobTaskForEachTaskTaskHealthRule) int { return v.Value }).(pulumi.IntOutput) } type JobTaskForEachTaskTaskHealthRuleArrayOutput struct{ *pulumi.OutputState } @@ -26041,7 +25665,6 @@ func (o JobTaskForEachTaskTaskLibraryPypiPtrOutput) Repo() pulumi.StringPtrOutpu type JobTaskForEachTaskTaskNewCluster struct { ApplyPolicyDefaultValues *bool `pulumi:"applyPolicyDefaultValues"` Autoscale *JobTaskForEachTaskTaskNewClusterAutoscale `pulumi:"autoscale"` - AutoterminationMinutes *int `pulumi:"autoterminationMinutes"` AwsAttributes *JobTaskForEachTaskTaskNewClusterAwsAttributes `pulumi:"awsAttributes"` AzureAttributes *JobTaskForEachTaskTaskNewClusterAzureAttributes `pulumi:"azureAttributes"` ClusterId *string `pulumi:"clusterId"` @@ -26087,7 +25710,6 @@ type JobTaskForEachTaskTaskNewClusterInput interface { type JobTaskForEachTaskTaskNewClusterArgs struct { ApplyPolicyDefaultValues pulumi.BoolPtrInput `pulumi:"applyPolicyDefaultValues"` Autoscale JobTaskForEachTaskTaskNewClusterAutoscalePtrInput `pulumi:"autoscale"` - AutoterminationMinutes pulumi.IntPtrInput `pulumi:"autoterminationMinutes"` AwsAttributes JobTaskForEachTaskTaskNewClusterAwsAttributesPtrInput `pulumi:"awsAttributes"` AzureAttributes JobTaskForEachTaskTaskNewClusterAzureAttributesPtrInput `pulumi:"azureAttributes"` ClusterId pulumi.StringPtrInput `pulumi:"clusterId"` @@ -26206,10 +25828,6 @@ func (o JobTaskForEachTaskTaskNewClusterOutput) Autoscale() JobTaskForEachTaskTa }).(JobTaskForEachTaskTaskNewClusterAutoscalePtrOutput) } -func (o JobTaskForEachTaskTaskNewClusterOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v JobTaskForEachTaskTaskNewCluster) *int { return v.AutoterminationMinutes }).(pulumi.IntPtrOutput) -} - func (o JobTaskForEachTaskTaskNewClusterOutput) AwsAttributes() JobTaskForEachTaskTaskNewClusterAwsAttributesPtrOutput { return o.ApplyT(func(v JobTaskForEachTaskTaskNewCluster) *JobTaskForEachTaskTaskNewClusterAwsAttributes { return v.AwsAttributes @@ -26381,15 +25999,6 @@ func (o JobTaskForEachTaskTaskNewClusterPtrOutput) Autoscale() JobTaskForEachTas }).(JobTaskForEachTaskTaskNewClusterAutoscalePtrOutput) } -func (o JobTaskForEachTaskTaskNewClusterPtrOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v *JobTaskForEachTaskTaskNewCluster) *int { - if v == nil { - return nil - } - return v.AutoterminationMinutes - }).(pulumi.IntPtrOutput) -} - func (o JobTaskForEachTaskTaskNewClusterPtrOutput) AwsAttributes() JobTaskForEachTaskTaskNewClusterAwsAttributesPtrOutput { return o.ApplyT(func(v *JobTaskForEachTaskTaskNewCluster) *JobTaskForEachTaskTaskNewClusterAwsAttributes { if v == nil { @@ -32263,7 +31872,7 @@ type JobTaskForEachTaskTaskSqlTask struct { // block consisting of single string field: `queryId` - identifier of the Databricks SQL Query (databricks_sql_query). Query *JobTaskForEachTaskTaskSqlTaskQuery `pulumi:"query"` // ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - WarehouseId *string `pulumi:"warehouseId"` + WarehouseId string `pulumi:"warehouseId"` } // JobTaskForEachTaskTaskSqlTaskInput is an input type that accepts JobTaskForEachTaskTaskSqlTaskArgs and JobTaskForEachTaskTaskSqlTaskOutput values. @@ -32289,7 +31898,7 @@ type JobTaskForEachTaskTaskSqlTaskArgs struct { // block consisting of single string field: `queryId` - identifier of the Databricks SQL Query (databricks_sql_query). Query JobTaskForEachTaskTaskSqlTaskQueryPtrInput `pulumi:"query"` // ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - WarehouseId pulumi.StringPtrInput `pulumi:"warehouseId"` + WarehouseId pulumi.StringInput `pulumi:"warehouseId"` } func (JobTaskForEachTaskTaskSqlTaskArgs) ElementType() reflect.Type { @@ -32395,8 +32004,8 @@ func (o JobTaskForEachTaskTaskSqlTaskOutput) Query() JobTaskForEachTaskTaskSqlTa } // ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. -func (o JobTaskForEachTaskTaskSqlTaskOutput) WarehouseId() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTaskForEachTaskTaskSqlTask) *string { return v.WarehouseId }).(pulumi.StringPtrOutput) +func (o JobTaskForEachTaskTaskSqlTaskOutput) WarehouseId() pulumi.StringOutput { + return o.ApplyT(func(v JobTaskForEachTaskTaskSqlTask) string { return v.WarehouseId }).(pulumi.StringOutput) } type JobTaskForEachTaskTaskSqlTaskPtrOutput struct{ *pulumi.OutputState } @@ -32479,7 +32088,7 @@ func (o JobTaskForEachTaskTaskSqlTaskPtrOutput) WarehouseId() pulumi.StringPtrOu if v == nil { return nil } - return v.WarehouseId + return &v.WarehouseId }).(pulumi.StringPtrOutput) } @@ -34346,11 +33955,11 @@ func (o JobTaskHealthPtrOutput) Rules() JobTaskHealthRuleArrayOutput { type JobTaskHealthRule struct { // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). - Metric *string `pulumi:"metric"` + Metric string `pulumi:"metric"` // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. - Op *string `pulumi:"op"` + Op string `pulumi:"op"` // integer value used to compare to the given metric. - Value *int `pulumi:"value"` + Value int `pulumi:"value"` } // JobTaskHealthRuleInput is an input type that accepts JobTaskHealthRuleArgs and JobTaskHealthRuleOutput values. @@ -34366,11 +33975,11 @@ type JobTaskHealthRuleInput interface { type JobTaskHealthRuleArgs struct { // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). - Metric pulumi.StringPtrInput `pulumi:"metric"` + Metric pulumi.StringInput `pulumi:"metric"` // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. - Op pulumi.StringPtrInput `pulumi:"op"` + Op pulumi.StringInput `pulumi:"op"` // integer value used to compare to the given metric. - Value pulumi.IntPtrInput `pulumi:"value"` + Value pulumi.IntInput `pulumi:"value"` } func (JobTaskHealthRuleArgs) ElementType() reflect.Type { @@ -34425,18 +34034,18 @@ func (o JobTaskHealthRuleOutput) ToJobTaskHealthRuleOutputWithContext(ctx contex } // string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). -func (o JobTaskHealthRuleOutput) Metric() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTaskHealthRule) *string { return v.Metric }).(pulumi.StringPtrOutput) +func (o JobTaskHealthRuleOutput) Metric() pulumi.StringOutput { + return o.ApplyT(func(v JobTaskHealthRule) string { return v.Metric }).(pulumi.StringOutput) } // string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. -func (o JobTaskHealthRuleOutput) Op() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTaskHealthRule) *string { return v.Op }).(pulumi.StringPtrOutput) +func (o JobTaskHealthRuleOutput) Op() pulumi.StringOutput { + return o.ApplyT(func(v JobTaskHealthRule) string { return v.Op }).(pulumi.StringOutput) } // integer value used to compare to the given metric. -func (o JobTaskHealthRuleOutput) Value() pulumi.IntPtrOutput { - return o.ApplyT(func(v JobTaskHealthRule) *int { return v.Value }).(pulumi.IntPtrOutput) +func (o JobTaskHealthRuleOutput) Value() pulumi.IntOutput { + return o.ApplyT(func(v JobTaskHealthRule) int { return v.Value }).(pulumi.IntOutput) } type JobTaskHealthRuleArrayOutput struct{ *pulumi.OutputState } @@ -35051,7 +34660,6 @@ func (o JobTaskLibraryPypiPtrOutput) Repo() pulumi.StringPtrOutput { type JobTaskNewCluster struct { ApplyPolicyDefaultValues *bool `pulumi:"applyPolicyDefaultValues"` Autoscale *JobTaskNewClusterAutoscale `pulumi:"autoscale"` - AutoterminationMinutes *int `pulumi:"autoterminationMinutes"` AwsAttributes *JobTaskNewClusterAwsAttributes `pulumi:"awsAttributes"` AzureAttributes *JobTaskNewClusterAzureAttributes `pulumi:"azureAttributes"` ClusterId *string `pulumi:"clusterId"` @@ -35097,7 +34705,6 @@ type JobTaskNewClusterInput interface { type JobTaskNewClusterArgs struct { ApplyPolicyDefaultValues pulumi.BoolPtrInput `pulumi:"applyPolicyDefaultValues"` Autoscale JobTaskNewClusterAutoscalePtrInput `pulumi:"autoscale"` - AutoterminationMinutes pulumi.IntPtrInput `pulumi:"autoterminationMinutes"` AwsAttributes JobTaskNewClusterAwsAttributesPtrInput `pulumi:"awsAttributes"` AzureAttributes JobTaskNewClusterAzureAttributesPtrInput `pulumi:"azureAttributes"` ClusterId pulumi.StringPtrInput `pulumi:"clusterId"` @@ -35214,10 +34821,6 @@ func (o JobTaskNewClusterOutput) Autoscale() JobTaskNewClusterAutoscalePtrOutput return o.ApplyT(func(v JobTaskNewCluster) *JobTaskNewClusterAutoscale { return v.Autoscale }).(JobTaskNewClusterAutoscalePtrOutput) } -func (o JobTaskNewClusterOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v JobTaskNewCluster) *int { return v.AutoterminationMinutes }).(pulumi.IntPtrOutput) -} - func (o JobTaskNewClusterOutput) AwsAttributes() JobTaskNewClusterAwsAttributesPtrOutput { return o.ApplyT(func(v JobTaskNewCluster) *JobTaskNewClusterAwsAttributes { return v.AwsAttributes }).(JobTaskNewClusterAwsAttributesPtrOutput) } @@ -35373,15 +34976,6 @@ func (o JobTaskNewClusterPtrOutput) Autoscale() JobTaskNewClusterAutoscalePtrOut }).(JobTaskNewClusterAutoscalePtrOutput) } -func (o JobTaskNewClusterPtrOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v *JobTaskNewCluster) *int { - if v == nil { - return nil - } - return v.AutoterminationMinutes - }).(pulumi.IntPtrOutput) -} - func (o JobTaskNewClusterPtrOutput) AwsAttributes() JobTaskNewClusterAwsAttributesPtrOutput { return o.ApplyT(func(v *JobTaskNewCluster) *JobTaskNewClusterAwsAttributes { if v == nil { @@ -41217,7 +40811,7 @@ type JobTaskSqlTask struct { // block consisting of single string field: `queryId` - identifier of the Databricks SQL Query (databricks_sql_query). Query *JobTaskSqlTaskQuery `pulumi:"query"` // ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - WarehouseId *string `pulumi:"warehouseId"` + WarehouseId string `pulumi:"warehouseId"` } // JobTaskSqlTaskInput is an input type that accepts JobTaskSqlTaskArgs and JobTaskSqlTaskOutput values. @@ -41243,7 +40837,7 @@ type JobTaskSqlTaskArgs struct { // block consisting of single string field: `queryId` - identifier of the Databricks SQL Query (databricks_sql_query). Query JobTaskSqlTaskQueryPtrInput `pulumi:"query"` // ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - WarehouseId pulumi.StringPtrInput `pulumi:"warehouseId"` + WarehouseId pulumi.StringInput `pulumi:"warehouseId"` } func (JobTaskSqlTaskArgs) ElementType() reflect.Type { @@ -41349,8 +40943,8 @@ func (o JobTaskSqlTaskOutput) Query() JobTaskSqlTaskQueryPtrOutput { } // ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. -func (o JobTaskSqlTaskOutput) WarehouseId() pulumi.StringPtrOutput { - return o.ApplyT(func(v JobTaskSqlTask) *string { return v.WarehouseId }).(pulumi.StringPtrOutput) +func (o JobTaskSqlTaskOutput) WarehouseId() pulumi.StringOutput { + return o.ApplyT(func(v JobTaskSqlTask) string { return v.WarehouseId }).(pulumi.StringOutput) } type JobTaskSqlTaskPtrOutput struct{ *pulumi.OutputState } @@ -41433,7 +41027,7 @@ func (o JobTaskSqlTaskPtrOutput) WarehouseId() pulumi.StringPtrOutput { if v == nil { return nil } - return v.WarehouseId + return &v.WarehouseId }).(pulumi.StringPtrOutput) } @@ -43153,9 +42747,8 @@ type JobTrigger struct { // configuration block to define a trigger for [File Arrival events](https://learn.microsoft.com/en-us/azure/databricks/workflows/jobs/file-arrival-triggers) consisting of following attributes: FileArrival *JobTriggerFileArrival `pulumi:"fileArrival"` // Indicate whether this trigger is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pauseStatus` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pauseStatus`. - PauseStatus *string `pulumi:"pauseStatus"` - Table *JobTriggerTable `pulumi:"table"` - // configuration block to define a trigger for Table Update events consisting of following attributes: + PauseStatus *string `pulumi:"pauseStatus"` + Table *JobTriggerTable `pulumi:"table"` TableUpdate *JobTriggerTableUpdate `pulumi:"tableUpdate"` } @@ -43174,9 +42767,8 @@ type JobTriggerArgs struct { // configuration block to define a trigger for [File Arrival events](https://learn.microsoft.com/en-us/azure/databricks/workflows/jobs/file-arrival-triggers) consisting of following attributes: FileArrival JobTriggerFileArrivalPtrInput `pulumi:"fileArrival"` // Indicate whether this trigger is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pauseStatus` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pauseStatus`. - PauseStatus pulumi.StringPtrInput `pulumi:"pauseStatus"` - Table JobTriggerTablePtrInput `pulumi:"table"` - // configuration block to define a trigger for Table Update events consisting of following attributes: + PauseStatus pulumi.StringPtrInput `pulumi:"pauseStatus"` + Table JobTriggerTablePtrInput `pulumi:"table"` TableUpdate JobTriggerTableUpdatePtrInput `pulumi:"tableUpdate"` } @@ -43271,7 +42863,6 @@ func (o JobTriggerOutput) Table() JobTriggerTablePtrOutput { return o.ApplyT(func(v JobTrigger) *JobTriggerTable { return v.Table }).(JobTriggerTablePtrOutput) } -// configuration block to define a trigger for Table Update events consisting of following attributes: func (o JobTriggerOutput) TableUpdate() JobTriggerTableUpdatePtrOutput { return o.ApplyT(func(v JobTrigger) *JobTriggerTableUpdate { return v.TableUpdate }).(JobTriggerTableUpdatePtrOutput) } @@ -43329,7 +42920,6 @@ func (o JobTriggerPtrOutput) Table() JobTriggerTablePtrOutput { }).(JobTriggerTablePtrOutput) } -// configuration block to define a trigger for Table Update events consisting of following attributes: func (o JobTriggerPtrOutput) TableUpdate() JobTriggerTableUpdatePtrOutput { return o.ApplyT(func(v *JobTrigger) *JobTriggerTableUpdate { if v == nil { @@ -43693,14 +43283,10 @@ func (o JobTriggerTablePtrOutput) WaitAfterLastChangeSeconds() pulumi.IntPtrOutp } type JobTriggerTableUpdate struct { - // The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - Condition *string `pulumi:"condition"` - // If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - MinTimeBetweenTriggersSeconds *int `pulumi:"minTimeBetweenTriggersSeconds"` - // A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - TableNames []string `pulumi:"tableNames"` - // If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - WaitAfterLastChangeSeconds *int `pulumi:"waitAfterLastChangeSeconds"` + Condition *string `pulumi:"condition"` + MinTimeBetweenTriggersSeconds *int `pulumi:"minTimeBetweenTriggersSeconds"` + TableNames []string `pulumi:"tableNames"` + WaitAfterLastChangeSeconds *int `pulumi:"waitAfterLastChangeSeconds"` } // JobTriggerTableUpdateInput is an input type that accepts JobTriggerTableUpdateArgs and JobTriggerTableUpdateOutput values. @@ -43715,14 +43301,10 @@ type JobTriggerTableUpdateInput interface { } type JobTriggerTableUpdateArgs struct { - // The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - Condition pulumi.StringPtrInput `pulumi:"condition"` - // If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - MinTimeBetweenTriggersSeconds pulumi.IntPtrInput `pulumi:"minTimeBetweenTriggersSeconds"` - // A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - TableNames pulumi.StringArrayInput `pulumi:"tableNames"` - // If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - WaitAfterLastChangeSeconds pulumi.IntPtrInput `pulumi:"waitAfterLastChangeSeconds"` + Condition pulumi.StringPtrInput `pulumi:"condition"` + MinTimeBetweenTriggersSeconds pulumi.IntPtrInput `pulumi:"minTimeBetweenTriggersSeconds"` + TableNames pulumi.StringArrayInput `pulumi:"tableNames"` + WaitAfterLastChangeSeconds pulumi.IntPtrInput `pulumi:"waitAfterLastChangeSeconds"` } func (JobTriggerTableUpdateArgs) ElementType() reflect.Type { @@ -43802,22 +43384,18 @@ func (o JobTriggerTableUpdateOutput) ToJobTriggerTableUpdatePtrOutputWithContext }).(JobTriggerTableUpdatePtrOutput) } -// The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. func (o JobTriggerTableUpdateOutput) Condition() pulumi.StringPtrOutput { return o.ApplyT(func(v JobTriggerTableUpdate) *string { return v.Condition }).(pulumi.StringPtrOutput) } -// If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. func (o JobTriggerTableUpdateOutput) MinTimeBetweenTriggersSeconds() pulumi.IntPtrOutput { return o.ApplyT(func(v JobTriggerTableUpdate) *int { return v.MinTimeBetweenTriggersSeconds }).(pulumi.IntPtrOutput) } -// A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. func (o JobTriggerTableUpdateOutput) TableNames() pulumi.StringArrayOutput { return o.ApplyT(func(v JobTriggerTableUpdate) []string { return v.TableNames }).(pulumi.StringArrayOutput) } -// If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. func (o JobTriggerTableUpdateOutput) WaitAfterLastChangeSeconds() pulumi.IntPtrOutput { return o.ApplyT(func(v JobTriggerTableUpdate) *int { return v.WaitAfterLastChangeSeconds }).(pulumi.IntPtrOutput) } @@ -43846,7 +43424,6 @@ func (o JobTriggerTableUpdatePtrOutput) Elem() JobTriggerTableUpdateOutput { }).(JobTriggerTableUpdateOutput) } -// The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. func (o JobTriggerTableUpdatePtrOutput) Condition() pulumi.StringPtrOutput { return o.ApplyT(func(v *JobTriggerTableUpdate) *string { if v == nil { @@ -43856,7 +43433,6 @@ func (o JobTriggerTableUpdatePtrOutput) Condition() pulumi.StringPtrOutput { }).(pulumi.StringPtrOutput) } -// If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. func (o JobTriggerTableUpdatePtrOutput) MinTimeBetweenTriggersSeconds() pulumi.IntPtrOutput { return o.ApplyT(func(v *JobTriggerTableUpdate) *int { if v == nil { @@ -43866,7 +43442,6 @@ func (o JobTriggerTableUpdatePtrOutput) MinTimeBetweenTriggersSeconds() pulumi.I }).(pulumi.IntPtrOutput) } -// A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. func (o JobTriggerTableUpdatePtrOutput) TableNames() pulumi.StringArrayOutput { return o.ApplyT(func(v *JobTriggerTableUpdate) []string { if v == nil { @@ -43876,7 +43451,6 @@ func (o JobTriggerTableUpdatePtrOutput) TableNames() pulumi.StringArrayOutput { }).(pulumi.StringArrayOutput) } -// If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. func (o JobTriggerTableUpdatePtrOutput) WaitAfterLastChangeSeconds() pulumi.IntPtrOutput { return o.ApplyT(func(v *JobTriggerTableUpdate) *int { if v == nil { @@ -80125,9 +79699,9 @@ func (o GetJobJobSettingsSettingsHealthPtrOutput) Rules() GetJobJobSettingsSetti } type GetJobJobSettingsSettingsHealthRule struct { - Metric *string `pulumi:"metric"` - Op *string `pulumi:"op"` - Value *int `pulumi:"value"` + Metric string `pulumi:"metric"` + Op string `pulumi:"op"` + Value int `pulumi:"value"` } // GetJobJobSettingsSettingsHealthRuleInput is an input type that accepts GetJobJobSettingsSettingsHealthRuleArgs and GetJobJobSettingsSettingsHealthRuleOutput values. @@ -80142,9 +79716,9 @@ type GetJobJobSettingsSettingsHealthRuleInput interface { } type GetJobJobSettingsSettingsHealthRuleArgs struct { - Metric pulumi.StringPtrInput `pulumi:"metric"` - Op pulumi.StringPtrInput `pulumi:"op"` - Value pulumi.IntPtrInput `pulumi:"value"` + Metric pulumi.StringInput `pulumi:"metric"` + Op pulumi.StringInput `pulumi:"op"` + Value pulumi.IntInput `pulumi:"value"` } func (GetJobJobSettingsSettingsHealthRuleArgs) ElementType() reflect.Type { @@ -80198,16 +79772,16 @@ func (o GetJobJobSettingsSettingsHealthRuleOutput) ToGetJobJobSettingsSettingsHe return o } -func (o GetJobJobSettingsSettingsHealthRuleOutput) Metric() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsHealthRule) *string { return v.Metric }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsHealthRuleOutput) Metric() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsHealthRule) string { return v.Metric }).(pulumi.StringOutput) } -func (o GetJobJobSettingsSettingsHealthRuleOutput) Op() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsHealthRule) *string { return v.Op }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsHealthRuleOutput) Op() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsHealthRule) string { return v.Op }).(pulumi.StringOutput) } -func (o GetJobJobSettingsSettingsHealthRuleOutput) Value() pulumi.IntPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsHealthRule) *int { return v.Value }).(pulumi.IntPtrOutput) +func (o GetJobJobSettingsSettingsHealthRuleOutput) Value() pulumi.IntOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsHealthRule) int { return v.Value }).(pulumi.IntOutput) } type GetJobJobSettingsSettingsHealthRuleArrayOutput struct{ *pulumi.OutputState } @@ -80231,8 +79805,8 @@ func (o GetJobJobSettingsSettingsHealthRuleArrayOutput) Index(i pulumi.IntInput) } type GetJobJobSettingsSettingsJobCluster struct { - JobClusterKey *string `pulumi:"jobClusterKey"` - NewCluster *GetJobJobSettingsSettingsJobClusterNewCluster `pulumi:"newCluster"` + JobClusterKey string `pulumi:"jobClusterKey"` + NewCluster GetJobJobSettingsSettingsJobClusterNewCluster `pulumi:"newCluster"` } // GetJobJobSettingsSettingsJobClusterInput is an input type that accepts GetJobJobSettingsSettingsJobClusterArgs and GetJobJobSettingsSettingsJobClusterOutput values. @@ -80247,8 +79821,8 @@ type GetJobJobSettingsSettingsJobClusterInput interface { } type GetJobJobSettingsSettingsJobClusterArgs struct { - JobClusterKey pulumi.StringPtrInput `pulumi:"jobClusterKey"` - NewCluster GetJobJobSettingsSettingsJobClusterNewClusterPtrInput `pulumi:"newCluster"` + JobClusterKey pulumi.StringInput `pulumi:"jobClusterKey"` + NewCluster GetJobJobSettingsSettingsJobClusterNewClusterInput `pulumi:"newCluster"` } func (GetJobJobSettingsSettingsJobClusterArgs) ElementType() reflect.Type { @@ -80302,14 +79876,14 @@ func (o GetJobJobSettingsSettingsJobClusterOutput) ToGetJobJobSettingsSettingsJo return o } -func (o GetJobJobSettingsSettingsJobClusterOutput) JobClusterKey() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsJobCluster) *string { return v.JobClusterKey }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsJobClusterOutput) JobClusterKey() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsJobCluster) string { return v.JobClusterKey }).(pulumi.StringOutput) } -func (o GetJobJobSettingsSettingsJobClusterOutput) NewCluster() GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsJobCluster) *GetJobJobSettingsSettingsJobClusterNewCluster { +func (o GetJobJobSettingsSettingsJobClusterOutput) NewCluster() GetJobJobSettingsSettingsJobClusterNewClusterOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsJobCluster) GetJobJobSettingsSettingsJobClusterNewCluster { return v.NewCluster - }).(GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) + }).(GetJobJobSettingsSettingsJobClusterNewClusterOutput) } type GetJobJobSettingsSettingsJobClusterArrayOutput struct{ *pulumi.OutputState } @@ -80421,47 +79995,6 @@ func (i GetJobJobSettingsSettingsJobClusterNewClusterArgs) ToGetJobJobSettingsSe return pulumi.ToOutputWithContext(ctx, i).(GetJobJobSettingsSettingsJobClusterNewClusterOutput) } -func (i GetJobJobSettingsSettingsJobClusterNewClusterArgs) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutput() GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return i.ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(context.Background()) -} - -func (i GetJobJobSettingsSettingsJobClusterNewClusterArgs) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(ctx context.Context) GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return pulumi.ToOutputWithContext(ctx, i).(GetJobJobSettingsSettingsJobClusterNewClusterOutput).ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(ctx) -} - -// GetJobJobSettingsSettingsJobClusterNewClusterPtrInput is an input type that accepts GetJobJobSettingsSettingsJobClusterNewClusterArgs, GetJobJobSettingsSettingsJobClusterNewClusterPtr and GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput values. -// You can construct a concrete instance of `GetJobJobSettingsSettingsJobClusterNewClusterPtrInput` via: -// -// GetJobJobSettingsSettingsJobClusterNewClusterArgs{...} -// -// or: -// -// nil -type GetJobJobSettingsSettingsJobClusterNewClusterPtrInput interface { - pulumi.Input - - ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutput() GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput - ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(context.Context) GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput -} - -type getJobJobSettingsSettingsJobClusterNewClusterPtrType GetJobJobSettingsSettingsJobClusterNewClusterArgs - -func GetJobJobSettingsSettingsJobClusterNewClusterPtr(v *GetJobJobSettingsSettingsJobClusterNewClusterArgs) GetJobJobSettingsSettingsJobClusterNewClusterPtrInput { - return (*getJobJobSettingsSettingsJobClusterNewClusterPtrType)(v) -} - -func (*getJobJobSettingsSettingsJobClusterNewClusterPtrType) ElementType() reflect.Type { - return reflect.TypeOf((**GetJobJobSettingsSettingsJobClusterNewCluster)(nil)).Elem() -} - -func (i *getJobJobSettingsSettingsJobClusterNewClusterPtrType) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutput() GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return i.ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(context.Background()) -} - -func (i *getJobJobSettingsSettingsJobClusterNewClusterPtrType) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(ctx context.Context) GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return pulumi.ToOutputWithContext(ctx, i).(GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) -} - type GetJobJobSettingsSettingsJobClusterNewClusterOutput struct{ *pulumi.OutputState } func (GetJobJobSettingsSettingsJobClusterNewClusterOutput) ElementType() reflect.Type { @@ -80476,16 +80009,6 @@ func (o GetJobJobSettingsSettingsJobClusterNewClusterOutput) ToGetJobJobSettings return o } -func (o GetJobJobSettingsSettingsJobClusterNewClusterOutput) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutput() GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return o.ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(context.Background()) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterOutput) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(ctx context.Context) GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return o.ApplyTWithContext(ctx, func(_ context.Context, v GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewCluster { - return &v - }).(GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) -} - func (o GetJobJobSettingsSettingsJobClusterNewClusterOutput) ApplyPolicyDefaultValues() pulumi.BoolPtrOutput { return o.ApplyT(func(v GetJobJobSettingsSettingsJobClusterNewCluster) *bool { return v.ApplyPolicyDefaultValues }).(pulumi.BoolPtrOutput) } @@ -80624,300 +80147,6 @@ func (o GetJobJobSettingsSettingsJobClusterNewClusterOutput) WorkloadType() GetJ }).(GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypePtrOutput) } -type GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput struct{ *pulumi.OutputState } - -func (GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ElementType() reflect.Type { - return reflect.TypeOf((**GetJobJobSettingsSettingsJobClusterNewCluster)(nil)).Elem() -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutput() GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return o -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ToGetJobJobSettingsSettingsJobClusterNewClusterPtrOutputWithContext(ctx context.Context) GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput { - return o -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) Elem() GetJobJobSettingsSettingsJobClusterNewClusterOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) GetJobJobSettingsSettingsJobClusterNewCluster { - if v != nil { - return *v - } - var ret GetJobJobSettingsSettingsJobClusterNewCluster - return ret - }).(GetJobJobSettingsSettingsJobClusterNewClusterOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ApplyPolicyDefaultValues() pulumi.BoolPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *bool { - if v == nil { - return nil - } - return v.ApplyPolicyDefaultValues - }).(pulumi.BoolPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) Autoscale() GetJobJobSettingsSettingsJobClusterNewClusterAutoscalePtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewClusterAutoscale { - if v == nil { - return nil - } - return v.Autoscale - }).(GetJobJobSettingsSettingsJobClusterNewClusterAutoscalePtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) AutoterminationMinutes() pulumi.IntPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *int { - if v == nil { - return nil - } - return v.AutoterminationMinutes - }).(pulumi.IntPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) AwsAttributes() GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributes { - if v == nil { - return nil - } - return v.AwsAttributes - }).(GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) AzureAttributes() GetJobJobSettingsSettingsJobClusterNewClusterAzureAttributesPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewClusterAzureAttributes { - if v == nil { - return nil - } - return v.AzureAttributes - }).(GetJobJobSettingsSettingsJobClusterNewClusterAzureAttributesPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ClusterId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.ClusterId - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ClusterLogConf() GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConf { - if v == nil { - return nil - } - return v.ClusterLogConf - }).(GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ClusterMountInfos() GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoArrayOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) []GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfo { - if v == nil { - return nil - } - return v.ClusterMountInfos - }).(GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoArrayOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) ClusterName() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.ClusterName - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) CustomTags() pulumi.MapOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) map[string]interface{} { - if v == nil { - return nil - } - return v.CustomTags - }).(pulumi.MapOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) DataSecurityMode() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.DataSecurityMode - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) DockerImage() GetJobJobSettingsSettingsJobClusterNewClusterDockerImagePtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewClusterDockerImage { - if v == nil { - return nil - } - return v.DockerImage - }).(GetJobJobSettingsSettingsJobClusterNewClusterDockerImagePtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) DriverInstancePoolId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return &v.DriverInstancePoolId - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) DriverNodeTypeId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return &v.DriverNodeTypeId - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) EnableElasticDisk() pulumi.BoolPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *bool { - if v == nil { - return nil - } - return &v.EnableElasticDisk - }).(pulumi.BoolPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) EnableLocalDiskEncryption() pulumi.BoolPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *bool { - if v == nil { - return nil - } - return &v.EnableLocalDiskEncryption - }).(pulumi.BoolPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) GcpAttributes() GetJobJobSettingsSettingsJobClusterNewClusterGcpAttributesPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewClusterGcpAttributes { - if v == nil { - return nil - } - return v.GcpAttributes - }).(GetJobJobSettingsSettingsJobClusterNewClusterGcpAttributesPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) IdempotencyToken() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.IdempotencyToken - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) InitScripts() GetJobJobSettingsSettingsJobClusterNewClusterInitScriptArrayOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) []GetJobJobSettingsSettingsJobClusterNewClusterInitScript { - if v == nil { - return nil - } - return v.InitScripts - }).(GetJobJobSettingsSettingsJobClusterNewClusterInitScriptArrayOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) InstancePoolId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.InstancePoolId - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) NodeTypeId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return &v.NodeTypeId - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) NumWorkers() pulumi.IntPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *int { - if v == nil { - return nil - } - return &v.NumWorkers - }).(pulumi.IntPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) PolicyId() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.PolicyId - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) RuntimeEngine() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.RuntimeEngine - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) SingleUserName() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return v.SingleUserName - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) SparkConf() pulumi.MapOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) map[string]interface{} { - if v == nil { - return nil - } - return v.SparkConf - }).(pulumi.MapOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) SparkEnvVars() pulumi.MapOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) map[string]interface{} { - if v == nil { - return nil - } - return v.SparkEnvVars - }).(pulumi.MapOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) SparkVersion() pulumi.StringPtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *string { - if v == nil { - return nil - } - return &v.SparkVersion - }).(pulumi.StringPtrOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) SshPublicKeys() pulumi.StringArrayOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) []string { - if v == nil { - return nil - } - return v.SshPublicKeys - }).(pulumi.StringArrayOutput) -} - -func (o GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput) WorkloadType() GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypePtrOutput { - return o.ApplyT(func(v *GetJobJobSettingsSettingsJobClusterNewCluster) *GetJobJobSettingsSettingsJobClusterNewClusterWorkloadType { - if v == nil { - return nil - } - return v.WorkloadType - }).(GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypePtrOutput) -} - type GetJobJobSettingsSettingsJobClusterNewClusterAutoscale struct { MaxWorkers *int `pulumi:"maxWorkers"` MinWorkers *int `pulumi:"minWorkers"` @@ -83958,7 +83187,6 @@ func init() { pulumi.RegisterInputType(reflect.TypeOf((*JobJobClusterInput)(nil)).Elem(), JobJobClusterArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*JobJobClusterArrayInput)(nil)).Elem(), JobJobClusterArray{}) pulumi.RegisterInputType(reflect.TypeOf((*JobJobClusterNewClusterInput)(nil)).Elem(), JobJobClusterNewClusterArgs{}) - pulumi.RegisterInputType(reflect.TypeOf((*JobJobClusterNewClusterPtrInput)(nil)).Elem(), JobJobClusterNewClusterArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*JobJobClusterNewClusterAutoscaleInput)(nil)).Elem(), JobJobClusterNewClusterAutoscaleArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*JobJobClusterNewClusterAutoscalePtrInput)(nil)).Elem(), JobJobClusterNewClusterAutoscaleArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*JobJobClusterNewClusterAwsAttributesInput)(nil)).Elem(), JobJobClusterNewClusterAwsAttributesArgs{}) @@ -84782,7 +84010,6 @@ func init() { pulumi.RegisterInputType(reflect.TypeOf((*GetJobJobSettingsSettingsJobClusterInput)(nil)).Elem(), GetJobJobSettingsSettingsJobClusterArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetJobJobSettingsSettingsJobClusterArrayInput)(nil)).Elem(), GetJobJobSettingsSettingsJobClusterArray{}) pulumi.RegisterInputType(reflect.TypeOf((*GetJobJobSettingsSettingsJobClusterNewClusterInput)(nil)).Elem(), GetJobJobSettingsSettingsJobClusterNewClusterArgs{}) - pulumi.RegisterInputType(reflect.TypeOf((*GetJobJobSettingsSettingsJobClusterNewClusterPtrInput)(nil)).Elem(), GetJobJobSettingsSettingsJobClusterNewClusterArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleInput)(nil)).Elem(), GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetJobJobSettingsSettingsJobClusterNewClusterAutoscalePtrInput)(nil)).Elem(), GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesInput)(nil)).Elem(), GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesArgs{}) @@ -84952,7 +84179,6 @@ func init() { pulumi.RegisterOutputType(JobJobClusterOutput{}) pulumi.RegisterOutputType(JobJobClusterArrayOutput{}) pulumi.RegisterOutputType(JobJobClusterNewClusterOutput{}) - pulumi.RegisterOutputType(JobJobClusterNewClusterPtrOutput{}) pulumi.RegisterOutputType(JobJobClusterNewClusterAutoscaleOutput{}) pulumi.RegisterOutputType(JobJobClusterNewClusterAutoscalePtrOutput{}) pulumi.RegisterOutputType(JobJobClusterNewClusterAwsAttributesOutput{}) @@ -85776,7 +85002,6 @@ func init() { pulumi.RegisterOutputType(GetJobJobSettingsSettingsJobClusterOutput{}) pulumi.RegisterOutputType(GetJobJobSettingsSettingsJobClusterArrayOutput{}) pulumi.RegisterOutputType(GetJobJobSettingsSettingsJobClusterNewClusterOutput{}) - pulumi.RegisterOutputType(GetJobJobSettingsSettingsJobClusterNewClusterPtrOutput{}) pulumi.RegisterOutputType(GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleOutput{}) pulumi.RegisterOutputType(GetJobJobSettingsSettingsJobClusterNewClusterAutoscalePtrOutput{}) pulumi.RegisterOutputType(GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesOutput{}) diff --git a/sdk/go/databricks/pulumiTypes1.go b/sdk/go/databricks/pulumiTypes1.go index 51640e81..6e03947b 100644 --- a/sdk/go/databricks/pulumiTypes1.go +++ b/sdk/go/databricks/pulumiTypes1.go @@ -6481,7 +6481,7 @@ type GetJobJobSettingsSettingsTask struct { SparkPythonTask *GetJobJobSettingsSettingsTaskSparkPythonTask `pulumi:"sparkPythonTask"` SparkSubmitTask *GetJobJobSettingsSettingsTaskSparkSubmitTask `pulumi:"sparkSubmitTask"` SqlTask *GetJobJobSettingsSettingsTaskSqlTask `pulumi:"sqlTask"` - TaskKey *string `pulumi:"taskKey"` + TaskKey string `pulumi:"taskKey"` TimeoutSeconds *int `pulumi:"timeoutSeconds"` WebhookNotifications *GetJobJobSettingsSettingsTaskWebhookNotifications `pulumi:"webhookNotifications"` } @@ -6523,7 +6523,7 @@ type GetJobJobSettingsSettingsTaskArgs struct { SparkPythonTask GetJobJobSettingsSettingsTaskSparkPythonTaskPtrInput `pulumi:"sparkPythonTask"` SparkSubmitTask GetJobJobSettingsSettingsTaskSparkSubmitTaskPtrInput `pulumi:"sparkSubmitTask"` SqlTask GetJobJobSettingsSettingsTaskSqlTaskPtrInput `pulumi:"sqlTask"` - TaskKey pulumi.StringPtrInput `pulumi:"taskKey"` + TaskKey pulumi.StringInput `pulumi:"taskKey"` TimeoutSeconds pulumi.IntPtrInput `pulumi:"timeoutSeconds"` WebhookNotifications GetJobJobSettingsSettingsTaskWebhookNotificationsPtrInput `pulumi:"webhookNotifications"` } @@ -6697,8 +6697,8 @@ func (o GetJobJobSettingsSettingsTaskOutput) SqlTask() GetJobJobSettingsSettings return o.ApplyT(func(v GetJobJobSettingsSettingsTask) *GetJobJobSettingsSettingsTaskSqlTask { return v.SqlTask }).(GetJobJobSettingsSettingsTaskSqlTaskPtrOutput) } -func (o GetJobJobSettingsSettingsTaskOutput) TaskKey() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTask) *string { return v.TaskKey }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskOutput) TaskKey() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTask) string { return v.TaskKey }).(pulumi.StringOutput) } func (o GetJobJobSettingsSettingsTaskOutput) TimeoutSeconds() pulumi.IntPtrOutput { @@ -7602,7 +7602,7 @@ type GetJobJobSettingsSettingsTaskForEachTaskTask struct { SparkPythonTask *GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTask `pulumi:"sparkPythonTask"` SparkSubmitTask *GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTask `pulumi:"sparkSubmitTask"` SqlTask *GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask `pulumi:"sqlTask"` - TaskKey *string `pulumi:"taskKey"` + TaskKey string `pulumi:"taskKey"` TimeoutSeconds *int `pulumi:"timeoutSeconds"` WebhookNotifications *GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotifications `pulumi:"webhookNotifications"` } @@ -7643,7 +7643,7 @@ type GetJobJobSettingsSettingsTaskForEachTaskTaskArgs struct { SparkPythonTask GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTaskPtrInput `pulumi:"sparkPythonTask"` SparkSubmitTask GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTaskPtrInput `pulumi:"sparkSubmitTask"` SqlTask GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskPtrInput `pulumi:"sqlTask"` - TaskKey pulumi.StringPtrInput `pulumi:"taskKey"` + TaskKey pulumi.StringInput `pulumi:"taskKey"` TimeoutSeconds pulumi.IntPtrInput `pulumi:"timeoutSeconds"` WebhookNotifications GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotificationsPtrInput `pulumi:"webhookNotifications"` } @@ -7853,8 +7853,8 @@ func (o GetJobJobSettingsSettingsTaskForEachTaskTaskOutput) SqlTask() GetJobJobS }).(GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskPtrOutput) } -func (o GetJobJobSettingsSettingsTaskForEachTaskTaskOutput) TaskKey() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTask) *string { return v.TaskKey }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskForEachTaskTaskOutput) TaskKey() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTask) string { return v.TaskKey }).(pulumi.StringOutput) } func (o GetJobJobSettingsSettingsTaskForEachTaskTaskOutput) TimeoutSeconds() pulumi.IntPtrOutput { @@ -8112,7 +8112,7 @@ func (o GetJobJobSettingsSettingsTaskForEachTaskTaskPtrOutput) TaskKey() pulumi. if v == nil { return nil } - return v.TaskKey + return &v.TaskKey }).(pulumi.StringPtrOutput) } @@ -8953,9 +8953,9 @@ func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthPtrOutput) Rules() Get } type GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule struct { - Metric *string `pulumi:"metric"` - Op *string `pulumi:"op"` - Value *int `pulumi:"value"` + Metric string `pulumi:"metric"` + Op string `pulumi:"op"` + Value int `pulumi:"value"` } // GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleInput is an input type that accepts GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs and GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput values. @@ -8970,9 +8970,9 @@ type GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleInput interface { } type GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs struct { - Metric pulumi.StringPtrInput `pulumi:"metric"` - Op pulumi.StringPtrInput `pulumi:"op"` - Value pulumi.IntPtrInput `pulumi:"value"` + Metric pulumi.StringInput `pulumi:"metric"` + Op pulumi.StringInput `pulumi:"op"` + Value pulumi.IntInput `pulumi:"value"` } func (GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs) ElementType() reflect.Type { @@ -9026,16 +9026,16 @@ func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput) ToGetJobJo return o } -func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput) Metric() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule) *string { return v.Metric }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput) Metric() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule) string { return v.Metric }).(pulumi.StringOutput) } -func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput) Op() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule) *string { return v.Op }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput) Op() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule) string { return v.Op }).(pulumi.StringOutput) } -func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput) Value() pulumi.IntPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule) *int { return v.Value }).(pulumi.IntPtrOutput) +func (o GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleOutput) Value() pulumi.IntOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule) int { return v.Value }).(pulumi.IntOutput) } type GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArrayOutput struct{ *pulumi.OutputState } @@ -14793,7 +14793,7 @@ type GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask struct { File *GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFile `pulumi:"file"` Parameters map[string]interface{} `pulumi:"parameters"` Query *GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQuery `pulumi:"query"` - WarehouseId *string `pulumi:"warehouseId"` + WarehouseId string `pulumi:"warehouseId"` } // GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskInput is an input type that accepts GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs and GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskOutput values. @@ -14813,7 +14813,7 @@ type GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs struct { File GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFilePtrInput `pulumi:"file"` Parameters pulumi.MapInput `pulumi:"parameters"` Query GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryPtrInput `pulumi:"query"` - WarehouseId pulumi.StringPtrInput `pulumi:"warehouseId"` + WarehouseId pulumi.StringInput `pulumi:"warehouseId"` } func (GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs) ElementType() reflect.Type { @@ -14923,8 +14923,8 @@ func (o GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskOutput) Query() GetJo }).(GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryPtrOutput) } -func (o GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskOutput) WarehouseId() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask) *string { return v.WarehouseId }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskOutput) WarehouseId() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask) string { return v.WarehouseId }).(pulumi.StringOutput) } type GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskPtrOutput struct{ *pulumi.OutputState } @@ -15001,7 +15001,7 @@ func (o GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskPtrOutput) WarehouseI if v == nil { return nil } - return v.WarehouseId + return &v.WarehouseId }).(pulumi.StringPtrOutput) } @@ -16551,9 +16551,9 @@ func (o GetJobJobSettingsSettingsTaskHealthPtrOutput) Rules() GetJobJobSettingsS } type GetJobJobSettingsSettingsTaskHealthRule struct { - Metric *string `pulumi:"metric"` - Op *string `pulumi:"op"` - Value *int `pulumi:"value"` + Metric string `pulumi:"metric"` + Op string `pulumi:"op"` + Value int `pulumi:"value"` } // GetJobJobSettingsSettingsTaskHealthRuleInput is an input type that accepts GetJobJobSettingsSettingsTaskHealthRuleArgs and GetJobJobSettingsSettingsTaskHealthRuleOutput values. @@ -16568,9 +16568,9 @@ type GetJobJobSettingsSettingsTaskHealthRuleInput interface { } type GetJobJobSettingsSettingsTaskHealthRuleArgs struct { - Metric pulumi.StringPtrInput `pulumi:"metric"` - Op pulumi.StringPtrInput `pulumi:"op"` - Value pulumi.IntPtrInput `pulumi:"value"` + Metric pulumi.StringInput `pulumi:"metric"` + Op pulumi.StringInput `pulumi:"op"` + Value pulumi.IntInput `pulumi:"value"` } func (GetJobJobSettingsSettingsTaskHealthRuleArgs) ElementType() reflect.Type { @@ -16624,16 +16624,16 @@ func (o GetJobJobSettingsSettingsTaskHealthRuleOutput) ToGetJobJobSettingsSettin return o } -func (o GetJobJobSettingsSettingsTaskHealthRuleOutput) Metric() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskHealthRule) *string { return v.Metric }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskHealthRuleOutput) Metric() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskHealthRule) string { return v.Metric }).(pulumi.StringOutput) } -func (o GetJobJobSettingsSettingsTaskHealthRuleOutput) Op() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskHealthRule) *string { return v.Op }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskHealthRuleOutput) Op() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskHealthRule) string { return v.Op }).(pulumi.StringOutput) } -func (o GetJobJobSettingsSettingsTaskHealthRuleOutput) Value() pulumi.IntPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskHealthRule) *int { return v.Value }).(pulumi.IntPtrOutput) +func (o GetJobJobSettingsSettingsTaskHealthRuleOutput) Value() pulumi.IntOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskHealthRule) int { return v.Value }).(pulumi.IntOutput) } type GetJobJobSettingsSettingsTaskHealthRuleArrayOutput struct{ *pulumi.OutputState } @@ -22291,7 +22291,7 @@ type GetJobJobSettingsSettingsTaskSqlTask struct { File *GetJobJobSettingsSettingsTaskSqlTaskFile `pulumi:"file"` Parameters map[string]interface{} `pulumi:"parameters"` Query *GetJobJobSettingsSettingsTaskSqlTaskQuery `pulumi:"query"` - WarehouseId *string `pulumi:"warehouseId"` + WarehouseId string `pulumi:"warehouseId"` } // GetJobJobSettingsSettingsTaskSqlTaskInput is an input type that accepts GetJobJobSettingsSettingsTaskSqlTaskArgs and GetJobJobSettingsSettingsTaskSqlTaskOutput values. @@ -22311,7 +22311,7 @@ type GetJobJobSettingsSettingsTaskSqlTaskArgs struct { File GetJobJobSettingsSettingsTaskSqlTaskFilePtrInput `pulumi:"file"` Parameters pulumi.MapInput `pulumi:"parameters"` Query GetJobJobSettingsSettingsTaskSqlTaskQueryPtrInput `pulumi:"query"` - WarehouseId pulumi.StringPtrInput `pulumi:"warehouseId"` + WarehouseId pulumi.StringInput `pulumi:"warehouseId"` } func (GetJobJobSettingsSettingsTaskSqlTaskArgs) ElementType() reflect.Type { @@ -22417,8 +22417,8 @@ func (o GetJobJobSettingsSettingsTaskSqlTaskOutput) Query() GetJobJobSettingsSet }).(GetJobJobSettingsSettingsTaskSqlTaskQueryPtrOutput) } -func (o GetJobJobSettingsSettingsTaskSqlTaskOutput) WarehouseId() pulumi.StringPtrOutput { - return o.ApplyT(func(v GetJobJobSettingsSettingsTaskSqlTask) *string { return v.WarehouseId }).(pulumi.StringPtrOutput) +func (o GetJobJobSettingsSettingsTaskSqlTaskOutput) WarehouseId() pulumi.StringOutput { + return o.ApplyT(func(v GetJobJobSettingsSettingsTaskSqlTask) string { return v.WarehouseId }).(pulumi.StringOutput) } type GetJobJobSettingsSettingsTaskSqlTaskPtrOutput struct{ *pulumi.OutputState } @@ -22495,7 +22495,7 @@ func (o GetJobJobSettingsSettingsTaskSqlTaskPtrOutput) WarehouseId() pulumi.Stri if v == nil { return nil } - return v.WarehouseId + return &v.WarehouseId }).(pulumi.StringPtrOutput) } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRule.java index 74ca3e96..909c92a3 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRule.java @@ -4,36 +4,35 @@ package com.pulumi.databricks.inputs; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsHealthRule extends com.pulumi.resources.InvokeArgs { public static final GetJobJobSettingsSettingsHealthRule Empty = new GetJobJobSettingsSettingsHealthRule(); - @Import(name="metric") - private @Nullable String metric; + @Import(name="metric", required=true) + private String metric; - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } - @Import(name="op") - private @Nullable String op; + @Import(name="op", required=true) + private String op; - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } - @Import(name="value") - private @Nullable Integer value; + @Import(name="value", required=true) + private Integer value; - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } private GetJobJobSettingsSettingsHealthRule() {} @@ -62,22 +61,31 @@ public Builder(GetJobJobSettingsSettingsHealthRule defaults) { $ = new GetJobJobSettingsSettingsHealthRule(Objects.requireNonNull(defaults)); } - public Builder metric(@Nullable String metric) { + public Builder metric(String metric) { $.metric = metric; return this; } - public Builder op(@Nullable String op) { + public Builder op(String op) { $.op = op; return this; } - public Builder value(@Nullable Integer value) { + public Builder value(Integer value) { $.value = value; return this; } public GetJobJobSettingsSettingsHealthRule build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRule", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRule", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRule", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRuleArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRuleArgs.java index 13aff4eb..2fcd63d1 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRuleArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsHealthRuleArgs.java @@ -5,36 +5,35 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsHealthRuleArgs extends com.pulumi.resources.ResourceArgs { public static final GetJobJobSettingsSettingsHealthRuleArgs Empty = new GetJobJobSettingsSettingsHealthRuleArgs(); - @Import(name="metric") - private @Nullable Output metric; + @Import(name="metric", required=true) + private Output metric; - public Optional> metric() { - return Optional.ofNullable(this.metric); + public Output metric() { + return this.metric; } - @Import(name="op") - private @Nullable Output op; + @Import(name="op", required=true) + private Output op; - public Optional> op() { - return Optional.ofNullable(this.op); + public Output op() { + return this.op; } - @Import(name="value") - private @Nullable Output value; + @Import(name="value", required=true) + private Output value; - public Optional> value() { - return Optional.ofNullable(this.value); + public Output value() { + return this.value; } private GetJobJobSettingsSettingsHealthRuleArgs() {} @@ -63,7 +62,7 @@ public Builder(GetJobJobSettingsSettingsHealthRuleArgs defaults) { $ = new GetJobJobSettingsSettingsHealthRuleArgs(Objects.requireNonNull(defaults)); } - public Builder metric(@Nullable Output metric) { + public Builder metric(Output metric) { $.metric = metric; return this; } @@ -72,7 +71,7 @@ public Builder metric(String metric) { return metric(Output.of(metric)); } - public Builder op(@Nullable Output op) { + public Builder op(Output op) { $.op = op; return this; } @@ -81,7 +80,7 @@ public Builder op(String op) { return op(Output.of(op)); } - public Builder value(@Nullable Output value) { + public Builder value(Output value) { $.value = value; return this; } @@ -91,6 +90,15 @@ public Builder value(Integer value) { } public GetJobJobSettingsSettingsHealthRuleArgs build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRuleArgs", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRuleArgs", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRuleArgs", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobCluster.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobCluster.java index c219e87b..0bd0e15f 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobCluster.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobCluster.java @@ -5,28 +5,27 @@ import com.pulumi.core.annotations.Import; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsJobClusterNewCluster; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsJobCluster extends com.pulumi.resources.InvokeArgs { public static final GetJobJobSettingsSettingsJobCluster Empty = new GetJobJobSettingsSettingsJobCluster(); - @Import(name="jobClusterKey") - private @Nullable String jobClusterKey; + @Import(name="jobClusterKey", required=true) + private String jobClusterKey; - public Optional jobClusterKey() { - return Optional.ofNullable(this.jobClusterKey); + public String jobClusterKey() { + return this.jobClusterKey; } - @Import(name="newCluster") - private @Nullable GetJobJobSettingsSettingsJobClusterNewCluster newCluster; + @Import(name="newCluster", required=true) + private GetJobJobSettingsSettingsJobClusterNewCluster newCluster; - public Optional newCluster() { - return Optional.ofNullable(this.newCluster); + public GetJobJobSettingsSettingsJobClusterNewCluster newCluster() { + return this.newCluster; } private GetJobJobSettingsSettingsJobCluster() {} @@ -54,17 +53,23 @@ public Builder(GetJobJobSettingsSettingsJobCluster defaults) { $ = new GetJobJobSettingsSettingsJobCluster(Objects.requireNonNull(defaults)); } - public Builder jobClusterKey(@Nullable String jobClusterKey) { + public Builder jobClusterKey(String jobClusterKey) { $.jobClusterKey = jobClusterKey; return this; } - public Builder newCluster(@Nullable GetJobJobSettingsSettingsJobClusterNewCluster newCluster) { + public Builder newCluster(GetJobJobSettingsSettingsJobClusterNewCluster newCluster) { $.newCluster = newCluster; return this; } public GetJobJobSettingsSettingsJobCluster build() { + if ($.jobClusterKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsJobCluster", "jobClusterKey"); + } + if ($.newCluster == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsJobCluster", "newCluster"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobClusterArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobClusterArgs.java index 3574139b..929aee65 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobClusterArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsJobClusterArgs.java @@ -6,28 +6,27 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsJobClusterNewClusterArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsJobClusterArgs extends com.pulumi.resources.ResourceArgs { public static final GetJobJobSettingsSettingsJobClusterArgs Empty = new GetJobJobSettingsSettingsJobClusterArgs(); - @Import(name="jobClusterKey") - private @Nullable Output jobClusterKey; + @Import(name="jobClusterKey", required=true) + private Output jobClusterKey; - public Optional> jobClusterKey() { - return Optional.ofNullable(this.jobClusterKey); + public Output jobClusterKey() { + return this.jobClusterKey; } - @Import(name="newCluster") - private @Nullable Output newCluster; + @Import(name="newCluster", required=true) + private Output newCluster; - public Optional> newCluster() { - return Optional.ofNullable(this.newCluster); + public Output newCluster() { + return this.newCluster; } private GetJobJobSettingsSettingsJobClusterArgs() {} @@ -55,7 +54,7 @@ public Builder(GetJobJobSettingsSettingsJobClusterArgs defaults) { $ = new GetJobJobSettingsSettingsJobClusterArgs(Objects.requireNonNull(defaults)); } - public Builder jobClusterKey(@Nullable Output jobClusterKey) { + public Builder jobClusterKey(Output jobClusterKey) { $.jobClusterKey = jobClusterKey; return this; } @@ -64,7 +63,7 @@ public Builder jobClusterKey(String jobClusterKey) { return jobClusterKey(Output.of(jobClusterKey)); } - public Builder newCluster(@Nullable Output newCluster) { + public Builder newCluster(Output newCluster) { $.newCluster = newCluster; return this; } @@ -74,6 +73,12 @@ public Builder newCluster(GetJobJobSettingsSettingsJobClusterNewClusterArgs newC } public GetJobJobSettingsSettingsJobClusterArgs build() { + if ($.jobClusterKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsJobClusterArgs", "jobClusterKey"); + } + if ($.newCluster == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsJobClusterArgs", "newCluster"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTask.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTask.java index 1f40e361..909b8e6b 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTask.java @@ -211,11 +211,11 @@ public Optional sqlTask() { return Optional.ofNullable(this.sqlTask); } - @Import(name="taskKey") - private @Nullable String taskKey; + @Import(name="taskKey", required=true) + private String taskKey; - public Optional taskKey() { - return Optional.ofNullable(this.taskKey); + public String taskKey() { + return this.taskKey; } @Import(name="timeoutSeconds") @@ -416,7 +416,7 @@ public Builder sqlTask(@Nullable GetJobJobSettingsSettingsTaskSqlTask sqlTask) { return this; } - public Builder taskKey(@Nullable String taskKey) { + public Builder taskKey(String taskKey) { $.taskKey = taskKey; return this; } @@ -435,6 +435,9 @@ public GetJobJobSettingsSettingsTask build() { if ($.retryOnTimeout == null) { throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTask", "retryOnTimeout"); } + if ($.taskKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTask", "taskKey"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskArgs.java index 2c972d6f..05316736 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskArgs.java @@ -212,11 +212,11 @@ public Optional> sqlTask() { return Optional.ofNullable(this.sqlTask); } - @Import(name="taskKey") - private @Nullable Output taskKey; + @Import(name="taskKey", required=true) + private Output taskKey; - public Optional> taskKey() { - return Optional.ofNullable(this.taskKey); + public Output taskKey() { + return this.taskKey; } @Import(name="timeoutSeconds") @@ -517,7 +517,7 @@ public Builder sqlTask(GetJobJobSettingsSettingsTaskSqlTaskArgs sqlTask) { return sqlTask(Output.of(sqlTask)); } - public Builder taskKey(@Nullable Output taskKey) { + public Builder taskKey(Output taskKey) { $.taskKey = taskKey; return this; } @@ -548,6 +548,9 @@ public GetJobJobSettingsSettingsTaskArgs build() { if ($.retryOnTimeout == null) { throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskArgs", "retryOnTimeout"); } + if ($.taskKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskArgs", "taskKey"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java index 36e2f551..0e555c42 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java @@ -203,11 +203,11 @@ public Optional sqlTask() { return Optional.ofNullable(this.sqlTask); } - @Import(name="taskKey") - private @Nullable String taskKey; + @Import(name="taskKey", required=true) + private String taskKey; - public Optional taskKey() { - return Optional.ofNullable(this.taskKey); + public String taskKey() { + return this.taskKey; } @Import(name="timeoutSeconds") @@ -402,7 +402,7 @@ public Builder sqlTask(@Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSql return this; } - public Builder taskKey(@Nullable String taskKey) { + public Builder taskKey(String taskKey) { $.taskKey = taskKey; return this; } @@ -421,6 +421,9 @@ public GetJobJobSettingsSettingsTaskForEachTaskTask build() { if ($.retryOnTimeout == null) { throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTask", "retryOnTimeout"); } + if ($.taskKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTask", "taskKey"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.java index 5e21747b..5505c75c 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskArgs.java @@ -204,11 +204,11 @@ public Optional> return Optional.ofNullable(this.sqlTask); } - @Import(name="taskKey") - private @Nullable Output taskKey; + @Import(name="taskKey", required=true) + private Output taskKey; - public Optional> taskKey() { - return Optional.ofNullable(this.taskKey); + public Output taskKey() { + return this.taskKey; } @Import(name="timeoutSeconds") @@ -499,7 +499,7 @@ public Builder sqlTask(GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs s return sqlTask(Output.of(sqlTask)); } - public Builder taskKey(@Nullable Output taskKey) { + public Builder taskKey(Output taskKey) { $.taskKey = taskKey; return this; } @@ -530,6 +530,9 @@ public GetJobJobSettingsSettingsTaskForEachTaskTaskArgs build() { if ($.retryOnTimeout == null) { throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskArgs", "retryOnTimeout"); } + if ($.taskKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskArgs", "taskKey"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java index 2ecda10f..02412a93 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java @@ -4,36 +4,35 @@ package com.pulumi.databricks.inputs; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule extends com.pulumi.resources.InvokeArgs { public static final GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule Empty = new GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule(); - @Import(name="metric") - private @Nullable String metric; + @Import(name="metric", required=true) + private String metric; - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } - @Import(name="op") - private @Nullable String op; + @Import(name="op", required=true) + private String op; - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } - @Import(name="value") - private @Nullable Integer value; + @Import(name="value", required=true) + private Integer value; - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } private GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule() {} @@ -62,22 +61,31 @@ public Builder(GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule defaults) $ = new GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule(Objects.requireNonNull(defaults)); } - public Builder metric(@Nullable String metric) { + public Builder metric(String metric) { $.metric = metric; return this; } - public Builder op(@Nullable String op) { + public Builder op(String op) { $.op = op; return this; } - public Builder value(@Nullable Integer value) { + public Builder value(Integer value) { $.value = value; return this; } public GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.java index ba263686..42f92bcc 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs.java @@ -5,36 +5,35 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs extends com.pulumi.resources.ResourceArgs { public static final GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs Empty = new GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs(); - @Import(name="metric") - private @Nullable Output metric; + @Import(name="metric", required=true) + private Output metric; - public Optional> metric() { - return Optional.ofNullable(this.metric); + public Output metric() { + return this.metric; } - @Import(name="op") - private @Nullable Output op; + @Import(name="op", required=true) + private Output op; - public Optional> op() { - return Optional.ofNullable(this.op); + public Output op() { + return this.op; } - @Import(name="value") - private @Nullable Output value; + @Import(name="value", required=true) + private Output value; - public Optional> value() { - return Optional.ofNullable(this.value); + public Output value() { + return this.value; } private GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs() {} @@ -63,7 +62,7 @@ public Builder(GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs defaul $ = new GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs(Objects.requireNonNull(defaults)); } - public Builder metric(@Nullable Output metric) { + public Builder metric(Output metric) { $.metric = metric; return this; } @@ -72,7 +71,7 @@ public Builder metric(String metric) { return metric(Output.of(metric)); } - public Builder op(@Nullable Output op) { + public Builder op(Output op) { $.op = op; return this; } @@ -81,7 +80,7 @@ public Builder op(String op) { return op(Output.of(op)); } - public Builder value(@Nullable Output value) { + public Builder value(Output value) { $.value = value; return this; } @@ -91,6 +90,15 @@ public Builder value(Integer value) { } public GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java index 0435b360..00a1cb27 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java @@ -8,6 +8,7 @@ import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskDashboard; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFile; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQuery; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -55,11 +56,11 @@ public Optional query( return Optional.ofNullable(this.query); } - @Import(name="warehouseId") - private @Nullable String warehouseId; + @Import(name="warehouseId", required=true) + private String warehouseId; - public Optional warehouseId() { - return Optional.ofNullable(this.warehouseId); + public String warehouseId() { + return this.warehouseId; } private GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask() {} @@ -116,12 +117,15 @@ public Builder query(@Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTa return this; } - public Builder warehouseId(@Nullable String warehouseId) { + public Builder warehouseId(String warehouseId) { $.warehouseId = warehouseId; return this; } public GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask build() { + if ($.warehouseId == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask", "warehouseId"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.java index a7721de4..a18e5ec7 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs.java @@ -9,6 +9,7 @@ import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskDashboardArgs; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFileArgs; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -56,11 +57,11 @@ public Optional warehouseId; + @Import(name="warehouseId", required=true) + private Output warehouseId; - public Optional> warehouseId() { - return Optional.ofNullable(this.warehouseId); + public Output warehouseId() { + return this.warehouseId; } private GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs() {} @@ -137,7 +138,7 @@ public Builder query(GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryArg return query(Output.of(query)); } - public Builder warehouseId(@Nullable Output warehouseId) { + public Builder warehouseId(Output warehouseId) { $.warehouseId = warehouseId; return this; } @@ -147,6 +148,9 @@ public Builder warehouseId(String warehouseId) { } public GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs build() { + if ($.warehouseId == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs", "warehouseId"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRule.java index 703b9907..9c2bce8e 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRule.java @@ -4,36 +4,35 @@ package com.pulumi.databricks.inputs; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsTaskHealthRule extends com.pulumi.resources.InvokeArgs { public static final GetJobJobSettingsSettingsTaskHealthRule Empty = new GetJobJobSettingsSettingsTaskHealthRule(); - @Import(name="metric") - private @Nullable String metric; + @Import(name="metric", required=true) + private String metric; - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } - @Import(name="op") - private @Nullable String op; + @Import(name="op", required=true) + private String op; - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } - @Import(name="value") - private @Nullable Integer value; + @Import(name="value", required=true) + private Integer value; - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } private GetJobJobSettingsSettingsTaskHealthRule() {} @@ -62,22 +61,31 @@ public Builder(GetJobJobSettingsSettingsTaskHealthRule defaults) { $ = new GetJobJobSettingsSettingsTaskHealthRule(Objects.requireNonNull(defaults)); } - public Builder metric(@Nullable String metric) { + public Builder metric(String metric) { $.metric = metric; return this; } - public Builder op(@Nullable String op) { + public Builder op(String op) { $.op = op; return this; } - public Builder value(@Nullable Integer value) { + public Builder value(Integer value) { $.value = value; return this; } public GetJobJobSettingsSettingsTaskHealthRule build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRule", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRule", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRule", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.java index 25dbe43d..067f5f28 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskHealthRuleArgs.java @@ -5,36 +5,35 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class GetJobJobSettingsSettingsTaskHealthRuleArgs extends com.pulumi.resources.ResourceArgs { public static final GetJobJobSettingsSettingsTaskHealthRuleArgs Empty = new GetJobJobSettingsSettingsTaskHealthRuleArgs(); - @Import(name="metric") - private @Nullable Output metric; + @Import(name="metric", required=true) + private Output metric; - public Optional> metric() { - return Optional.ofNullable(this.metric); + public Output metric() { + return this.metric; } - @Import(name="op") - private @Nullable Output op; + @Import(name="op", required=true) + private Output op; - public Optional> op() { - return Optional.ofNullable(this.op); + public Output op() { + return this.op; } - @Import(name="value") - private @Nullable Output value; + @Import(name="value", required=true) + private Output value; - public Optional> value() { - return Optional.ofNullable(this.value); + public Output value() { + return this.value; } private GetJobJobSettingsSettingsTaskHealthRuleArgs() {} @@ -63,7 +62,7 @@ public Builder(GetJobJobSettingsSettingsTaskHealthRuleArgs defaults) { $ = new GetJobJobSettingsSettingsTaskHealthRuleArgs(Objects.requireNonNull(defaults)); } - public Builder metric(@Nullable Output metric) { + public Builder metric(Output metric) { $.metric = metric; return this; } @@ -72,7 +71,7 @@ public Builder metric(String metric) { return metric(Output.of(metric)); } - public Builder op(@Nullable Output op) { + public Builder op(Output op) { $.op = op; return this; } @@ -81,7 +80,7 @@ public Builder op(String op) { return op(Output.of(op)); } - public Builder value(@Nullable Output value) { + public Builder value(Output value) { $.value = value; return this; } @@ -91,6 +90,15 @@ public Builder value(Integer value) { } public GetJobJobSettingsSettingsTaskHealthRuleArgs build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRuleArgs", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRuleArgs", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRuleArgs", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTask.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTask.java index 51123d87..36a7f3bf 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTask.java @@ -8,6 +8,7 @@ import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskSqlTaskDashboard; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskSqlTaskFile; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskSqlTaskQuery; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -55,11 +56,11 @@ public Optional query() { return Optional.ofNullable(this.query); } - @Import(name="warehouseId") - private @Nullable String warehouseId; + @Import(name="warehouseId", required=true) + private String warehouseId; - public Optional warehouseId() { - return Optional.ofNullable(this.warehouseId); + public String warehouseId() { + return this.warehouseId; } private GetJobJobSettingsSettingsTaskSqlTask() {} @@ -116,12 +117,15 @@ public Builder query(@Nullable GetJobJobSettingsSettingsTaskSqlTaskQuery query) return this; } - public Builder warehouseId(@Nullable String warehouseId) { + public Builder warehouseId(String warehouseId) { $.warehouseId = warehouseId; return this; } public GetJobJobSettingsSettingsTaskSqlTask build() { + if ($.warehouseId == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskSqlTask", "warehouseId"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.java index 766a5e5d..52e3036e 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/GetJobJobSettingsSettingsTaskSqlTaskArgs.java @@ -9,6 +9,7 @@ import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskSqlTaskDashboardArgs; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskSqlTaskFileArgs; import com.pulumi.databricks.inputs.GetJobJobSettingsSettingsTaskSqlTaskQueryArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -56,11 +57,11 @@ public Optional> query() { return Optional.ofNullable(this.query); } - @Import(name="warehouseId") - private @Nullable Output warehouseId; + @Import(name="warehouseId", required=true) + private Output warehouseId; - public Optional> warehouseId() { - return Optional.ofNullable(this.warehouseId); + public Output warehouseId() { + return this.warehouseId; } private GetJobJobSettingsSettingsTaskSqlTaskArgs() {} @@ -137,7 +138,7 @@ public Builder query(GetJobJobSettingsSettingsTaskSqlTaskQueryArgs query) { return query(Output.of(query)); } - public Builder warehouseId(@Nullable Output warehouseId) { + public Builder warehouseId(Output warehouseId) { $.warehouseId = warehouseId; return this; } @@ -147,6 +148,9 @@ public Builder warehouseId(String warehouseId) { } public GetJobJobSettingsSettingsTaskSqlTaskArgs build() { + if ($.warehouseId == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskSqlTaskArgs", "warehouseId"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobHealthRuleArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobHealthRuleArgs.java index c27749c2..dc6241df 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobHealthRuleArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobHealthRuleArgs.java @@ -5,11 +5,10 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class JobHealthRuleArgs extends com.pulumi.resources.ResourceArgs { @@ -20,45 +19,45 @@ public final class JobHealthRuleArgs extends com.pulumi.resources.ResourceArgs { * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - @Import(name="metric") - private @Nullable Output metric; + @Import(name="metric", required=true) + private Output metric; /** * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - public Optional> metric() { - return Optional.ofNullable(this.metric); + public Output metric() { + return this.metric; } /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - @Import(name="op") - private @Nullable Output op; + @Import(name="op", required=true) + private Output op; /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - public Optional> op() { - return Optional.ofNullable(this.op); + public Output op() { + return this.op; } /** * integer value used to compare to the given metric. * */ - @Import(name="value") - private @Nullable Output value; + @Import(name="value", required=true) + private Output value; /** * @return integer value used to compare to the given metric. * */ - public Optional> value() { - return Optional.ofNullable(this.value); + public Output value() { + return this.value; } private JobHealthRuleArgs() {} @@ -93,7 +92,7 @@ public Builder(JobHealthRuleArgs defaults) { * @return builder * */ - public Builder metric(@Nullable Output metric) { + public Builder metric(Output metric) { $.metric = metric; return this; } @@ -114,7 +113,7 @@ public Builder metric(String metric) { * @return builder * */ - public Builder op(@Nullable Output op) { + public Builder op(Output op) { $.op = op; return this; } @@ -135,7 +134,7 @@ public Builder op(String op) { * @return builder * */ - public Builder value(@Nullable Output value) { + public Builder value(Output value) { $.value = value; return this; } @@ -151,6 +150,15 @@ public Builder value(Integer value) { } public JobHealthRuleArgs build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("JobHealthRuleArgs", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("JobHealthRuleArgs", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("JobHealthRuleArgs", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterArgs.java index 23d1b87a..56c9434b 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterArgs.java @@ -6,10 +6,9 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; import com.pulumi.databricks.inputs.JobJobClusterNewClusterArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class JobJobClusterArgs extends com.pulumi.resources.ResourceArgs { @@ -20,30 +19,30 @@ public final class JobJobClusterArgs extends com.pulumi.resources.ResourceArgs { * Identifier that can be referenced in `task` block, so that cluster is shared between tasks * */ - @Import(name="jobClusterKey") - private @Nullable Output jobClusterKey; + @Import(name="jobClusterKey", required=true) + private Output jobClusterKey; /** * @return Identifier that can be referenced in `task` block, so that cluster is shared between tasks * */ - public Optional> jobClusterKey() { - return Optional.ofNullable(this.jobClusterKey); + public Output jobClusterKey() { + return this.jobClusterKey; } /** * Same set of parameters as for databricks.Cluster resource. * */ - @Import(name="newCluster") - private @Nullable Output newCluster; + @Import(name="newCluster", required=true) + private Output newCluster; /** * @return Same set of parameters as for databricks.Cluster resource. * */ - public Optional> newCluster() { - return Optional.ofNullable(this.newCluster); + public Output newCluster() { + return this.newCluster; } private JobJobClusterArgs() {} @@ -77,7 +76,7 @@ public Builder(JobJobClusterArgs defaults) { * @return builder * */ - public Builder jobClusterKey(@Nullable Output jobClusterKey) { + public Builder jobClusterKey(Output jobClusterKey) { $.jobClusterKey = jobClusterKey; return this; } @@ -98,7 +97,7 @@ public Builder jobClusterKey(String jobClusterKey) { * @return builder * */ - public Builder newCluster(@Nullable Output newCluster) { + public Builder newCluster(Output newCluster) { $.newCluster = newCluster; return this; } @@ -114,6 +113,12 @@ public Builder newCluster(JobJobClusterNewClusterArgs newCluster) { } public JobJobClusterArgs build() { + if ($.jobClusterKey == null) { + throw new MissingRequiredPropertyException("JobJobClusterArgs", "jobClusterKey"); + } + if ($.newCluster == null) { + throw new MissingRequiredPropertyException("JobJobClusterArgs", "newCluster"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterArgs.java index 5c104fc8..113b9947 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterArgs.java @@ -45,13 +45,6 @@ public Optional> autoscale() { return Optional.ofNullable(this.autoscale); } - @Import(name="autoterminationMinutes") - private @Nullable Output autoterminationMinutes; - - public Optional> autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } - @Import(name="awsAttributes") private @Nullable Output awsAttributes; @@ -261,7 +254,6 @@ private JobJobClusterNewClusterArgs() {} private JobJobClusterNewClusterArgs(JobJobClusterNewClusterArgs $) { this.applyPolicyDefaultValues = $.applyPolicyDefaultValues; this.autoscale = $.autoscale; - this.autoterminationMinutes = $.autoterminationMinutes; this.awsAttributes = $.awsAttributes; this.azureAttributes = $.azureAttributes; this.clusterId = $.clusterId; @@ -328,15 +320,6 @@ public Builder autoscale(JobJobClusterNewClusterAutoscaleArgs autoscale) { return autoscale(Output.of(autoscale)); } - public Builder autoterminationMinutes(@Nullable Output autoterminationMinutes) { - $.autoterminationMinutes = autoterminationMinutes; - return this; - } - - public Builder autoterminationMinutes(Integer autoterminationMinutes) { - return autoterminationMinutes(Output.of(autoterminationMinutes)); - } - public Builder awsAttributes(@Nullable Output awsAttributes) { $.awsAttributes = awsAttributes; return this; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterArgs.java index b73aae2d..c2f99041 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterArgs.java @@ -45,13 +45,6 @@ public Optional> autoscale() { return Optional.ofNullable(this.autoscale); } - @Import(name="autoterminationMinutes") - private @Nullable Output autoterminationMinutes; - - public Optional> autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } - @Import(name="awsAttributes") private @Nullable Output awsAttributes; @@ -261,7 +254,6 @@ private JobNewClusterArgs() {} private JobNewClusterArgs(JobNewClusterArgs $) { this.applyPolicyDefaultValues = $.applyPolicyDefaultValues; this.autoscale = $.autoscale; - this.autoterminationMinutes = $.autoterminationMinutes; this.awsAttributes = $.awsAttributes; this.azureAttributes = $.azureAttributes; this.clusterId = $.clusterId; @@ -328,15 +320,6 @@ public Builder autoscale(JobNewClusterAutoscaleArgs autoscale) { return autoscale(Output.of(autoscale)); } - public Builder autoterminationMinutes(@Nullable Output autoterminationMinutes) { - $.autoterminationMinutes = autoterminationMinutes; - return this; - } - - public Builder autoterminationMinutes(Integer autoterminationMinutes) { - return autoterminationMinutes(Output.of(autoterminationMinutes)); - } - public Builder awsAttributes(@Nullable Output awsAttributes) { $.awsAttributes = awsAttributes; return this; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskArgs.java index 20bd0999..63945227 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskArgs.java @@ -23,6 +23,7 @@ import com.pulumi.databricks.inputs.JobTaskSparkSubmitTaskArgs; import com.pulumi.databricks.inputs.JobTaskSqlTaskArgs; import com.pulumi.databricks.inputs.JobTaskWebhookNotificationsArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Boolean; import java.lang.Integer; import java.lang.String; @@ -331,16 +332,16 @@ public Optional> sqlTask() { * * `*_task` - (Required) one of the specific task blocks described below: * */ - @Import(name="taskKey") - private @Nullable Output taskKey; + @Import(name="taskKey", required=true) + private Output taskKey; /** * @return string specifying an unique key for a given task. * * `*_task` - (Required) one of the specific task blocks described below: * */ - public Optional> taskKey() { - return Optional.ofNullable(this.taskKey); + public Output taskKey() { + return this.taskKey; } /** @@ -846,7 +847,7 @@ public Builder sqlTask(JobTaskSqlTaskArgs sqlTask) { * @return builder * */ - public Builder taskKey(@Nullable Output taskKey) { + public Builder taskKey(Output taskKey) { $.taskKey = taskKey; return this; } @@ -905,6 +906,9 @@ public Builder webhookNotifications(JobTaskWebhookNotificationsArgs webhookNotif } public JobTaskArgs build() { + if ($.taskKey == null) { + throw new MissingRequiredPropertyException("JobTaskArgs", "taskKey"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskArgs.java index 2aecbd19..47f3a4eb 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskArgs.java @@ -22,6 +22,7 @@ import com.pulumi.databricks.inputs.JobTaskForEachTaskTaskSparkSubmitTaskArgs; import com.pulumi.databricks.inputs.JobTaskForEachTaskTaskSqlTaskArgs; import com.pulumi.databricks.inputs.JobTaskForEachTaskTaskWebhookNotificationsArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Boolean; import java.lang.Integer; import java.lang.String; @@ -323,16 +324,16 @@ public Optional> sqlTask() { * * `*_task` - (Required) one of the specific task blocks described below: * */ - @Import(name="taskKey") - private @Nullable Output taskKey; + @Import(name="taskKey", required=true) + private Output taskKey; /** * @return string specifying an unique key for a given task. * * `*_task` - (Required) one of the specific task blocks described below: * */ - public Optional> taskKey() { - return Optional.ofNullable(this.taskKey); + public Output taskKey() { + return this.taskKey; } /** @@ -828,7 +829,7 @@ public Builder sqlTask(JobTaskForEachTaskTaskSqlTaskArgs sqlTask) { * @return builder * */ - public Builder taskKey(@Nullable Output taskKey) { + public Builder taskKey(Output taskKey) { $.taskKey = taskKey; return this; } @@ -887,6 +888,9 @@ public Builder webhookNotifications(JobTaskForEachTaskTaskWebhookNotificationsAr } public JobTaskForEachTaskTaskArgs build() { + if ($.taskKey == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskArgs", "taskKey"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskHealthRuleArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskHealthRuleArgs.java index 16ce499f..869039fa 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskHealthRuleArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskHealthRuleArgs.java @@ -5,11 +5,10 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class JobTaskForEachTaskTaskHealthRuleArgs extends com.pulumi.resources.ResourceArgs { @@ -20,45 +19,45 @@ public final class JobTaskForEachTaskTaskHealthRuleArgs extends com.pulumi.resou * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - @Import(name="metric") - private @Nullable Output metric; + @Import(name="metric", required=true) + private Output metric; /** * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - public Optional> metric() { - return Optional.ofNullable(this.metric); + public Output metric() { + return this.metric; } /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - @Import(name="op") - private @Nullable Output op; + @Import(name="op", required=true) + private Output op; /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - public Optional> op() { - return Optional.ofNullable(this.op); + public Output op() { + return this.op; } /** * integer value used to compare to the given metric. * */ - @Import(name="value") - private @Nullable Output value; + @Import(name="value", required=true) + private Output value; /** * @return integer value used to compare to the given metric. * */ - public Optional> value() { - return Optional.ofNullable(this.value); + public Output value() { + return this.value; } private JobTaskForEachTaskTaskHealthRuleArgs() {} @@ -93,7 +92,7 @@ public Builder(JobTaskForEachTaskTaskHealthRuleArgs defaults) { * @return builder * */ - public Builder metric(@Nullable Output metric) { + public Builder metric(Output metric) { $.metric = metric; return this; } @@ -114,7 +113,7 @@ public Builder metric(String metric) { * @return builder * */ - public Builder op(@Nullable Output op) { + public Builder op(Output op) { $.op = op; return this; } @@ -135,7 +134,7 @@ public Builder op(String op) { * @return builder * */ - public Builder value(@Nullable Output value) { + public Builder value(Output value) { $.value = value; return this; } @@ -151,6 +150,15 @@ public Builder value(Integer value) { } public JobTaskForEachTaskTaskHealthRuleArgs build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskHealthRuleArgs", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskHealthRuleArgs", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskHealthRuleArgs", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskNewClusterArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskNewClusterArgs.java index db7c5234..f9de2cb2 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskNewClusterArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskNewClusterArgs.java @@ -45,13 +45,6 @@ public Optional> autoscale return Optional.ofNullable(this.autoscale); } - @Import(name="autoterminationMinutes") - private @Nullable Output autoterminationMinutes; - - public Optional> autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } - @Import(name="awsAttributes") private @Nullable Output awsAttributes; @@ -261,7 +254,6 @@ private JobTaskForEachTaskTaskNewClusterArgs() {} private JobTaskForEachTaskTaskNewClusterArgs(JobTaskForEachTaskTaskNewClusterArgs $) { this.applyPolicyDefaultValues = $.applyPolicyDefaultValues; this.autoscale = $.autoscale; - this.autoterminationMinutes = $.autoterminationMinutes; this.awsAttributes = $.awsAttributes; this.azureAttributes = $.azureAttributes; this.clusterId = $.clusterId; @@ -328,15 +320,6 @@ public Builder autoscale(JobTaskForEachTaskTaskNewClusterAutoscaleArgs autoscale return autoscale(Output.of(autoscale)); } - public Builder autoterminationMinutes(@Nullable Output autoterminationMinutes) { - $.autoterminationMinutes = autoterminationMinutes; - return this; - } - - public Builder autoterminationMinutes(Integer autoterminationMinutes) { - return autoterminationMinutes(Output.of(autoterminationMinutes)); - } - public Builder awsAttributes(@Nullable Output awsAttributes) { $.awsAttributes = awsAttributes; return this; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskSqlTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskSqlTaskArgs.java index b6ee10d2..b2243c1e 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskSqlTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskForEachTaskTaskSqlTaskArgs.java @@ -9,6 +9,7 @@ import com.pulumi.databricks.inputs.JobTaskForEachTaskTaskSqlTaskDashboardArgs; import com.pulumi.databricks.inputs.JobTaskForEachTaskTaskSqlTaskFileArgs; import com.pulumi.databricks.inputs.JobTaskForEachTaskTaskSqlTaskQueryArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -100,15 +101,15 @@ public Optional> query() { * ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - @Import(name="warehouseId") - private @Nullable Output warehouseId; + @Import(name="warehouseId", required=true) + private Output warehouseId; /** * @return ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - public Optional> warehouseId() { - return Optional.ofNullable(this.warehouseId); + public Output warehouseId() { + return this.warehouseId; } private JobTaskForEachTaskTaskSqlTaskArgs() {} @@ -251,7 +252,7 @@ public Builder query(JobTaskForEachTaskTaskSqlTaskQueryArgs query) { * @return builder * */ - public Builder warehouseId(@Nullable Output warehouseId) { + public Builder warehouseId(Output warehouseId) { $.warehouseId = warehouseId; return this; } @@ -267,6 +268,9 @@ public Builder warehouseId(String warehouseId) { } public JobTaskForEachTaskTaskSqlTaskArgs build() { + if ($.warehouseId == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskSqlTaskArgs", "warehouseId"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskHealthRuleArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskHealthRuleArgs.java index 2b781ff4..d3ebe8c6 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskHealthRuleArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskHealthRuleArgs.java @@ -5,11 +5,10 @@ import com.pulumi.core.Output; import com.pulumi.core.annotations.Import; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; public final class JobTaskHealthRuleArgs extends com.pulumi.resources.ResourceArgs { @@ -20,45 +19,45 @@ public final class JobTaskHealthRuleArgs extends com.pulumi.resources.ResourceAr * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - @Import(name="metric") - private @Nullable Output metric; + @Import(name="metric", required=true) + private Output metric; /** * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - public Optional> metric() { - return Optional.ofNullable(this.metric); + public Output metric() { + return this.metric; } /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - @Import(name="op") - private @Nullable Output op; + @Import(name="op", required=true) + private Output op; /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - public Optional> op() { - return Optional.ofNullable(this.op); + public Output op() { + return this.op; } /** * integer value used to compare to the given metric. * */ - @Import(name="value") - private @Nullable Output value; + @Import(name="value", required=true) + private Output value; /** * @return integer value used to compare to the given metric. * */ - public Optional> value() { - return Optional.ofNullable(this.value); + public Output value() { + return this.value; } private JobTaskHealthRuleArgs() {} @@ -93,7 +92,7 @@ public Builder(JobTaskHealthRuleArgs defaults) { * @return builder * */ - public Builder metric(@Nullable Output metric) { + public Builder metric(Output metric) { $.metric = metric; return this; } @@ -114,7 +113,7 @@ public Builder metric(String metric) { * @return builder * */ - public Builder op(@Nullable Output op) { + public Builder op(Output op) { $.op = op; return this; } @@ -135,7 +134,7 @@ public Builder op(String op) { * @return builder * */ - public Builder value(@Nullable Output value) { + public Builder value(Output value) { $.value = value; return this; } @@ -151,6 +150,15 @@ public Builder value(Integer value) { } public JobTaskHealthRuleArgs build() { + if ($.metric == null) { + throw new MissingRequiredPropertyException("JobTaskHealthRuleArgs", "metric"); + } + if ($.op == null) { + throw new MissingRequiredPropertyException("JobTaskHealthRuleArgs", "op"); + } + if ($.value == null) { + throw new MissingRequiredPropertyException("JobTaskHealthRuleArgs", "value"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterArgs.java index c22f6fb2..e85426bb 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterArgs.java @@ -45,13 +45,6 @@ public Optional> autoscale() { return Optional.ofNullable(this.autoscale); } - @Import(name="autoterminationMinutes") - private @Nullable Output autoterminationMinutes; - - public Optional> autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } - @Import(name="awsAttributes") private @Nullable Output awsAttributes; @@ -261,7 +254,6 @@ private JobTaskNewClusterArgs() {} private JobTaskNewClusterArgs(JobTaskNewClusterArgs $) { this.applyPolicyDefaultValues = $.applyPolicyDefaultValues; this.autoscale = $.autoscale; - this.autoterminationMinutes = $.autoterminationMinutes; this.awsAttributes = $.awsAttributes; this.azureAttributes = $.azureAttributes; this.clusterId = $.clusterId; @@ -328,15 +320,6 @@ public Builder autoscale(JobTaskNewClusterAutoscaleArgs autoscale) { return autoscale(Output.of(autoscale)); } - public Builder autoterminationMinutes(@Nullable Output autoterminationMinutes) { - $.autoterminationMinutes = autoterminationMinutes; - return this; - } - - public Builder autoterminationMinutes(Integer autoterminationMinutes) { - return autoterminationMinutes(Output.of(autoterminationMinutes)); - } - public Builder awsAttributes(@Nullable Output awsAttributes) { $.awsAttributes = awsAttributes; return this; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskSqlTaskArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskSqlTaskArgs.java index 6a7375fa..9d821568 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskSqlTaskArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskSqlTaskArgs.java @@ -9,6 +9,7 @@ import com.pulumi.databricks.inputs.JobTaskSqlTaskDashboardArgs; import com.pulumi.databricks.inputs.JobTaskSqlTaskFileArgs; import com.pulumi.databricks.inputs.JobTaskSqlTaskQueryArgs; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -100,15 +101,15 @@ public Optional> query() { * ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - @Import(name="warehouseId") - private @Nullable Output warehouseId; + @Import(name="warehouseId", required=true) + private Output warehouseId; /** * @return ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - public Optional> warehouseId() { - return Optional.ofNullable(this.warehouseId); + public Output warehouseId() { + return this.warehouseId; } private JobTaskSqlTaskArgs() {} @@ -251,7 +252,7 @@ public Builder query(JobTaskSqlTaskQueryArgs query) { * @return builder * */ - public Builder warehouseId(@Nullable Output warehouseId) { + public Builder warehouseId(Output warehouseId) { $.warehouseId = warehouseId; return this; } @@ -267,6 +268,9 @@ public Builder warehouseId(String warehouseId) { } public JobTaskSqlTaskArgs build() { + if ($.warehouseId == null) { + throw new MissingRequiredPropertyException("JobTaskSqlTaskArgs", "warehouseId"); + } return $; } } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerArgs.java index 4b82ac29..5aaed5e2 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerArgs.java @@ -55,17 +55,9 @@ public Optional> table() { return Optional.ofNullable(this.table); } - /** - * configuration block to define a trigger for Table Update events consisting of following attributes: - * - */ @Import(name="tableUpdate") private @Nullable Output tableUpdate; - /** - * @return configuration block to define a trigger for Table Update events consisting of following attributes: - * - */ public Optional> tableUpdate() { return Optional.ofNullable(this.tableUpdate); } @@ -148,23 +140,11 @@ public Builder table(JobTriggerTableArgs table) { return table(Output.of(table)); } - /** - * @param tableUpdate configuration block to define a trigger for Table Update events consisting of following attributes: - * - * @return builder - * - */ public Builder tableUpdate(@Nullable Output tableUpdate) { $.tableUpdate = tableUpdate; return this; } - /** - * @param tableUpdate configuration block to define a trigger for Table Update events consisting of following attributes: - * - * @return builder - * - */ public Builder tableUpdate(JobTriggerTableUpdateArgs tableUpdate) { return tableUpdate(Output.of(tableUpdate)); } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerTableUpdateArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerTableUpdateArgs.java index 79fe9fc9..69c9c8d7 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerTableUpdateArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerTableUpdateArgs.java @@ -18,62 +18,30 @@ public final class JobTriggerTableUpdateArgs extends com.pulumi.resources.Resour public static final JobTriggerTableUpdateArgs Empty = new JobTriggerTableUpdateArgs(); - /** - * The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - * - */ @Import(name="condition") private @Nullable Output condition; - /** - * @return The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - * - */ public Optional> condition() { return Optional.ofNullable(this.condition); } - /** - * If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - * - */ @Import(name="minTimeBetweenTriggersSeconds") private @Nullable Output minTimeBetweenTriggersSeconds; - /** - * @return If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - * - */ public Optional> minTimeBetweenTriggersSeconds() { return Optional.ofNullable(this.minTimeBetweenTriggersSeconds); } - /** - * A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - * - */ @Import(name="tableNames", required=true) private Output> tableNames; - /** - * @return A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - * - */ public Output> tableNames() { return this.tableNames; } - /** - * If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - * - */ @Import(name="waitAfterLastChangeSeconds") private @Nullable Output waitAfterLastChangeSeconds; - /** - * @return If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - * - */ public Optional> waitAfterLastChangeSeconds() { return Optional.ofNullable(this.waitAfterLastChangeSeconds); } @@ -105,96 +73,42 @@ public Builder(JobTriggerTableUpdateArgs defaults) { $ = new JobTriggerTableUpdateArgs(Objects.requireNonNull(defaults)); } - /** - * @param condition The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - * - * @return builder - * - */ public Builder condition(@Nullable Output condition) { $.condition = condition; return this; } - /** - * @param condition The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - * - * @return builder - * - */ public Builder condition(String condition) { return condition(Output.of(condition)); } - /** - * @param minTimeBetweenTriggersSeconds If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - * - * @return builder - * - */ public Builder minTimeBetweenTriggersSeconds(@Nullable Output minTimeBetweenTriggersSeconds) { $.minTimeBetweenTriggersSeconds = minTimeBetweenTriggersSeconds; return this; } - /** - * @param minTimeBetweenTriggersSeconds If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - * - * @return builder - * - */ public Builder minTimeBetweenTriggersSeconds(Integer minTimeBetweenTriggersSeconds) { return minTimeBetweenTriggersSeconds(Output.of(minTimeBetweenTriggersSeconds)); } - /** - * @param tableNames A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - * - * @return builder - * - */ public Builder tableNames(Output> tableNames) { $.tableNames = tableNames; return this; } - /** - * @param tableNames A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - * - * @return builder - * - */ public Builder tableNames(List tableNames) { return tableNames(Output.of(tableNames)); } - /** - * @param tableNames A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - * - * @return builder - * - */ public Builder tableNames(String... tableNames) { return tableNames(List.of(tableNames)); } - /** - * @param waitAfterLastChangeSeconds If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - * - * @return builder - * - */ public Builder waitAfterLastChangeSeconds(@Nullable Output waitAfterLastChangeSeconds) { $.waitAfterLastChangeSeconds = waitAfterLastChangeSeconds; return this; } - /** - * @param waitAfterLastChangeSeconds If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - * - * @return builder - * - */ public Builder waitAfterLastChangeSeconds(Integer waitAfterLastChangeSeconds) { return waitAfterLastChangeSeconds(Output.of(waitAfterLastChangeSeconds)); } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsHealthRule.java index 1b105b40..6eeadc3c 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsHealthRule.java @@ -4,27 +4,26 @@ package com.pulumi.databricks.outputs; import com.pulumi.core.annotations.CustomType; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class GetJobJobSettingsSettingsHealthRule { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; private GetJobJobSettingsSettingsHealthRule() {} - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } public static Builder builder() { @@ -36,9 +35,9 @@ public static Builder builder(GetJobJobSettingsSettingsHealthRule defaults) { } @CustomType.Builder public static final class Builder { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; public Builder() {} public Builder(GetJobJobSettingsSettingsHealthRule defaults) { Objects.requireNonNull(defaults); @@ -48,20 +47,26 @@ public Builder(GetJobJobSettingsSettingsHealthRule defaults) { } @CustomType.Setter - public Builder metric(@Nullable String metric) { - + public Builder metric(String metric) { + if (metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRule", "metric"); + } this.metric = metric; return this; } @CustomType.Setter - public Builder op(@Nullable String op) { - + public Builder op(String op) { + if (op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRule", "op"); + } this.op = op; return this; } @CustomType.Setter - public Builder value(@Nullable Integer value) { - + public Builder value(Integer value) { + if (value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsHealthRule", "value"); + } this.value = value; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsJobCluster.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsJobCluster.java index b1e64a7c..8360b14b 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsJobCluster.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsJobCluster.java @@ -5,22 +5,21 @@ import com.pulumi.core.annotations.CustomType; import com.pulumi.databricks.outputs.GetJobJobSettingsSettingsJobClusterNewCluster; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class GetJobJobSettingsSettingsJobCluster { - private @Nullable String jobClusterKey; - private @Nullable GetJobJobSettingsSettingsJobClusterNewCluster newCluster; + private String jobClusterKey; + private GetJobJobSettingsSettingsJobClusterNewCluster newCluster; private GetJobJobSettingsSettingsJobCluster() {} - public Optional jobClusterKey() { - return Optional.ofNullable(this.jobClusterKey); + public String jobClusterKey() { + return this.jobClusterKey; } - public Optional newCluster() { - return Optional.ofNullable(this.newCluster); + public GetJobJobSettingsSettingsJobClusterNewCluster newCluster() { + return this.newCluster; } public static Builder builder() { @@ -32,8 +31,8 @@ public static Builder builder(GetJobJobSettingsSettingsJobCluster defaults) { } @CustomType.Builder public static final class Builder { - private @Nullable String jobClusterKey; - private @Nullable GetJobJobSettingsSettingsJobClusterNewCluster newCluster; + private String jobClusterKey; + private GetJobJobSettingsSettingsJobClusterNewCluster newCluster; public Builder() {} public Builder(GetJobJobSettingsSettingsJobCluster defaults) { Objects.requireNonNull(defaults); @@ -42,14 +41,18 @@ public Builder(GetJobJobSettingsSettingsJobCluster defaults) { } @CustomType.Setter - public Builder jobClusterKey(@Nullable String jobClusterKey) { - + public Builder jobClusterKey(String jobClusterKey) { + if (jobClusterKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsJobCluster", "jobClusterKey"); + } this.jobClusterKey = jobClusterKey; return this; } @CustomType.Setter - public Builder newCluster(@Nullable GetJobJobSettingsSettingsJobClusterNewCluster newCluster) { - + public Builder newCluster(GetJobJobSettingsSettingsJobClusterNewCluster newCluster) { + if (newCluster == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsJobCluster", "newCluster"); + } this.newCluster = newCluster; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTask.java index 63573183..c1c1fc8c 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTask.java @@ -58,7 +58,7 @@ public final class GetJobJobSettingsSettingsTask { private @Nullable GetJobJobSettingsSettingsTaskSparkPythonTask sparkPythonTask; private @Nullable GetJobJobSettingsSettingsTaskSparkSubmitTask sparkSubmitTask; private @Nullable GetJobJobSettingsSettingsTaskSqlTask sqlTask; - private @Nullable String taskKey; + private String taskKey; private @Nullable Integer timeoutSeconds; private @Nullable GetJobJobSettingsSettingsTaskWebhookNotifications webhookNotifications; @@ -138,8 +138,8 @@ public Optional sparkSubmitTask() public Optional sqlTask() { return Optional.ofNullable(this.sqlTask); } - public Optional taskKey() { - return Optional.ofNullable(this.taskKey); + public String taskKey() { + return this.taskKey; } public Optional timeoutSeconds() { return Optional.ofNullable(this.timeoutSeconds); @@ -182,7 +182,7 @@ public static final class Builder { private @Nullable GetJobJobSettingsSettingsTaskSparkPythonTask sparkPythonTask; private @Nullable GetJobJobSettingsSettingsTaskSparkSubmitTask sparkSubmitTask; private @Nullable GetJobJobSettingsSettingsTaskSqlTask sqlTask; - private @Nullable String taskKey; + private String taskKey; private @Nullable Integer timeoutSeconds; private @Nullable GetJobJobSettingsSettingsTaskWebhookNotifications webhookNotifications; public Builder() {} @@ -377,8 +377,10 @@ public Builder sqlTask(@Nullable GetJobJobSettingsSettingsTaskSqlTask sqlTask) { return this; } @CustomType.Setter - public Builder taskKey(@Nullable String taskKey) { - + public Builder taskKey(String taskKey) { + if (taskKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTask", "taskKey"); + } this.taskKey = taskKey; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java index 8db93a1d..808eabb6 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTask.java @@ -56,7 +56,7 @@ public final class GetJobJobSettingsSettingsTaskForEachTaskTask { private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTask sparkPythonTask; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTask sparkSubmitTask; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask sqlTask; - private @Nullable String taskKey; + private String taskKey; private @Nullable Integer timeoutSeconds; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotifications webhookNotifications; @@ -133,8 +133,8 @@ public Optional spa public Optional sqlTask() { return Optional.ofNullable(this.sqlTask); } - public Optional taskKey() { - return Optional.ofNullable(this.taskKey); + public String taskKey() { + return this.taskKey; } public Optional timeoutSeconds() { return Optional.ofNullable(this.timeoutSeconds); @@ -176,7 +176,7 @@ public static final class Builder { private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTask sparkPythonTask; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTask sparkSubmitTask; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask sqlTask; - private @Nullable String taskKey; + private String taskKey; private @Nullable Integer timeoutSeconds; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotifications webhookNotifications; public Builder() {} @@ -364,8 +364,10 @@ public Builder sqlTask(@Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSql return this; } @CustomType.Setter - public Builder taskKey(@Nullable String taskKey) { - + public Builder taskKey(String taskKey) { + if (taskKey == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTask", "taskKey"); + } this.taskKey = taskKey; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java index de6d44b8..df365c8b 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule.java @@ -4,27 +4,26 @@ package com.pulumi.databricks.outputs; import com.pulumi.core.annotations.CustomType; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; private GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule() {} - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } public static Builder builder() { @@ -36,9 +35,9 @@ public static Builder builder(GetJobJobSettingsSettingsTaskForEachTaskTaskHealth } @CustomType.Builder public static final class Builder { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; public Builder() {} public Builder(GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule defaults) { Objects.requireNonNull(defaults); @@ -48,20 +47,26 @@ public Builder(GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule defaults) } @CustomType.Setter - public Builder metric(@Nullable String metric) { - + public Builder metric(String metric) { + if (metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule", "metric"); + } this.metric = metric; return this; } @CustomType.Setter - public Builder op(@Nullable String op) { - + public Builder op(String op) { + if (op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule", "op"); + } this.op = op; return this; } @CustomType.Setter - public Builder value(@Nullable Integer value) { - + public Builder value(Integer value) { + if (value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule", "value"); + } this.value = value; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java index e48d8256..ef64fbc7 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask.java @@ -8,6 +8,7 @@ import com.pulumi.databricks.outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskDashboard; import com.pulumi.databricks.outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFile; import com.pulumi.databricks.outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQuery; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -22,7 +23,7 @@ public final class GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask { private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFile file; private @Nullable Map parameters; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQuery query; - private @Nullable String warehouseId; + private String warehouseId; private GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask() {} public Optional alert() { @@ -40,8 +41,8 @@ public Map parameters() { public Optional query() { return Optional.ofNullable(this.query); } - public Optional warehouseId() { - return Optional.ofNullable(this.warehouseId); + public String warehouseId() { + return this.warehouseId; } public static Builder builder() { @@ -58,7 +59,7 @@ public static final class Builder { private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFile file; private @Nullable Map parameters; private @Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQuery query; - private @Nullable String warehouseId; + private String warehouseId; public Builder() {} public Builder(GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask defaults) { Objects.requireNonNull(defaults); @@ -101,8 +102,10 @@ public Builder query(@Nullable GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTa return this; } @CustomType.Setter - public Builder warehouseId(@Nullable String warehouseId) { - + public Builder warehouseId(String warehouseId) { + if (warehouseId == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask", "warehouseId"); + } this.warehouseId = warehouseId; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskHealthRule.java index 50600513..4011b8b9 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskHealthRule.java @@ -4,27 +4,26 @@ package com.pulumi.databricks.outputs; import com.pulumi.core.annotations.CustomType; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class GetJobJobSettingsSettingsTaskHealthRule { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; private GetJobJobSettingsSettingsTaskHealthRule() {} - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } public static Builder builder() { @@ -36,9 +35,9 @@ public static Builder builder(GetJobJobSettingsSettingsTaskHealthRule defaults) } @CustomType.Builder public static final class Builder { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; public Builder() {} public Builder(GetJobJobSettingsSettingsTaskHealthRule defaults) { Objects.requireNonNull(defaults); @@ -48,20 +47,26 @@ public Builder(GetJobJobSettingsSettingsTaskHealthRule defaults) { } @CustomType.Setter - public Builder metric(@Nullable String metric) { - + public Builder metric(String metric) { + if (metric == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRule", "metric"); + } this.metric = metric; return this; } @CustomType.Setter - public Builder op(@Nullable String op) { - + public Builder op(String op) { + if (op == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRule", "op"); + } this.op = op; return this; } @CustomType.Setter - public Builder value(@Nullable Integer value) { - + public Builder value(Integer value) { + if (value == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskHealthRule", "value"); + } this.value = value; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskSqlTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskSqlTask.java index 9c3dd626..2553012d 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskSqlTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/GetJobJobSettingsSettingsTaskSqlTask.java @@ -8,6 +8,7 @@ import com.pulumi.databricks.outputs.GetJobJobSettingsSettingsTaskSqlTaskDashboard; import com.pulumi.databricks.outputs.GetJobJobSettingsSettingsTaskSqlTaskFile; import com.pulumi.databricks.outputs.GetJobJobSettingsSettingsTaskSqlTaskQuery; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -22,7 +23,7 @@ public final class GetJobJobSettingsSettingsTaskSqlTask { private @Nullable GetJobJobSettingsSettingsTaskSqlTaskFile file; private @Nullable Map parameters; private @Nullable GetJobJobSettingsSettingsTaskSqlTaskQuery query; - private @Nullable String warehouseId; + private String warehouseId; private GetJobJobSettingsSettingsTaskSqlTask() {} public Optional alert() { @@ -40,8 +41,8 @@ public Map parameters() { public Optional query() { return Optional.ofNullable(this.query); } - public Optional warehouseId() { - return Optional.ofNullable(this.warehouseId); + public String warehouseId() { + return this.warehouseId; } public static Builder builder() { @@ -58,7 +59,7 @@ public static final class Builder { private @Nullable GetJobJobSettingsSettingsTaskSqlTaskFile file; private @Nullable Map parameters; private @Nullable GetJobJobSettingsSettingsTaskSqlTaskQuery query; - private @Nullable String warehouseId; + private String warehouseId; public Builder() {} public Builder(GetJobJobSettingsSettingsTaskSqlTask defaults) { Objects.requireNonNull(defaults); @@ -101,8 +102,10 @@ public Builder query(@Nullable GetJobJobSettingsSettingsTaskSqlTaskQuery query) return this; } @CustomType.Setter - public Builder warehouseId(@Nullable String warehouseId) { - + public Builder warehouseId(String warehouseId) { + if (warehouseId == null) { + throw new MissingRequiredPropertyException("GetJobJobSettingsSettingsTaskSqlTask", "warehouseId"); + } this.warehouseId = warehouseId; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobHealthRule.java index 690051c4..b225fcf9 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobHealthRule.java @@ -4,11 +4,10 @@ package com.pulumi.databricks.outputs; import com.pulumi.core.annotations.CustomType; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class JobHealthRule { @@ -16,39 +15,39 @@ public final class JobHealthRule { * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - private @Nullable String metric; + private String metric; /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - private @Nullable String op; + private String op; /** * @return integer value used to compare to the given metric. * */ - private @Nullable Integer value; + private Integer value; private JobHealthRule() {} /** * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } /** * @return integer value used to compare to the given metric. * */ - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } public static Builder builder() { @@ -60,9 +59,9 @@ public static Builder builder(JobHealthRule defaults) { } @CustomType.Builder public static final class Builder { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; public Builder() {} public Builder(JobHealthRule defaults) { Objects.requireNonNull(defaults); @@ -72,20 +71,26 @@ public Builder(JobHealthRule defaults) { } @CustomType.Setter - public Builder metric(@Nullable String metric) { - + public Builder metric(String metric) { + if (metric == null) { + throw new MissingRequiredPropertyException("JobHealthRule", "metric"); + } this.metric = metric; return this; } @CustomType.Setter - public Builder op(@Nullable String op) { - + public Builder op(String op) { + if (op == null) { + throw new MissingRequiredPropertyException("JobHealthRule", "op"); + } this.op = op; return this; } @CustomType.Setter - public Builder value(@Nullable Integer value) { - + public Builder value(Integer value) { + if (value == null) { + throw new MissingRequiredPropertyException("JobHealthRule", "value"); + } this.value = value; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobCluster.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobCluster.java index f114b841..8d866ad7 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobCluster.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobCluster.java @@ -5,10 +5,9 @@ import com.pulumi.core.annotations.CustomType; import com.pulumi.databricks.outputs.JobJobClusterNewCluster; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class JobJobCluster { @@ -16,27 +15,27 @@ public final class JobJobCluster { * @return Identifier that can be referenced in `task` block, so that cluster is shared between tasks * */ - private @Nullable String jobClusterKey; + private String jobClusterKey; /** * @return Same set of parameters as for databricks.Cluster resource. * */ - private @Nullable JobJobClusterNewCluster newCluster; + private JobJobClusterNewCluster newCluster; private JobJobCluster() {} /** * @return Identifier that can be referenced in `task` block, so that cluster is shared between tasks * */ - public Optional jobClusterKey() { - return Optional.ofNullable(this.jobClusterKey); + public String jobClusterKey() { + return this.jobClusterKey; } /** * @return Same set of parameters as for databricks.Cluster resource. * */ - public Optional newCluster() { - return Optional.ofNullable(this.newCluster); + public JobJobClusterNewCluster newCluster() { + return this.newCluster; } public static Builder builder() { @@ -48,8 +47,8 @@ public static Builder builder(JobJobCluster defaults) { } @CustomType.Builder public static final class Builder { - private @Nullable String jobClusterKey; - private @Nullable JobJobClusterNewCluster newCluster; + private String jobClusterKey; + private JobJobClusterNewCluster newCluster; public Builder() {} public Builder(JobJobCluster defaults) { Objects.requireNonNull(defaults); @@ -58,14 +57,18 @@ public Builder(JobJobCluster defaults) { } @CustomType.Setter - public Builder jobClusterKey(@Nullable String jobClusterKey) { - + public Builder jobClusterKey(String jobClusterKey) { + if (jobClusterKey == null) { + throw new MissingRequiredPropertyException("JobJobCluster", "jobClusterKey"); + } this.jobClusterKey = jobClusterKey; return this; } @CustomType.Setter - public Builder newCluster(@Nullable JobJobClusterNewCluster newCluster) { - + public Builder newCluster(JobJobClusterNewCluster newCluster) { + if (newCluster == null) { + throw new MissingRequiredPropertyException("JobJobCluster", "newCluster"); + } this.newCluster = newCluster; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewCluster.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewCluster.java index fdbd5602..863db446 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewCluster.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewCluster.java @@ -29,7 +29,6 @@ public final class JobJobClusterNewCluster { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobJobClusterNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobJobClusterNewClusterAwsAttributes awsAttributes; private @Nullable JobJobClusterNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -70,9 +69,6 @@ public Optional applyPolicyDefaultValues() { public Optional autoscale() { return Optional.ofNullable(this.autoscale); } - public Optional autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } public Optional awsAttributes() { return Optional.ofNullable(this.awsAttributes); } @@ -173,7 +169,6 @@ public static Builder builder(JobJobClusterNewCluster defaults) { public static final class Builder { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobJobClusterNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobJobClusterNewClusterAwsAttributes awsAttributes; private @Nullable JobJobClusterNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -207,7 +202,6 @@ public Builder(JobJobClusterNewCluster defaults) { Objects.requireNonNull(defaults); this.applyPolicyDefaultValues = defaults.applyPolicyDefaultValues; this.autoscale = defaults.autoscale; - this.autoterminationMinutes = defaults.autoterminationMinutes; this.awsAttributes = defaults.awsAttributes; this.azureAttributes = defaults.azureAttributes; this.clusterId = defaults.clusterId; @@ -251,12 +245,6 @@ public Builder autoscale(@Nullable JobJobClusterNewClusterAutoscale autoscale) { return this; } @CustomType.Setter - public Builder autoterminationMinutes(@Nullable Integer autoterminationMinutes) { - - this.autoterminationMinutes = autoterminationMinutes; - return this; - } - @CustomType.Setter public Builder awsAttributes(@Nullable JobJobClusterNewClusterAwsAttributes awsAttributes) { this.awsAttributes = awsAttributes; @@ -442,7 +430,6 @@ public JobJobClusterNewCluster build() { final var _resultValue = new JobJobClusterNewCluster(); _resultValue.applyPolicyDefaultValues = applyPolicyDefaultValues; _resultValue.autoscale = autoscale; - _resultValue.autoterminationMinutes = autoterminationMinutes; _resultValue.awsAttributes = awsAttributes; _resultValue.azureAttributes = azureAttributes; _resultValue.clusterId = clusterId; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewCluster.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewCluster.java index f11c7311..00ba2c48 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewCluster.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewCluster.java @@ -29,7 +29,6 @@ public final class JobNewCluster { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobNewClusterAwsAttributes awsAttributes; private @Nullable JobNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -70,9 +69,6 @@ public Optional applyPolicyDefaultValues() { public Optional autoscale() { return Optional.ofNullable(this.autoscale); } - public Optional autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } public Optional awsAttributes() { return Optional.ofNullable(this.awsAttributes); } @@ -173,7 +169,6 @@ public static Builder builder(JobNewCluster defaults) { public static final class Builder { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobNewClusterAwsAttributes awsAttributes; private @Nullable JobNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -207,7 +202,6 @@ public Builder(JobNewCluster defaults) { Objects.requireNonNull(defaults); this.applyPolicyDefaultValues = defaults.applyPolicyDefaultValues; this.autoscale = defaults.autoscale; - this.autoterminationMinutes = defaults.autoterminationMinutes; this.awsAttributes = defaults.awsAttributes; this.azureAttributes = defaults.azureAttributes; this.clusterId = defaults.clusterId; @@ -251,12 +245,6 @@ public Builder autoscale(@Nullable JobNewClusterAutoscale autoscale) { return this; } @CustomType.Setter - public Builder autoterminationMinutes(@Nullable Integer autoterminationMinutes) { - - this.autoterminationMinutes = autoterminationMinutes; - return this; - } - @CustomType.Setter public Builder awsAttributes(@Nullable JobNewClusterAwsAttributes awsAttributes) { this.awsAttributes = awsAttributes; @@ -442,7 +430,6 @@ public JobNewCluster build() { final var _resultValue = new JobNewCluster(); _resultValue.applyPolicyDefaultValues = applyPolicyDefaultValues; _resultValue.autoscale = autoscale; - _resultValue.autoterminationMinutes = autoterminationMinutes; _resultValue.awsAttributes = awsAttributes; _resultValue.azureAttributes = azureAttributes; _resultValue.clusterId = clusterId; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTask.java index 70d8ac99..31ad6a1f 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTask.java @@ -22,6 +22,7 @@ import com.pulumi.databricks.outputs.JobTaskSparkSubmitTask; import com.pulumi.databricks.outputs.JobTaskSqlTask; import com.pulumi.databricks.outputs.JobTaskWebhookNotifications; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Boolean; import java.lang.Integer; import java.lang.String; @@ -117,7 +118,7 @@ public final class JobTask { * * `*_task` - (Required) one of the specific task blocks described below: * */ - private @Nullable String taskKey; + private String taskKey; /** * @return (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. * @@ -267,8 +268,8 @@ public Optional sqlTask() { * * `*_task` - (Required) one of the specific task blocks described below: * */ - public Optional taskKey() { - return Optional.ofNullable(this.taskKey); + public String taskKey() { + return this.taskKey; } /** * @return (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. @@ -320,7 +321,7 @@ public static final class Builder { private @Nullable JobTaskSparkPythonTask sparkPythonTask; private @Nullable JobTaskSparkSubmitTask sparkSubmitTask; private @Nullable JobTaskSqlTask sqlTask; - private @Nullable String taskKey; + private String taskKey; private @Nullable Integer timeoutSeconds; private @Nullable JobTaskWebhookNotifications webhookNotifications; public Builder() {} @@ -520,8 +521,10 @@ public Builder sqlTask(@Nullable JobTaskSqlTask sqlTask) { return this; } @CustomType.Setter - public Builder taskKey(@Nullable String taskKey) { - + public Builder taskKey(String taskKey) { + if (taskKey == null) { + throw new MissingRequiredPropertyException("JobTask", "taskKey"); + } this.taskKey = taskKey; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTask.java index 5bd8843e..27f87bf4 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTask.java @@ -21,6 +21,7 @@ import com.pulumi.databricks.outputs.JobTaskForEachTaskTaskSparkSubmitTask; import com.pulumi.databricks.outputs.JobTaskForEachTaskTaskSqlTask; import com.pulumi.databricks.outputs.JobTaskForEachTaskTaskWebhookNotifications; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Boolean; import java.lang.Integer; import java.lang.String; @@ -115,7 +116,7 @@ public final class JobTaskForEachTaskTask { * * `*_task` - (Required) one of the specific task blocks described below: * */ - private @Nullable String taskKey; + private String taskKey; /** * @return (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. * @@ -262,8 +263,8 @@ public Optional sqlTask() { * * `*_task` - (Required) one of the specific task blocks described below: * */ - public Optional taskKey() { - return Optional.ofNullable(this.taskKey); + public String taskKey() { + return this.taskKey; } /** * @return (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. @@ -314,7 +315,7 @@ public static final class Builder { private @Nullable JobTaskForEachTaskTaskSparkPythonTask sparkPythonTask; private @Nullable JobTaskForEachTaskTaskSparkSubmitTask sparkSubmitTask; private @Nullable JobTaskForEachTaskTaskSqlTask sqlTask; - private @Nullable String taskKey; + private String taskKey; private @Nullable Integer timeoutSeconds; private @Nullable JobTaskForEachTaskTaskWebhookNotifications webhookNotifications; public Builder() {} @@ -507,8 +508,10 @@ public Builder sqlTask(@Nullable JobTaskForEachTaskTaskSqlTask sqlTask) { return this; } @CustomType.Setter - public Builder taskKey(@Nullable String taskKey) { - + public Builder taskKey(String taskKey) { + if (taskKey == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTask", "taskKey"); + } this.taskKey = taskKey; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskHealthRule.java index f6af2a91..c5ce4244 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskHealthRule.java @@ -4,11 +4,10 @@ package com.pulumi.databricks.outputs; import com.pulumi.core.annotations.CustomType; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class JobTaskForEachTaskTaskHealthRule { @@ -16,39 +15,39 @@ public final class JobTaskForEachTaskTaskHealthRule { * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - private @Nullable String metric; + private String metric; /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - private @Nullable String op; + private String op; /** * @return integer value used to compare to the given metric. * */ - private @Nullable Integer value; + private Integer value; private JobTaskForEachTaskTaskHealthRule() {} /** * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } /** * @return integer value used to compare to the given metric. * */ - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } public static Builder builder() { @@ -60,9 +59,9 @@ public static Builder builder(JobTaskForEachTaskTaskHealthRule defaults) { } @CustomType.Builder public static final class Builder { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; public Builder() {} public Builder(JobTaskForEachTaskTaskHealthRule defaults) { Objects.requireNonNull(defaults); @@ -72,20 +71,26 @@ public Builder(JobTaskForEachTaskTaskHealthRule defaults) { } @CustomType.Setter - public Builder metric(@Nullable String metric) { - + public Builder metric(String metric) { + if (metric == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskHealthRule", "metric"); + } this.metric = metric; return this; } @CustomType.Setter - public Builder op(@Nullable String op) { - + public Builder op(String op) { + if (op == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskHealthRule", "op"); + } this.op = op; return this; } @CustomType.Setter - public Builder value(@Nullable Integer value) { - + public Builder value(Integer value) { + if (value == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskHealthRule", "value"); + } this.value = value; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskNewCluster.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskNewCluster.java index 92dcb34c..bfc3c14e 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskNewCluster.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskNewCluster.java @@ -29,7 +29,6 @@ public final class JobTaskForEachTaskTaskNewCluster { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobTaskForEachTaskTaskNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobTaskForEachTaskTaskNewClusterAwsAttributes awsAttributes; private @Nullable JobTaskForEachTaskTaskNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -70,9 +69,6 @@ public Optional applyPolicyDefaultValues() { public Optional autoscale() { return Optional.ofNullable(this.autoscale); } - public Optional autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } public Optional awsAttributes() { return Optional.ofNullable(this.awsAttributes); } @@ -173,7 +169,6 @@ public static Builder builder(JobTaskForEachTaskTaskNewCluster defaults) { public static final class Builder { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobTaskForEachTaskTaskNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobTaskForEachTaskTaskNewClusterAwsAttributes awsAttributes; private @Nullable JobTaskForEachTaskTaskNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -207,7 +202,6 @@ public Builder(JobTaskForEachTaskTaskNewCluster defaults) { Objects.requireNonNull(defaults); this.applyPolicyDefaultValues = defaults.applyPolicyDefaultValues; this.autoscale = defaults.autoscale; - this.autoterminationMinutes = defaults.autoterminationMinutes; this.awsAttributes = defaults.awsAttributes; this.azureAttributes = defaults.azureAttributes; this.clusterId = defaults.clusterId; @@ -251,12 +245,6 @@ public Builder autoscale(@Nullable JobTaskForEachTaskTaskNewClusterAutoscale aut return this; } @CustomType.Setter - public Builder autoterminationMinutes(@Nullable Integer autoterminationMinutes) { - - this.autoterminationMinutes = autoterminationMinutes; - return this; - } - @CustomType.Setter public Builder awsAttributes(@Nullable JobTaskForEachTaskTaskNewClusterAwsAttributes awsAttributes) { this.awsAttributes = awsAttributes; @@ -442,7 +430,6 @@ public JobTaskForEachTaskTaskNewCluster build() { final var _resultValue = new JobTaskForEachTaskTaskNewCluster(); _resultValue.applyPolicyDefaultValues = applyPolicyDefaultValues; _resultValue.autoscale = autoscale; - _resultValue.autoterminationMinutes = autoterminationMinutes; _resultValue.awsAttributes = awsAttributes; _resultValue.azureAttributes = azureAttributes; _resultValue.clusterId = clusterId; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskSqlTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskSqlTask.java index abffaa42..0e07d1b4 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskSqlTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskForEachTaskTaskSqlTask.java @@ -8,6 +8,7 @@ import com.pulumi.databricks.outputs.JobTaskForEachTaskTaskSqlTaskDashboard; import com.pulumi.databricks.outputs.JobTaskForEachTaskTaskSqlTaskFile; import com.pulumi.databricks.outputs.JobTaskForEachTaskTaskSqlTaskQuery; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -46,7 +47,7 @@ public final class JobTaskForEachTaskTaskSqlTask { * @return ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - private @Nullable String warehouseId; + private String warehouseId; private JobTaskForEachTaskTaskSqlTask() {} /** @@ -88,8 +89,8 @@ public Optional query() { * @return ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - public Optional warehouseId() { - return Optional.ofNullable(this.warehouseId); + public String warehouseId() { + return this.warehouseId; } public static Builder builder() { @@ -106,7 +107,7 @@ public static final class Builder { private @Nullable JobTaskForEachTaskTaskSqlTaskFile file; private @Nullable Map parameters; private @Nullable JobTaskForEachTaskTaskSqlTaskQuery query; - private @Nullable String warehouseId; + private String warehouseId; public Builder() {} public Builder(JobTaskForEachTaskTaskSqlTask defaults) { Objects.requireNonNull(defaults); @@ -149,8 +150,10 @@ public Builder query(@Nullable JobTaskForEachTaskTaskSqlTaskQuery query) { return this; } @CustomType.Setter - public Builder warehouseId(@Nullable String warehouseId) { - + public Builder warehouseId(String warehouseId) { + if (warehouseId == null) { + throw new MissingRequiredPropertyException("JobTaskForEachTaskTaskSqlTask", "warehouseId"); + } this.warehouseId = warehouseId; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskHealthRule.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskHealthRule.java index 2ffe5ab9..4001be7f 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskHealthRule.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskHealthRule.java @@ -4,11 +4,10 @@ package com.pulumi.databricks.outputs; import com.pulumi.core.annotations.CustomType; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Integer; import java.lang.String; import java.util.Objects; -import java.util.Optional; -import javax.annotation.Nullable; @CustomType public final class JobTaskHealthRule { @@ -16,39 +15,39 @@ public final class JobTaskHealthRule { * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - private @Nullable String metric; + private String metric; /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - private @Nullable String op; + private String op; /** * @return integer value used to compare to the given metric. * */ - private @Nullable Integer value; + private Integer value; private JobTaskHealthRule() {} /** * @return string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). * */ - public Optional metric() { - return Optional.ofNullable(this.metric); + public String metric() { + return this.metric; } /** * @return string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. * */ - public Optional op() { - return Optional.ofNullable(this.op); + public String op() { + return this.op; } /** * @return integer value used to compare to the given metric. * */ - public Optional value() { - return Optional.ofNullable(this.value); + public Integer value() { + return this.value; } public static Builder builder() { @@ -60,9 +59,9 @@ public static Builder builder(JobTaskHealthRule defaults) { } @CustomType.Builder public static final class Builder { - private @Nullable String metric; - private @Nullable String op; - private @Nullable Integer value; + private String metric; + private String op; + private Integer value; public Builder() {} public Builder(JobTaskHealthRule defaults) { Objects.requireNonNull(defaults); @@ -72,20 +71,26 @@ public Builder(JobTaskHealthRule defaults) { } @CustomType.Setter - public Builder metric(@Nullable String metric) { - + public Builder metric(String metric) { + if (metric == null) { + throw new MissingRequiredPropertyException("JobTaskHealthRule", "metric"); + } this.metric = metric; return this; } @CustomType.Setter - public Builder op(@Nullable String op) { - + public Builder op(String op) { + if (op == null) { + throw new MissingRequiredPropertyException("JobTaskHealthRule", "op"); + } this.op = op; return this; } @CustomType.Setter - public Builder value(@Nullable Integer value) { - + public Builder value(Integer value) { + if (value == null) { + throw new MissingRequiredPropertyException("JobTaskHealthRule", "value"); + } this.value = value; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewCluster.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewCluster.java index cae5c2b5..73c1b82a 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewCluster.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewCluster.java @@ -29,7 +29,6 @@ public final class JobTaskNewCluster { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobTaskNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobTaskNewClusterAwsAttributes awsAttributes; private @Nullable JobTaskNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -70,9 +69,6 @@ public Optional applyPolicyDefaultValues() { public Optional autoscale() { return Optional.ofNullable(this.autoscale); } - public Optional autoterminationMinutes() { - return Optional.ofNullable(this.autoterminationMinutes); - } public Optional awsAttributes() { return Optional.ofNullable(this.awsAttributes); } @@ -173,7 +169,6 @@ public static Builder builder(JobTaskNewCluster defaults) { public static final class Builder { private @Nullable Boolean applyPolicyDefaultValues; private @Nullable JobTaskNewClusterAutoscale autoscale; - private @Nullable Integer autoterminationMinutes; private @Nullable JobTaskNewClusterAwsAttributes awsAttributes; private @Nullable JobTaskNewClusterAzureAttributes azureAttributes; private @Nullable String clusterId; @@ -207,7 +202,6 @@ public Builder(JobTaskNewCluster defaults) { Objects.requireNonNull(defaults); this.applyPolicyDefaultValues = defaults.applyPolicyDefaultValues; this.autoscale = defaults.autoscale; - this.autoterminationMinutes = defaults.autoterminationMinutes; this.awsAttributes = defaults.awsAttributes; this.azureAttributes = defaults.azureAttributes; this.clusterId = defaults.clusterId; @@ -251,12 +245,6 @@ public Builder autoscale(@Nullable JobTaskNewClusterAutoscale autoscale) { return this; } @CustomType.Setter - public Builder autoterminationMinutes(@Nullable Integer autoterminationMinutes) { - - this.autoterminationMinutes = autoterminationMinutes; - return this; - } - @CustomType.Setter public Builder awsAttributes(@Nullable JobTaskNewClusterAwsAttributes awsAttributes) { this.awsAttributes = awsAttributes; @@ -442,7 +430,6 @@ public JobTaskNewCluster build() { final var _resultValue = new JobTaskNewCluster(); _resultValue.applyPolicyDefaultValues = applyPolicyDefaultValues; _resultValue.autoscale = autoscale; - _resultValue.autoterminationMinutes = autoterminationMinutes; _resultValue.awsAttributes = awsAttributes; _resultValue.azureAttributes = azureAttributes; _resultValue.clusterId = clusterId; diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskSqlTask.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskSqlTask.java index cb3c319e..182e1e74 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskSqlTask.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskSqlTask.java @@ -8,6 +8,7 @@ import com.pulumi.databricks.outputs.JobTaskSqlTaskDashboard; import com.pulumi.databricks.outputs.JobTaskSqlTaskFile; import com.pulumi.databricks.outputs.JobTaskSqlTaskQuery; +import com.pulumi.exceptions.MissingRequiredPropertyException; import java.lang.Object; import java.lang.String; import java.util.Map; @@ -46,7 +47,7 @@ public final class JobTaskSqlTask { * @return ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - private @Nullable String warehouseId; + private String warehouseId; private JobTaskSqlTask() {} /** @@ -88,8 +89,8 @@ public Optional query() { * @return ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. * */ - public Optional warehouseId() { - return Optional.ofNullable(this.warehouseId); + public String warehouseId() { + return this.warehouseId; } public static Builder builder() { @@ -106,7 +107,7 @@ public static final class Builder { private @Nullable JobTaskSqlTaskFile file; private @Nullable Map parameters; private @Nullable JobTaskSqlTaskQuery query; - private @Nullable String warehouseId; + private String warehouseId; public Builder() {} public Builder(JobTaskSqlTask defaults) { Objects.requireNonNull(defaults); @@ -149,8 +150,10 @@ public Builder query(@Nullable JobTaskSqlTaskQuery query) { return this; } @CustomType.Setter - public Builder warehouseId(@Nullable String warehouseId) { - + public Builder warehouseId(String warehouseId) { + if (warehouseId == null) { + throw new MissingRequiredPropertyException("JobTaskSqlTask", "warehouseId"); + } this.warehouseId = warehouseId; return this; } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTrigger.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTrigger.java index 7092df64..dbd1eb0f 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTrigger.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTrigger.java @@ -25,10 +25,6 @@ public final class JobTrigger { */ private @Nullable String pauseStatus; private @Nullable JobTriggerTable table; - /** - * @return configuration block to define a trigger for Table Update events consisting of following attributes: - * - */ private @Nullable JobTriggerTableUpdate tableUpdate; private JobTrigger() {} @@ -49,10 +45,6 @@ public Optional pauseStatus() { public Optional table() { return Optional.ofNullable(this.table); } - /** - * @return configuration block to define a trigger for Table Update events consisting of following attributes: - * - */ public Optional tableUpdate() { return Optional.ofNullable(this.tableUpdate); } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerTableUpdate.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerTableUpdate.java index 6fd5777b..adfcc316 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerTableUpdate.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerTableUpdate.java @@ -14,53 +14,21 @@ @CustomType public final class JobTriggerTableUpdate { - /** - * @return The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - * - */ private @Nullable String condition; - /** - * @return If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - * - */ private @Nullable Integer minTimeBetweenTriggersSeconds; - /** - * @return A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - * - */ private List tableNames; - /** - * @return If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - * - */ private @Nullable Integer waitAfterLastChangeSeconds; private JobTriggerTableUpdate() {} - /** - * @return The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - * - */ public Optional condition() { return Optional.ofNullable(this.condition); } - /** - * @return If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - * - */ public Optional minTimeBetweenTriggersSeconds() { return Optional.ofNullable(this.minTimeBetweenTriggersSeconds); } - /** - * @return A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - * - */ public List tableNames() { return this.tableNames; } - /** - * @return If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - * - */ public Optional waitAfterLastChangeSeconds() { return Optional.ofNullable(this.waitAfterLastChangeSeconds); } diff --git a/sdk/nodejs/types/input.ts b/sdk/nodejs/types/input.ts index c18ea5a0..08b78342 100644 --- a/sdk/nodejs/types/input.ts +++ b/sdk/nodejs/types/input.ts @@ -1844,25 +1844,25 @@ export interface GetJobJobSettingsSettingsHealthArgs { } export interface GetJobJobSettingsSettingsHealthRule { - metric?: string; - op?: string; - value?: number; + metric: string; + op: string; + value: number; } export interface GetJobJobSettingsSettingsHealthRuleArgs { - metric?: pulumi.Input; - op?: pulumi.Input; - value?: pulumi.Input; + metric: pulumi.Input; + op: pulumi.Input; + value: pulumi.Input; } export interface GetJobJobSettingsSettingsJobCluster { - jobClusterKey?: string; - newCluster?: inputs.GetJobJobSettingsSettingsJobClusterNewCluster; + jobClusterKey: string; + newCluster: inputs.GetJobJobSettingsSettingsJobClusterNewCluster; } export interface GetJobJobSettingsSettingsJobClusterArgs { - jobClusterKey?: pulumi.Input; - newCluster?: pulumi.Input; + jobClusterKey: pulumi.Input; + newCluster: pulumi.Input; } export interface GetJobJobSettingsSettingsJobClusterNewCluster { @@ -2707,7 +2707,7 @@ export interface GetJobJobSettingsSettingsTask { sparkPythonTask?: inputs.GetJobJobSettingsSettingsTaskSparkPythonTask; sparkSubmitTask?: inputs.GetJobJobSettingsSettingsTaskSparkSubmitTask; sqlTask?: inputs.GetJobJobSettingsSettingsTaskSqlTask; - taskKey?: string; + taskKey: string; timeoutSeconds?: number; webhookNotifications?: inputs.GetJobJobSettingsSettingsTaskWebhookNotifications; } @@ -2738,7 +2738,7 @@ export interface GetJobJobSettingsSettingsTaskArgs { sparkPythonTask?: pulumi.Input; sparkSubmitTask?: pulumi.Input; sqlTask?: pulumi.Input; - taskKey?: pulumi.Input; + taskKey: pulumi.Input; timeoutSeconds?: pulumi.Input; webhookNotifications?: pulumi.Input; } @@ -2838,7 +2838,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTask { sparkPythonTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTask; sparkSubmitTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTask; sqlTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask; - taskKey?: string; + taskKey: string; timeoutSeconds?: number; webhookNotifications?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotifications; } @@ -2868,7 +2868,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskArgs { sparkPythonTask?: pulumi.Input; sparkSubmitTask?: pulumi.Input; sqlTask?: pulumi.Input; - taskKey?: pulumi.Input; + taskKey: pulumi.Input; timeoutSeconds?: pulumi.Input; webhookNotifications?: pulumi.Input; } @@ -2940,15 +2940,15 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskHealthArgs { } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule { - metric?: string; - op?: string; - value?: number; + metric: string; + op: string; + value: number; } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs { - metric?: pulumi.Input; - op?: pulumi.Input; - value?: pulumi.Input; + metric: pulumi.Input; + op: pulumi.Input; + value: pulumi.Input; } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskLibrary { @@ -3415,7 +3415,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask { file?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFile; parameters?: {[key: string]: any}; query?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQuery; - warehouseId?: string; + warehouseId: string; } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs { @@ -3424,7 +3424,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs { file?: pulumi.Input; parameters?: pulumi.Input<{[key: string]: any}>; query?: pulumi.Input; - warehouseId?: pulumi.Input; + warehouseId: pulumi.Input; } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlert { @@ -3570,15 +3570,15 @@ export interface GetJobJobSettingsSettingsTaskHealthArgs { } export interface GetJobJobSettingsSettingsTaskHealthRule { - metric?: string; - op?: string; - value?: number; + metric: string; + op: string; + value: number; } export interface GetJobJobSettingsSettingsTaskHealthRuleArgs { - metric?: pulumi.Input; - op?: pulumi.Input; - value?: pulumi.Input; + metric: pulumi.Input; + op: pulumi.Input; + value: pulumi.Input; } export interface GetJobJobSettingsSettingsTaskLibrary { @@ -4045,7 +4045,7 @@ export interface GetJobJobSettingsSettingsTaskSqlTask { file?: inputs.GetJobJobSettingsSettingsTaskSqlTaskFile; parameters?: {[key: string]: any}; query?: inputs.GetJobJobSettingsSettingsTaskSqlTaskQuery; - warehouseId?: string; + warehouseId: string; } export interface GetJobJobSettingsSettingsTaskSqlTaskArgs { @@ -4054,7 +4054,7 @@ export interface GetJobJobSettingsSettingsTaskSqlTaskArgs { file?: pulumi.Input; parameters?: pulumi.Input<{[key: string]: any}>; query?: pulumi.Input; - warehouseId?: pulumi.Input; + warehouseId: pulumi.Input; } export interface GetJobJobSettingsSettingsTaskSqlTaskAlert { @@ -5421,32 +5421,31 @@ export interface JobHealthRule { /** * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). */ - metric?: pulumi.Input; + metric: pulumi.Input; /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. */ - op?: pulumi.Input; + op: pulumi.Input; /** * integer value used to compare to the given metric. */ - value?: pulumi.Input; + value: pulumi.Input; } export interface JobJobCluster { /** * Identifier that can be referenced in `task` block, so that cluster is shared between tasks */ - jobClusterKey?: pulumi.Input; + jobClusterKey: pulumi.Input; /** * Same set of parameters as for databricks.Cluster resource. */ - newCluster?: pulumi.Input; + newCluster: pulumi.Input; } export interface JobJobClusterNewCluster { applyPolicyDefaultValues?: pulumi.Input; autoscale?: pulumi.Input; - autoterminationMinutes?: pulumi.Input; awsAttributes?: pulumi.Input; azureAttributes?: pulumi.Input; clusterId?: pulumi.Input; @@ -5673,7 +5672,6 @@ export interface JobLibraryPypi { export interface JobNewCluster { applyPolicyDefaultValues?: pulumi.Input; autoscale?: pulumi.Input; - autoterminationMinutes?: pulumi.Input; awsAttributes?: pulumi.Input; azureAttributes?: pulumi.Input; clusterId?: pulumi.Input; @@ -6109,7 +6107,7 @@ export interface JobTask { * string specifying an unique key for a given task. * * `*_task` - (Required) one of the specific task blocks described below: */ - taskKey?: pulumi.Input; + taskKey: pulumi.Input; /** * (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. */ @@ -6291,7 +6289,7 @@ export interface JobTaskForEachTaskTask { * string specifying an unique key for a given task. * * `*_task` - (Required) one of the specific task blocks described below: */ - taskKey?: pulumi.Input; + taskKey: pulumi.Input; /** * (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. */ @@ -6398,15 +6396,15 @@ export interface JobTaskForEachTaskTaskHealthRule { /** * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). */ - metric?: pulumi.Input; + metric: pulumi.Input; /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. */ - op?: pulumi.Input; + op: pulumi.Input; /** * integer value used to compare to the given metric. */ - value?: pulumi.Input; + value: pulumi.Input; } export interface JobTaskForEachTaskTaskLibrary { @@ -6438,7 +6436,6 @@ export interface JobTaskForEachTaskTaskLibraryPypi { export interface JobTaskForEachTaskTaskNewCluster { applyPolicyDefaultValues?: pulumi.Input; autoscale?: pulumi.Input; - autoterminationMinutes?: pulumi.Input; awsAttributes?: pulumi.Input; azureAttributes?: pulumi.Input; clusterId?: pulumi.Input; @@ -6785,7 +6782,7 @@ export interface JobTaskForEachTaskTaskSqlTask { /** * ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. */ - warehouseId?: pulumi.Input; + warehouseId: pulumi.Input; } export interface JobTaskForEachTaskTaskSqlTaskAlert { @@ -6955,15 +6952,15 @@ export interface JobTaskHealthRule { /** * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). */ - metric?: pulumi.Input; + metric: pulumi.Input; /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. */ - op?: pulumi.Input; + op: pulumi.Input; /** * integer value used to compare to the given metric. */ - value?: pulumi.Input; + value: pulumi.Input; } export interface JobTaskLibrary { @@ -6995,7 +6992,6 @@ export interface JobTaskLibraryPypi { export interface JobTaskNewCluster { applyPolicyDefaultValues?: pulumi.Input; autoscale?: pulumi.Input; - autoterminationMinutes?: pulumi.Input; awsAttributes?: pulumi.Input; azureAttributes?: pulumi.Input; clusterId?: pulumi.Input; @@ -7342,7 +7338,7 @@ export interface JobTaskSqlTask { /** * ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. */ - warehouseId?: pulumi.Input; + warehouseId: pulumi.Input; } export interface JobTaskSqlTaskAlert { @@ -7511,9 +7507,6 @@ export interface JobTrigger { */ pauseStatus?: pulumi.Input; table?: pulumi.Input; - /** - * configuration block to define a trigger for Table Update events consisting of following attributes: - */ tableUpdate?: pulumi.Input; } @@ -7540,21 +7533,9 @@ export interface JobTriggerTable { } export interface JobTriggerTableUpdate { - /** - * The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - */ condition?: pulumi.Input; - /** - * If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - */ minTimeBetweenTriggersSeconds?: pulumi.Input; - /** - * A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - */ tableNames: pulumi.Input[]>; - /** - * If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - */ waitAfterLastChangeSeconds?: pulumi.Input; } diff --git a/sdk/nodejs/types/output.ts b/sdk/nodejs/types/output.ts index dc90fecc..2e94a312 100644 --- a/sdk/nodejs/types/output.ts +++ b/sdk/nodejs/types/output.ts @@ -1153,14 +1153,14 @@ export interface GetJobJobSettingsSettingsHealth { } export interface GetJobJobSettingsSettingsHealthRule { - metric?: string; - op?: string; - value?: number; + metric: string; + op: string; + value: number; } export interface GetJobJobSettingsSettingsJobCluster { - jobClusterKey?: string; - newCluster?: outputs.GetJobJobSettingsSettingsJobClusterNewCluster; + jobClusterKey: string; + newCluster: outputs.GetJobJobSettingsSettingsJobClusterNewCluster; } export interface GetJobJobSettingsSettingsJobClusterNewCluster { @@ -1597,7 +1597,7 @@ export interface GetJobJobSettingsSettingsTask { sparkPythonTask?: outputs.GetJobJobSettingsSettingsTaskSparkPythonTask; sparkSubmitTask?: outputs.GetJobJobSettingsSettingsTaskSparkSubmitTask; sqlTask?: outputs.GetJobJobSettingsSettingsTaskSqlTask; - taskKey?: string; + taskKey: string; timeoutSeconds?: number; webhookNotifications?: outputs.GetJobJobSettingsSettingsTaskWebhookNotifications; } @@ -1662,7 +1662,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTask { sparkPythonTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTask; sparkSubmitTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTask; sqlTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask; - taskKey?: string; + taskKey: string; timeoutSeconds?: number; webhookNotifications?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotifications; } @@ -1701,9 +1701,9 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskHealth { } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRule { - metric?: string; - op?: string; - value?: number; + metric: string; + op: string; + value: number; } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskLibrary { @@ -1941,7 +1941,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask { file?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFile; parameters?: {[key: string]: any}; query?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQuery; - warehouseId?: string; + warehouseId: string; } export interface GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlert { @@ -2016,9 +2016,9 @@ export interface GetJobJobSettingsSettingsTaskHealth { } export interface GetJobJobSettingsSettingsTaskHealthRule { - metric?: string; - op?: string; - value?: number; + metric: string; + op: string; + value: number; } export interface GetJobJobSettingsSettingsTaskLibrary { @@ -2256,7 +2256,7 @@ export interface GetJobJobSettingsSettingsTaskSqlTask { file?: outputs.GetJobJobSettingsSettingsTaskSqlTaskFile; parameters?: {[key: string]: any}; query?: outputs.GetJobJobSettingsSettingsTaskSqlTaskQuery; - warehouseId?: string; + warehouseId: string; } export interface GetJobJobSettingsSettingsTaskSqlTaskAlert { @@ -3073,32 +3073,31 @@ export interface JobHealthRule { /** * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). */ - metric?: string; + metric: string; /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. */ - op?: string; + op: string; /** * integer value used to compare to the given metric. */ - value?: number; + value: number; } export interface JobJobCluster { /** * Identifier that can be referenced in `task` block, so that cluster is shared between tasks */ - jobClusterKey?: string; + jobClusterKey: string; /** * Same set of parameters as for databricks.Cluster resource. */ - newCluster?: outputs.JobJobClusterNewCluster; + newCluster: outputs.JobJobClusterNewCluster; } export interface JobJobClusterNewCluster { applyPolicyDefaultValues?: boolean; autoscale?: outputs.JobJobClusterNewClusterAutoscale; - autoterminationMinutes?: number; awsAttributes?: outputs.JobJobClusterNewClusterAwsAttributes; azureAttributes?: outputs.JobJobClusterNewClusterAzureAttributes; clusterId?: string; @@ -3325,7 +3324,6 @@ export interface JobLibraryPypi { export interface JobNewCluster { applyPolicyDefaultValues?: boolean; autoscale?: outputs.JobNewClusterAutoscale; - autoterminationMinutes?: number; awsAttributes?: outputs.JobNewClusterAwsAttributes; azureAttributes?: outputs.JobNewClusterAzureAttributes; clusterId?: string; @@ -3761,7 +3759,7 @@ export interface JobTask { * string specifying an unique key for a given task. * * `*_task` - (Required) one of the specific task blocks described below: */ - taskKey?: string; + taskKey: string; /** * (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. */ @@ -3943,7 +3941,7 @@ export interface JobTaskForEachTaskTask { * string specifying an unique key for a given task. * * `*_task` - (Required) one of the specific task blocks described below: */ - taskKey?: string; + taskKey: string; /** * (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. */ @@ -4050,15 +4048,15 @@ export interface JobTaskForEachTaskTaskHealthRule { /** * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). */ - metric?: string; + metric: string; /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. */ - op?: string; + op: string; /** * integer value used to compare to the given metric. */ - value?: number; + value: number; } export interface JobTaskForEachTaskTaskLibrary { @@ -4090,7 +4088,6 @@ export interface JobTaskForEachTaskTaskLibraryPypi { export interface JobTaskForEachTaskTaskNewCluster { applyPolicyDefaultValues?: boolean; autoscale?: outputs.JobTaskForEachTaskTaskNewClusterAutoscale; - autoterminationMinutes?: number; awsAttributes?: outputs.JobTaskForEachTaskTaskNewClusterAwsAttributes; azureAttributes?: outputs.JobTaskForEachTaskTaskNewClusterAzureAttributes; clusterId?: string; @@ -4437,7 +4434,7 @@ export interface JobTaskForEachTaskTaskSqlTask { /** * ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. */ - warehouseId?: string; + warehouseId: string; } export interface JobTaskForEachTaskTaskSqlTaskAlert { @@ -4607,15 +4604,15 @@ export interface JobTaskHealthRule { /** * string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). */ - metric?: string; + metric: string; /** * string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. */ - op?: string; + op: string; /** * integer value used to compare to the given metric. */ - value?: number; + value: number; } export interface JobTaskLibrary { @@ -4647,7 +4644,6 @@ export interface JobTaskLibraryPypi { export interface JobTaskNewCluster { applyPolicyDefaultValues?: boolean; autoscale?: outputs.JobTaskNewClusterAutoscale; - autoterminationMinutes?: number; awsAttributes?: outputs.JobTaskNewClusterAwsAttributes; azureAttributes?: outputs.JobTaskNewClusterAzureAttributes; clusterId?: string; @@ -4994,7 +4990,7 @@ export interface JobTaskSqlTask { /** * ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. */ - warehouseId?: string; + warehouseId: string; } export interface JobTaskSqlTaskAlert { @@ -5163,9 +5159,6 @@ export interface JobTrigger { */ pauseStatus?: string; table?: outputs.JobTriggerTable; - /** - * configuration block to define a trigger for Table Update events consisting of following attributes: - */ tableUpdate?: outputs.JobTriggerTableUpdate; } @@ -5192,21 +5185,9 @@ export interface JobTriggerTable { } export interface JobTriggerTableUpdate { - /** - * The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - */ condition?: string; - /** - * If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - */ minTimeBetweenTriggersSeconds?: number; - /** - * A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - */ tableNames: string[]; - /** - * If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - */ waitAfterLastChangeSeconds?: number; } @@ -5531,11 +5512,11 @@ export interface ModelServingConfigServedEntity { /** * The workload size of the served entity. The workload size corresponds to a range of provisioned concurrency that the compute autoscales between. A single unit of provisioned concurrency can process one request at a time. Valid workload sizes are `Small` (4 - 4 provisioned concurrency), `Medium` (8 - 16 provisioned concurrency), and `Large` (16 - 64 provisioned concurrency). If `scale-to-zero` is enabled, the lower bound of the provisioned concurrency for each workload size is 0. */ - workloadSize?: string; + workloadSize: string; /** * The workload type of the served entity. The workload type selects which type of compute to use in the endpoint. The default value for this parameter is `CPU`. For deep learning workloads, GPU acceleration is available by selecting workload types like `GPU_SMALL` and others. See the available [GPU types](https://docs.databricks.com/machine-learning/model-serving/create-manage-serving-endpoints.html#gpu-workload-types). */ - workloadType?: string; + workloadType: string; } export interface ModelServingConfigServedEntityExternalModel { diff --git a/sdk/python/pulumi_databricks/_inputs.py b/sdk/python/pulumi_databricks/_inputs.py index 1ca274d5..99e8a169 100644 --- a/sdk/python/pulumi_databricks/_inputs.py +++ b/sdk/python/pulumi_databricks/_inputs.py @@ -3648,94 +3648,89 @@ def rules(self, value: pulumi.Input[Sequence[pulumi.Input['JobHealthRuleArgs']]] @pulumi.input_type class JobHealthRuleArgs: def __init__(__self__, *, - metric: Optional[pulumi.Input[str]] = None, - op: Optional[pulumi.Input[str]] = None, - value: Optional[pulumi.Input[int]] = None): + metric: pulumi.Input[str], + op: pulumi.Input[str], + value: pulumi.Input[int]): """ :param pulumi.Input[str] metric: string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). :param pulumi.Input[str] op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param pulumi.Input[int] value: integer value used to compare to the given metric. """ - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[pulumi.Input[str]]: + def metric(self) -> pulumi.Input[str]: """ string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). """ return pulumi.get(self, "metric") @metric.setter - def metric(self, value: Optional[pulumi.Input[str]]): + def metric(self, value: pulumi.Input[str]): pulumi.set(self, "metric", value) @property @pulumi.getter - def op(self) -> Optional[pulumi.Input[str]]: + def op(self) -> pulumi.Input[str]: """ string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ return pulumi.get(self, "op") @op.setter - def op(self, value: Optional[pulumi.Input[str]]): + def op(self, value: pulumi.Input[str]): pulumi.set(self, "op", value) @property @pulumi.getter - def value(self) -> Optional[pulumi.Input[int]]: + def value(self) -> pulumi.Input[int]: """ integer value used to compare to the given metric. """ return pulumi.get(self, "value") @value.setter - def value(self, value: Optional[pulumi.Input[int]]): + def value(self, value: pulumi.Input[int]): pulumi.set(self, "value", value) @pulumi.input_type class JobJobClusterArgs: def __init__(__self__, *, - job_cluster_key: Optional[pulumi.Input[str]] = None, - new_cluster: Optional[pulumi.Input['JobJobClusterNewClusterArgs']] = None): + job_cluster_key: pulumi.Input[str], + new_cluster: pulumi.Input['JobJobClusterNewClusterArgs']): """ :param pulumi.Input[str] job_cluster_key: Identifier that can be referenced in `task` block, so that cluster is shared between tasks :param pulumi.Input['JobJobClusterNewClusterArgs'] new_cluster: Same set of parameters as for Cluster resource. """ - if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) - if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + pulumi.set(__self__, "job_cluster_key", job_cluster_key) + pulumi.set(__self__, "new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") - def job_cluster_key(self) -> Optional[pulumi.Input[str]]: + def job_cluster_key(self) -> pulumi.Input[str]: """ Identifier that can be referenced in `task` block, so that cluster is shared between tasks """ return pulumi.get(self, "job_cluster_key") @job_cluster_key.setter - def job_cluster_key(self, value: Optional[pulumi.Input[str]]): + def job_cluster_key(self, value: pulumi.Input[str]): pulumi.set(self, "job_cluster_key", value) @property @pulumi.getter(name="newCluster") - def new_cluster(self) -> Optional[pulumi.Input['JobJobClusterNewClusterArgs']]: + def new_cluster(self) -> pulumi.Input['JobJobClusterNewClusterArgs']: """ Same set of parameters as for Cluster resource. """ return pulumi.get(self, "new_cluster") @new_cluster.setter - def new_cluster(self, value: Optional[pulumi.Input['JobJobClusterNewClusterArgs']]): + def new_cluster(self, value: pulumi.Input['JobJobClusterNewClusterArgs']): pulumi.set(self, "new_cluster", value) @@ -3745,7 +3740,6 @@ def __init__(__self__, *, spark_version: pulumi.Input[str], apply_policy_default_values: Optional[pulumi.Input[bool]] = None, autoscale: Optional[pulumi.Input['JobJobClusterNewClusterAutoscaleArgs']] = None, - autotermination_minutes: Optional[pulumi.Input[int]] = None, aws_attributes: Optional[pulumi.Input['JobJobClusterNewClusterAwsAttributesArgs']] = None, azure_attributes: Optional[pulumi.Input['JobJobClusterNewClusterAzureAttributesArgs']] = None, cluster_id: Optional[pulumi.Input[str]] = None, @@ -3781,8 +3775,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -3865,15 +3857,6 @@ def autoscale(self) -> Optional[pulumi.Input['JobJobClusterNewClusterAutoscaleAr def autoscale(self, value: Optional[pulumi.Input['JobJobClusterNewClusterAutoscaleArgs']]): pulumi.set(self, "autoscale", value) - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[pulumi.Input[int]]: - return pulumi.get(self, "autotermination_minutes") - - @autotermination_minutes.setter - def autotermination_minutes(self, value: Optional[pulumi.Input[int]]): - pulumi.set(self, "autotermination_minutes", value) - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional[pulumi.Input['JobJobClusterNewClusterAwsAttributesArgs']]: @@ -5395,7 +5378,6 @@ def __init__(__self__, *, spark_version: pulumi.Input[str], apply_policy_default_values: Optional[pulumi.Input[bool]] = None, autoscale: Optional[pulumi.Input['JobNewClusterAutoscaleArgs']] = None, - autotermination_minutes: Optional[pulumi.Input[int]] = None, aws_attributes: Optional[pulumi.Input['JobNewClusterAwsAttributesArgs']] = None, azure_attributes: Optional[pulumi.Input['JobNewClusterAzureAttributesArgs']] = None, cluster_id: Optional[pulumi.Input[str]] = None, @@ -5431,8 +5413,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -5515,15 +5495,6 @@ def autoscale(self) -> Optional[pulumi.Input['JobNewClusterAutoscaleArgs']]: def autoscale(self, value: Optional[pulumi.Input['JobNewClusterAutoscaleArgs']]): pulumi.set(self, "autoscale", value) - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[pulumi.Input[int]]: - return pulumi.get(self, "autotermination_minutes") - - @autotermination_minutes.setter - def autotermination_minutes(self, value: Optional[pulumi.Input[int]]): - pulumi.set(self, "autotermination_minutes", value) - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional[pulumi.Input['JobNewClusterAwsAttributesArgs']]: @@ -7426,6 +7397,7 @@ def parameters(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]) @pulumi.input_type class JobTaskArgs: def __init__(__self__, *, + task_key: pulumi.Input[str], condition_task: Optional[pulumi.Input['JobTaskConditionTaskArgs']] = None, dbt_task: Optional[pulumi.Input['JobTaskDbtTaskArgs']] = None, depends_ons: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskDependsOnArgs']]]] = None, @@ -7452,10 +7424,11 @@ def __init__(__self__, *, spark_python_task: Optional[pulumi.Input['JobTaskSparkPythonTaskArgs']] = None, spark_submit_task: Optional[pulumi.Input['JobTaskSparkSubmitTaskArgs']] = None, sql_task: Optional[pulumi.Input['JobTaskSqlTaskArgs']] = None, - task_key: Optional[pulumi.Input[str]] = None, timeout_seconds: Optional[pulumi.Input[int]] = None, webhook_notifications: Optional[pulumi.Input['JobTaskWebhookNotificationsArgs']] = None): """ + :param pulumi.Input[str] task_key: string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: :param pulumi.Input[Sequence[pulumi.Input['JobTaskDependsOnArgs']]] depends_ons: block specifying dependency(-ies) for a given task. :param pulumi.Input[str] description: An optional description for the job. The maximum length is 1024 characters in UTF-8 encoding. :param pulumi.Input['JobTaskEmailNotificationsArgs'] email_notifications: (List) An optional set of email addresses notified when this task begins, completes or fails. The default behavior is to not send any emails. This field is a block and is documented below. @@ -7471,11 +7444,10 @@ def __init__(__self__, *, :param pulumi.Input['JobTaskNotificationSettingsArgs'] notification_settings: An optional block controlling the notification settings on the job level documented below. :param pulumi.Input[bool] retry_on_timeout: (Bool) An optional policy to specify whether to retry a job when it times out. The default behavior is to not retry on timeout. :param pulumi.Input[str] run_if: An optional value indicating the condition that determines whether the task should be run once its dependencies have been completed. One of `ALL_SUCCESS`, `AT_LEAST_ONE_SUCCESS`, `NONE_FAILED`, `ALL_DONE`, `AT_LEAST_ONE_FAILED` or `ALL_FAILED`. When omitted, defaults to `ALL_SUCCESS`. - :param pulumi.Input[str] task_key: string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: :param pulumi.Input[int] timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. :param pulumi.Input['JobTaskWebhookNotificationsArgs'] webhook_notifications: (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. """ + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -7528,13 +7500,24 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: pulumi.set(__self__, "webhook_notifications", webhook_notifications) + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> pulumi.Input[str]: + """ + string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: + """ + return pulumi.get(self, "task_key") + + @task_key.setter + def task_key(self, value: pulumi.Input[str]): + pulumi.set(self, "task_key", value) + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional[pulumi.Input['JobTaskConditionTaskArgs']]: @@ -7810,19 +7793,6 @@ def sql_task(self) -> Optional[pulumi.Input['JobTaskSqlTaskArgs']]: def sql_task(self, value: Optional[pulumi.Input['JobTaskSqlTaskArgs']]): pulumi.set(self, "sql_task", value) - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[pulumi.Input[str]]: - """ - string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: - """ - return pulumi.get(self, "task_key") - - @task_key.setter - def task_key(self, value: Optional[pulumi.Input[str]]): - pulumi.set(self, "task_key", value) - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[pulumi.Input[int]]: @@ -8211,6 +8181,7 @@ def concurrency(self, value: Optional[pulumi.Input[int]]): @pulumi.input_type class JobTaskForEachTaskTaskArgs: def __init__(__self__, *, + task_key: pulumi.Input[str], condition_task: Optional[pulumi.Input['JobTaskForEachTaskTaskConditionTaskArgs']] = None, dbt_task: Optional[pulumi.Input['JobTaskForEachTaskTaskDbtTaskArgs']] = None, depends_ons: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskForEachTaskTaskDependsOnArgs']]]] = None, @@ -8236,10 +8207,11 @@ def __init__(__self__, *, spark_python_task: Optional[pulumi.Input['JobTaskForEachTaskTaskSparkPythonTaskArgs']] = None, spark_submit_task: Optional[pulumi.Input['JobTaskForEachTaskTaskSparkSubmitTaskArgs']] = None, sql_task: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskArgs']] = None, - task_key: Optional[pulumi.Input[str]] = None, timeout_seconds: Optional[pulumi.Input[int]] = None, webhook_notifications: Optional[pulumi.Input['JobTaskForEachTaskTaskWebhookNotificationsArgs']] = None): """ + :param pulumi.Input[str] task_key: string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: :param pulumi.Input[Sequence[pulumi.Input['JobTaskForEachTaskTaskDependsOnArgs']]] depends_ons: block specifying dependency(-ies) for a given task. :param pulumi.Input[str] description: An optional description for the job. The maximum length is 1024 characters in UTF-8 encoding. :param pulumi.Input['JobTaskForEachTaskTaskEmailNotificationsArgs'] email_notifications: (List) An optional set of email addresses notified when this task begins, completes or fails. The default behavior is to not send any emails. This field is a block and is documented below. @@ -8255,11 +8227,10 @@ def __init__(__self__, *, :param pulumi.Input['JobTaskForEachTaskTaskNotificationSettingsArgs'] notification_settings: An optional block controlling the notification settings on the job level documented below. :param pulumi.Input[bool] retry_on_timeout: (Bool) An optional policy to specify whether to retry a job when it times out. The default behavior is to not retry on timeout. :param pulumi.Input[str] run_if: An optional value indicating the condition that determines whether the task should be run once its dependencies have been completed. One of `ALL_SUCCESS`, `AT_LEAST_ONE_SUCCESS`, `NONE_FAILED`, `ALL_DONE`, `AT_LEAST_ONE_FAILED` or `ALL_FAILED`. When omitted, defaults to `ALL_SUCCESS`. - :param pulumi.Input[str] task_key: string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: :param pulumi.Input[int] timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. :param pulumi.Input['JobTaskForEachTaskTaskWebhookNotificationsArgs'] webhook_notifications: (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. """ + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -8310,13 +8281,24 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: pulumi.set(__self__, "webhook_notifications", webhook_notifications) + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> pulumi.Input[str]: + """ + string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: + """ + return pulumi.get(self, "task_key") + + @task_key.setter + def task_key(self, value: pulumi.Input[str]): + pulumi.set(self, "task_key", value) + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional[pulumi.Input['JobTaskForEachTaskTaskConditionTaskArgs']]: @@ -8583,19 +8565,6 @@ def sql_task(self) -> Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskArgs'] def sql_task(self, value: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskArgs']]): pulumi.set(self, "sql_task", value) - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[pulumi.Input[str]]: - """ - string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: - """ - return pulumi.get(self, "task_key") - - @task_key.setter - def task_key(self, value: Optional[pulumi.Input[str]]): - pulumi.set(self, "task_key", value) - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[pulumi.Input[int]]: @@ -8953,55 +8922,52 @@ def rules(self, value: pulumi.Input[Sequence[pulumi.Input['JobTaskForEachTaskTas @pulumi.input_type class JobTaskForEachTaskTaskHealthRuleArgs: def __init__(__self__, *, - metric: Optional[pulumi.Input[str]] = None, - op: Optional[pulumi.Input[str]] = None, - value: Optional[pulumi.Input[int]] = None): + metric: pulumi.Input[str], + op: pulumi.Input[str], + value: pulumi.Input[int]): """ :param pulumi.Input[str] metric: string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). :param pulumi.Input[str] op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param pulumi.Input[int] value: integer value used to compare to the given metric. """ - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[pulumi.Input[str]]: + def metric(self) -> pulumi.Input[str]: """ string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). """ return pulumi.get(self, "metric") @metric.setter - def metric(self, value: Optional[pulumi.Input[str]]): + def metric(self, value: pulumi.Input[str]): pulumi.set(self, "metric", value) @property @pulumi.getter - def op(self) -> Optional[pulumi.Input[str]]: + def op(self) -> pulumi.Input[str]: """ string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ return pulumi.get(self, "op") @op.setter - def op(self, value: Optional[pulumi.Input[str]]): + def op(self, value: pulumi.Input[str]): pulumi.set(self, "op", value) @property @pulumi.getter - def value(self) -> Optional[pulumi.Input[int]]: + def value(self) -> pulumi.Input[int]: """ integer value used to compare to the given metric. """ return pulumi.get(self, "value") @value.setter - def value(self, value: Optional[pulumi.Input[int]]): + def value(self, value: pulumi.Input[int]): pulumi.set(self, "value", value) @@ -9196,7 +9162,6 @@ def __init__(__self__, *, spark_version: pulumi.Input[str], apply_policy_default_values: Optional[pulumi.Input[bool]] = None, autoscale: Optional[pulumi.Input['JobTaskForEachTaskTaskNewClusterAutoscaleArgs']] = None, - autotermination_minutes: Optional[pulumi.Input[int]] = None, aws_attributes: Optional[pulumi.Input['JobTaskForEachTaskTaskNewClusterAwsAttributesArgs']] = None, azure_attributes: Optional[pulumi.Input['JobTaskForEachTaskTaskNewClusterAzureAttributesArgs']] = None, cluster_id: Optional[pulumi.Input[str]] = None, @@ -9232,8 +9197,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -9316,15 +9279,6 @@ def autoscale(self) -> Optional[pulumi.Input['JobTaskForEachTaskTaskNewClusterAu def autoscale(self, value: Optional[pulumi.Input['JobTaskForEachTaskTaskNewClusterAutoscaleArgs']]): pulumi.set(self, "autoscale", value) - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[pulumi.Input[int]]: - return pulumi.get(self, "autotermination_minutes") - - @autotermination_minutes.setter - def autotermination_minutes(self, value: Optional[pulumi.Input[int]]): - pulumi.set(self, "autotermination_minutes", value) - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional[pulumi.Input['JobTaskForEachTaskTaskNewClusterAwsAttributesArgs']]: @@ -11179,20 +11133,21 @@ def parameters(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]) @pulumi.input_type class JobTaskForEachTaskTaskSqlTaskArgs: def __init__(__self__, *, + warehouse_id: pulumi.Input[str], alert: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskAlertArgs']] = None, dashboard: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskDashboardArgs']] = None, file: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskFileArgs']] = None, parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, - query: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskQueryArgs']] = None, - warehouse_id: Optional[pulumi.Input[str]] = None): + query: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskQueryArgs']] = None): """ + :param pulumi.Input[str] warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. :param pulumi.Input['JobTaskForEachTaskTaskSqlTaskAlertArgs'] alert: block consisting of following fields: :param pulumi.Input['JobTaskForEachTaskTaskSqlTaskDashboardArgs'] dashboard: block consisting of following fields: :param pulumi.Input['JobTaskForEachTaskTaskSqlTaskFileArgs'] file: block consisting of single string fields: :param pulumi.Input[Mapping[str, Any]] parameters: (Map) parameters to be used for each run of this task. The SQL alert task does not support custom parameters. :param pulumi.Input['JobTaskForEachTaskTaskSqlTaskQueryArgs'] query: block consisting of single string field: `query_id` - identifier of the Databricks SQL Query (databricks_sql_query). - :param pulumi.Input[str] warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. """ + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -11203,8 +11158,18 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> pulumi.Input[str]: + """ + ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. + """ + return pulumi.get(self, "warehouse_id") + + @warehouse_id.setter + def warehouse_id(self, value: pulumi.Input[str]): + pulumi.set(self, "warehouse_id", value) @property @pulumi.getter @@ -11266,18 +11231,6 @@ def query(self) -> Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskQueryArgs def query(self, value: Optional[pulumi.Input['JobTaskForEachTaskTaskSqlTaskQueryArgs']]): pulumi.set(self, "query", value) - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[pulumi.Input[str]]: - """ - ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - """ - return pulumi.get(self, "warehouse_id") - - @warehouse_id.setter - def warehouse_id(self, value: Optional[pulumi.Input[str]]): - pulumi.set(self, "warehouse_id", value) - @pulumi.input_type class JobTaskForEachTaskTaskSqlTaskAlertArgs: @@ -11796,55 +11749,52 @@ def rules(self, value: pulumi.Input[Sequence[pulumi.Input['JobTaskHealthRuleArgs @pulumi.input_type class JobTaskHealthRuleArgs: def __init__(__self__, *, - metric: Optional[pulumi.Input[str]] = None, - op: Optional[pulumi.Input[str]] = None, - value: Optional[pulumi.Input[int]] = None): + metric: pulumi.Input[str], + op: pulumi.Input[str], + value: pulumi.Input[int]): """ :param pulumi.Input[str] metric: string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). :param pulumi.Input[str] op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param pulumi.Input[int] value: integer value used to compare to the given metric. """ - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[pulumi.Input[str]]: + def metric(self) -> pulumi.Input[str]: """ string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). """ return pulumi.get(self, "metric") @metric.setter - def metric(self, value: Optional[pulumi.Input[str]]): + def metric(self, value: pulumi.Input[str]): pulumi.set(self, "metric", value) @property @pulumi.getter - def op(self) -> Optional[pulumi.Input[str]]: + def op(self) -> pulumi.Input[str]: """ string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ return pulumi.get(self, "op") @op.setter - def op(self, value: Optional[pulumi.Input[str]]): + def op(self, value: pulumi.Input[str]): pulumi.set(self, "op", value) @property @pulumi.getter - def value(self) -> Optional[pulumi.Input[int]]: + def value(self) -> pulumi.Input[int]: """ integer value used to compare to the given metric. """ return pulumi.get(self, "value") @value.setter - def value(self, value: Optional[pulumi.Input[int]]): + def value(self, value: pulumi.Input[int]): pulumi.set(self, "value", value) @@ -12039,7 +11989,6 @@ def __init__(__self__, *, spark_version: pulumi.Input[str], apply_policy_default_values: Optional[pulumi.Input[bool]] = None, autoscale: Optional[pulumi.Input['JobTaskNewClusterAutoscaleArgs']] = None, - autotermination_minutes: Optional[pulumi.Input[int]] = None, aws_attributes: Optional[pulumi.Input['JobTaskNewClusterAwsAttributesArgs']] = None, azure_attributes: Optional[pulumi.Input['JobTaskNewClusterAzureAttributesArgs']] = None, cluster_id: Optional[pulumi.Input[str]] = None, @@ -12075,8 +12024,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -12159,15 +12106,6 @@ def autoscale(self) -> Optional[pulumi.Input['JobTaskNewClusterAutoscaleArgs']]: def autoscale(self, value: Optional[pulumi.Input['JobTaskNewClusterAutoscaleArgs']]): pulumi.set(self, "autoscale", value) - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[pulumi.Input[int]]: - return pulumi.get(self, "autotermination_minutes") - - @autotermination_minutes.setter - def autotermination_minutes(self, value: Optional[pulumi.Input[int]]): - pulumi.set(self, "autotermination_minutes", value) - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional[pulumi.Input['JobTaskNewClusterAwsAttributesArgs']]: @@ -14022,20 +13960,21 @@ def parameters(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]) @pulumi.input_type class JobTaskSqlTaskArgs: def __init__(__self__, *, + warehouse_id: pulumi.Input[str], alert: Optional[pulumi.Input['JobTaskSqlTaskAlertArgs']] = None, dashboard: Optional[pulumi.Input['JobTaskSqlTaskDashboardArgs']] = None, file: Optional[pulumi.Input['JobTaskSqlTaskFileArgs']] = None, parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, - query: Optional[pulumi.Input['JobTaskSqlTaskQueryArgs']] = None, - warehouse_id: Optional[pulumi.Input[str]] = None): + query: Optional[pulumi.Input['JobTaskSqlTaskQueryArgs']] = None): """ + :param pulumi.Input[str] warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. :param pulumi.Input['JobTaskSqlTaskAlertArgs'] alert: block consisting of following fields: :param pulumi.Input['JobTaskSqlTaskDashboardArgs'] dashboard: block consisting of following fields: :param pulumi.Input['JobTaskSqlTaskFileArgs'] file: block consisting of single string fields: :param pulumi.Input[Mapping[str, Any]] parameters: (Map) parameters to be used for each run of this task. The SQL alert task does not support custom parameters. :param pulumi.Input['JobTaskSqlTaskQueryArgs'] query: block consisting of single string field: `query_id` - identifier of the Databricks SQL Query (databricks_sql_query). - :param pulumi.Input[str] warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. """ + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -14046,8 +13985,18 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> pulumi.Input[str]: + """ + ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. + """ + return pulumi.get(self, "warehouse_id") + + @warehouse_id.setter + def warehouse_id(self, value: pulumi.Input[str]): + pulumi.set(self, "warehouse_id", value) @property @pulumi.getter @@ -14109,18 +14058,6 @@ def query(self) -> Optional[pulumi.Input['JobTaskSqlTaskQueryArgs']]: def query(self, value: Optional[pulumi.Input['JobTaskSqlTaskQueryArgs']]): pulumi.set(self, "query", value) - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[pulumi.Input[str]]: - """ - ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - """ - return pulumi.get(self, "warehouse_id") - - @warehouse_id.setter - def warehouse_id(self, value: Optional[pulumi.Input[str]]): - pulumi.set(self, "warehouse_id", value) - @pulumi.input_type class JobTaskSqlTaskAlertArgs: @@ -14624,7 +14561,6 @@ def __init__(__self__, *, """ :param pulumi.Input['JobTriggerFileArrivalArgs'] file_arrival: configuration block to define a trigger for [File Arrival events](https://learn.microsoft.com/en-us/azure/databricks/workflows/jobs/file-arrival-triggers) consisting of following attributes: :param pulumi.Input[str] pause_status: Indicate whether this trigger is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pause_status`. - :param pulumi.Input['JobTriggerTableUpdateArgs'] table_update: configuration block to define a trigger for Table Update events consisting of following attributes: """ if file_arrival is not None: pulumi.set(__self__, "file_arrival", file_arrival) @@ -14671,9 +14607,6 @@ def table(self, value: Optional[pulumi.Input['JobTriggerTableArgs']]): @property @pulumi.getter(name="tableUpdate") def table_update(self) -> Optional[pulumi.Input['JobTriggerTableUpdateArgs']]: - """ - configuration block to define a trigger for Table Update events consisting of following attributes: - """ return pulumi.get(self, "table_update") @table_update.setter @@ -14795,12 +14728,6 @@ def __init__(__self__, *, condition: Optional[pulumi.Input[str]] = None, min_time_between_triggers_seconds: Optional[pulumi.Input[int]] = None, wait_after_last_change_seconds: Optional[pulumi.Input[int]] = None): - """ - :param pulumi.Input[Sequence[pulumi.Input[str]]] table_names: A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - :param pulumi.Input[str] condition: The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - :param pulumi.Input[int] min_time_between_triggers_seconds: If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - :param pulumi.Input[int] wait_after_last_change_seconds: If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - """ pulumi.set(__self__, "table_names", table_names) if condition is not None: pulumi.set(__self__, "condition", condition) @@ -14812,9 +14739,6 @@ def __init__(__self__, *, @property @pulumi.getter(name="tableNames") def table_names(self) -> pulumi.Input[Sequence[pulumi.Input[str]]]: - """ - A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - """ return pulumi.get(self, "table_names") @table_names.setter @@ -14824,9 +14748,6 @@ def table_names(self, value: pulumi.Input[Sequence[pulumi.Input[str]]]): @property @pulumi.getter def condition(self) -> Optional[pulumi.Input[str]]: - """ - The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - """ return pulumi.get(self, "condition") @condition.setter @@ -14836,9 +14757,6 @@ def condition(self, value: Optional[pulumi.Input[str]]): @property @pulumi.getter(name="minTimeBetweenTriggersSeconds") def min_time_between_triggers_seconds(self) -> Optional[pulumi.Input[int]]: - """ - If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - """ return pulumi.get(self, "min_time_between_triggers_seconds") @min_time_between_triggers_seconds.setter @@ -14848,9 +14766,6 @@ def min_time_between_triggers_seconds(self, value: Optional[pulumi.Input[int]]): @property @pulumi.getter(name="waitAfterLastChangeSeconds") def wait_after_last_change_seconds(self) -> Optional[pulumi.Input[int]]: - """ - If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - """ return pulumi.get(self, "wait_after_last_change_seconds") @wait_after_last_change_seconds.setter @@ -26846,70 +26761,65 @@ def rules(self, value: Sequence['GetJobJobSettingsSettingsHealthRuleArgs']): @pulumi.input_type class GetJobJobSettingsSettingsHealthRuleArgs: def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + metric: str, + op: str, + value: int): + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: return pulumi.get(self, "metric") @metric.setter - def metric(self, value: Optional[str]): + def metric(self, value: str): pulumi.set(self, "metric", value) @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: return pulumi.get(self, "op") @op.setter - def op(self, value: Optional[str]): + def op(self, value: str): pulumi.set(self, "op", value) @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: return pulumi.get(self, "value") @value.setter - def value(self, value: Optional[int]): + def value(self, value: int): pulumi.set(self, "value", value) @pulumi.input_type class GetJobJobSettingsSettingsJobClusterArgs: def __init__(__self__, *, - job_cluster_key: Optional[str] = None, - new_cluster: Optional['GetJobJobSettingsSettingsJobClusterNewClusterArgs'] = None): - if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) - if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + job_cluster_key: str, + new_cluster: 'GetJobJobSettingsSettingsJobClusterNewClusterArgs'): + pulumi.set(__self__, "job_cluster_key", job_cluster_key) + pulumi.set(__self__, "new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") - def job_cluster_key(self) -> Optional[str]: + def job_cluster_key(self) -> str: return pulumi.get(self, "job_cluster_key") @job_cluster_key.setter - def job_cluster_key(self, value: Optional[str]): + def job_cluster_key(self, value: str): pulumi.set(self, "job_cluster_key", value) @property @pulumi.getter(name="newCluster") - def new_cluster(self) -> Optional['GetJobJobSettingsSettingsJobClusterNewClusterArgs']: + def new_cluster(self) -> 'GetJobJobSettingsSettingsJobClusterNewClusterArgs': return pulumi.get(self, "new_cluster") @new_cluster.setter - def new_cluster(self, value: Optional['GetJobJobSettingsSettingsJobClusterNewClusterArgs']): + def new_cluster(self, value: 'GetJobJobSettingsSettingsJobClusterNewClusterArgs'): pulumi.set(self, "new_cluster", value) @@ -29865,6 +29775,7 @@ def parameters(self, value: Optional[Sequence[str]]): class GetJobJobSettingsSettingsTaskArgs: def __init__(__self__, *, retry_on_timeout: bool, + task_key: str, condition_task: Optional['GetJobJobSettingsSettingsTaskConditionTaskArgs'] = None, dbt_task: Optional['GetJobJobSettingsSettingsTaskDbtTaskArgs'] = None, depends_ons: Optional[Sequence['GetJobJobSettingsSettingsTaskDependsOnArgs']] = None, @@ -29889,10 +29800,10 @@ def __init__(__self__, *, spark_python_task: Optional['GetJobJobSettingsSettingsTaskSparkPythonTaskArgs'] = None, spark_submit_task: Optional['GetJobJobSettingsSettingsTaskSparkSubmitTaskArgs'] = None, sql_task: Optional['GetJobJobSettingsSettingsTaskSqlTaskArgs'] = None, - task_key: Optional[str] = None, timeout_seconds: Optional[int] = None, webhook_notifications: Optional['GetJobJobSettingsSettingsTaskWebhookNotificationsArgs'] = None): pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -29941,8 +29852,6 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: @@ -29957,6 +29866,15 @@ def retry_on_timeout(self) -> bool: def retry_on_timeout(self, value: bool): pulumi.set(self, "retry_on_timeout", value) + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> str: + return pulumi.get(self, "task_key") + + @task_key.setter + def task_key(self, value: str): + pulumi.set(self, "task_key", value) + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional['GetJobJobSettingsSettingsTaskConditionTaskArgs']: @@ -30173,15 +30091,6 @@ def sql_task(self) -> Optional['GetJobJobSettingsSettingsTaskSqlTaskArgs']: def sql_task(self, value: Optional['GetJobJobSettingsSettingsTaskSqlTaskArgs']): pulumi.set(self, "sql_task", value) - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[str]: - return pulumi.get(self, "task_key") - - @task_key.setter - def task_key(self, value: Optional[str]): - pulumi.set(self, "task_key", value) - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[int]: @@ -30463,6 +30372,7 @@ def concurrency(self, value: Optional[int]): class GetJobJobSettingsSettingsTaskForEachTaskTaskArgs: def __init__(__self__, *, retry_on_timeout: bool, + task_key: str, condition_task: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTaskArgs'] = None, dbt_task: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTaskArgs'] = None, depends_ons: Optional[Sequence['GetJobJobSettingsSettingsTaskForEachTaskTaskDependsOnArgs']] = None, @@ -30486,10 +30396,10 @@ def __init__(__self__, *, spark_python_task: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTaskArgs'] = None, spark_submit_task: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTaskArgs'] = None, sql_task: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs'] = None, - task_key: Optional[str] = None, timeout_seconds: Optional[int] = None, webhook_notifications: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotificationsArgs'] = None): pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -30536,8 +30446,6 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: @@ -30552,6 +30460,15 @@ def retry_on_timeout(self) -> bool: def retry_on_timeout(self, value: bool): pulumi.set(self, "retry_on_timeout", value) + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> str: + return pulumi.get(self, "task_key") + + @task_key.setter + def task_key(self, value: str): + pulumi.set(self, "task_key", value) + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTaskArgs']: @@ -30759,15 +30676,6 @@ def sql_task(self) -> Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlT def sql_task(self, value: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs']): pulumi.set(self, "sql_task", value) - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[str]: - return pulumi.get(self, "task_key") - - @task_key.setter - def task_key(self, value: Optional[str]): - pulumi.set(self, "task_key", value) - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[int]: @@ -31025,41 +30933,38 @@ def rules(self, value: Sequence['GetJobJobSettingsSettingsTaskForEachTaskTaskHea @pulumi.input_type class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleArgs: def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + metric: str, + op: str, + value: int): + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: return pulumi.get(self, "metric") @metric.setter - def metric(self, value: Optional[str]): + def metric(self, value: str): pulumi.set(self, "metric", value) @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: return pulumi.get(self, "op") @op.setter - def op(self, value: Optional[str]): + def op(self, value: str): pulumi.set(self, "op", value) @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: return pulumi.get(self, "value") @value.setter - def value(self, value: Optional[int]): + def value(self, value: int): pulumi.set(self, "value", value) @@ -32730,12 +32635,13 @@ def parameters(self, value: Optional[Sequence[str]]): @pulumi.input_type class GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskArgs: def __init__(__self__, *, + warehouse_id: str, alert: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlertArgs'] = None, dashboard: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskDashboardArgs'] = None, file: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFileArgs'] = None, parameters: Optional[Mapping[str, Any]] = None, - query: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryArgs'] = None, - warehouse_id: Optional[str] = None): + query: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryArgs'] = None): + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -32746,8 +32652,15 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> str: + return pulumi.get(self, "warehouse_id") + + @warehouse_id.setter + def warehouse_id(self, value: str): + pulumi.set(self, "warehouse_id", value) @property @pulumi.getter @@ -32794,15 +32707,6 @@ def query(self) -> Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTask def query(self, value: Optional['GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryArgs']): pulumi.set(self, "query", value) - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[str]: - return pulumi.get(self, "warehouse_id") - - @warehouse_id.setter - def warehouse_id(self, value: Optional[str]): - pulumi.set(self, "warehouse_id", value) - @pulumi.input_type class GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlertArgs: @@ -33157,41 +33061,38 @@ def rules(self, value: Sequence['GetJobJobSettingsSettingsTaskHealthRuleArgs']): @pulumi.input_type class GetJobJobSettingsSettingsTaskHealthRuleArgs: def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + metric: str, + op: str, + value: int): + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: return pulumi.get(self, "metric") @metric.setter - def metric(self, value: Optional[str]): + def metric(self, value: str): pulumi.set(self, "metric", value) @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: return pulumi.get(self, "op") @op.setter - def op(self, value: Optional[str]): + def op(self, value: str): pulumi.set(self, "op", value) @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: return pulumi.get(self, "value") @value.setter - def value(self, value: Optional[int]): + def value(self, value: int): pulumi.set(self, "value", value) @@ -34862,12 +34763,13 @@ def parameters(self, value: Optional[Sequence[str]]): @pulumi.input_type class GetJobJobSettingsSettingsTaskSqlTaskArgs: def __init__(__self__, *, + warehouse_id: str, alert: Optional['GetJobJobSettingsSettingsTaskSqlTaskAlertArgs'] = None, dashboard: Optional['GetJobJobSettingsSettingsTaskSqlTaskDashboardArgs'] = None, file: Optional['GetJobJobSettingsSettingsTaskSqlTaskFileArgs'] = None, parameters: Optional[Mapping[str, Any]] = None, - query: Optional['GetJobJobSettingsSettingsTaskSqlTaskQueryArgs'] = None, - warehouse_id: Optional[str] = None): + query: Optional['GetJobJobSettingsSettingsTaskSqlTaskQueryArgs'] = None): + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -34878,8 +34780,15 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> str: + return pulumi.get(self, "warehouse_id") + + @warehouse_id.setter + def warehouse_id(self, value: str): + pulumi.set(self, "warehouse_id", value) @property @pulumi.getter @@ -34926,15 +34835,6 @@ def query(self) -> Optional['GetJobJobSettingsSettingsTaskSqlTaskQueryArgs']: def query(self, value: Optional['GetJobJobSettingsSettingsTaskSqlTaskQueryArgs']): pulumi.set(self, "query", value) - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[str]: - return pulumi.get(self, "warehouse_id") - - @warehouse_id.setter - def warehouse_id(self, value: Optional[str]): - pulumi.set(self, "warehouse_id", value) - @pulumi.input_type class GetJobJobSettingsSettingsTaskSqlTaskAlertArgs: diff --git a/sdk/python/pulumi_databricks/outputs.py b/sdk/python/pulumi_databricks/outputs.py index 5feef233..fb834e97 100644 --- a/sdk/python/pulumi_databricks/outputs.py +++ b/sdk/python/pulumi_databricks/outputs.py @@ -3669,24 +3669,21 @@ def rules(self) -> Sequence['outputs.JobHealthRule']: @pulumi.output_type class JobHealthRule(dict): def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): + metric: str, + op: str, + value: int): """ :param str metric: string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). :param str op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param int value: integer value used to compare to the given metric. """ - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: """ string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). """ @@ -3694,7 +3691,7 @@ def metric(self) -> Optional[str]: @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: """ string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ @@ -3702,7 +3699,7 @@ def op(self) -> Optional[str]: @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: """ integer value used to compare to the given metric. """ @@ -3731,20 +3728,18 @@ def get(self, key: str, default = None) -> Any: return super().get(key, default) def __init__(__self__, *, - job_cluster_key: Optional[str] = None, - new_cluster: Optional['outputs.JobJobClusterNewCluster'] = None): + job_cluster_key: str, + new_cluster: 'outputs.JobJobClusterNewCluster'): """ :param str job_cluster_key: Identifier that can be referenced in `task` block, so that cluster is shared between tasks :param 'JobJobClusterNewClusterArgs' new_cluster: Same set of parameters as for Cluster resource. """ - if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) - if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + pulumi.set(__self__, "job_cluster_key", job_cluster_key) + pulumi.set(__self__, "new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") - def job_cluster_key(self) -> Optional[str]: + def job_cluster_key(self) -> str: """ Identifier that can be referenced in `task` block, so that cluster is shared between tasks """ @@ -3752,7 +3747,7 @@ def job_cluster_key(self) -> Optional[str]: @property @pulumi.getter(name="newCluster") - def new_cluster(self) -> Optional['outputs.JobJobClusterNewCluster']: + def new_cluster(self) -> 'outputs.JobJobClusterNewCluster': """ Same set of parameters as for Cluster resource. """ @@ -3768,8 +3763,6 @@ def __key_warning(key: str): suggest = "spark_version" elif key == "applyPolicyDefaultValues": suggest = "apply_policy_default_values" - elif key == "autoterminationMinutes": - suggest = "autotermination_minutes" elif key == "awsAttributes": suggest = "aws_attributes" elif key == "azureAttributes": @@ -3838,7 +3831,6 @@ def __init__(__self__, *, spark_version: str, apply_policy_default_values: Optional[bool] = None, autoscale: Optional['outputs.JobJobClusterNewClusterAutoscale'] = None, - autotermination_minutes: Optional[int] = None, aws_attributes: Optional['outputs.JobJobClusterNewClusterAwsAttributes'] = None, azure_attributes: Optional['outputs.JobJobClusterNewClusterAzureAttributes'] = None, cluster_id: Optional[str] = None, @@ -3874,8 +3866,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -3946,11 +3936,6 @@ def apply_policy_default_values(self) -> Optional[bool]: def autoscale(self) -> Optional['outputs.JobJobClusterNewClusterAutoscale']: return pulumi.get(self, "autoscale") - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[int]: - return pulumi.get(self, "autotermination_minutes") - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional['outputs.JobJobClusterNewClusterAwsAttributes']: @@ -5208,8 +5193,6 @@ def __key_warning(key: str): suggest = "spark_version" elif key == "applyPolicyDefaultValues": suggest = "apply_policy_default_values" - elif key == "autoterminationMinutes": - suggest = "autotermination_minutes" elif key == "awsAttributes": suggest = "aws_attributes" elif key == "azureAttributes": @@ -5278,7 +5261,6 @@ def __init__(__self__, *, spark_version: str, apply_policy_default_values: Optional[bool] = None, autoscale: Optional['outputs.JobNewClusterAutoscale'] = None, - autotermination_minutes: Optional[int] = None, aws_attributes: Optional['outputs.JobNewClusterAwsAttributes'] = None, azure_attributes: Optional['outputs.JobNewClusterAzureAttributes'] = None, cluster_id: Optional[str] = None, @@ -5314,8 +5296,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -5386,11 +5366,6 @@ def apply_policy_default_values(self) -> Optional[bool]: def autoscale(self) -> Optional['outputs.JobNewClusterAutoscale']: return pulumi.get(self, "autoscale") - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[int]: - return pulumi.get(self, "autotermination_minutes") - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional['outputs.JobNewClusterAwsAttributes']: @@ -7143,7 +7118,9 @@ class JobTask(dict): @staticmethod def __key_warning(key: str): suggest = None - if key == "conditionTask": + if key == "taskKey": + suggest = "task_key" + elif key == "conditionTask": suggest = "condition_task" elif key == "dbtTask": suggest = "dbt_task" @@ -7189,8 +7166,6 @@ def __key_warning(key: str): suggest = "spark_submit_task" elif key == "sqlTask": suggest = "sql_task" - elif key == "taskKey": - suggest = "task_key" elif key == "timeoutSeconds": suggest = "timeout_seconds" elif key == "webhookNotifications": @@ -7208,6 +7183,7 @@ def get(self, key: str, default = None) -> Any: return super().get(key, default) def __init__(__self__, *, + task_key: str, condition_task: Optional['outputs.JobTaskConditionTask'] = None, dbt_task: Optional['outputs.JobTaskDbtTask'] = None, depends_ons: Optional[Sequence['outputs.JobTaskDependsOn']] = None, @@ -7234,10 +7210,11 @@ def __init__(__self__, *, spark_python_task: Optional['outputs.JobTaskSparkPythonTask'] = None, spark_submit_task: Optional['outputs.JobTaskSparkSubmitTask'] = None, sql_task: Optional['outputs.JobTaskSqlTask'] = None, - task_key: Optional[str] = None, timeout_seconds: Optional[int] = None, webhook_notifications: Optional['outputs.JobTaskWebhookNotifications'] = None): """ + :param str task_key: string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: :param Sequence['JobTaskDependsOnArgs'] depends_ons: block specifying dependency(-ies) for a given task. :param str description: An optional description for the job. The maximum length is 1024 characters in UTF-8 encoding. :param 'JobTaskEmailNotificationsArgs' email_notifications: (List) An optional set of email addresses notified when this task begins, completes or fails. The default behavior is to not send any emails. This field is a block and is documented below. @@ -7253,11 +7230,10 @@ def __init__(__self__, *, :param 'JobTaskNotificationSettingsArgs' notification_settings: An optional block controlling the notification settings on the job level documented below. :param bool retry_on_timeout: (Bool) An optional policy to specify whether to retry a job when it times out. The default behavior is to not retry on timeout. :param str run_if: An optional value indicating the condition that determines whether the task should be run once its dependencies have been completed. One of `ALL_SUCCESS`, `AT_LEAST_ONE_SUCCESS`, `NONE_FAILED`, `ALL_DONE`, `AT_LEAST_ONE_FAILED` or `ALL_FAILED`. When omitted, defaults to `ALL_SUCCESS`. - :param str task_key: string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: :param int timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. :param 'JobTaskWebhookNotificationsArgs' webhook_notifications: (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. """ + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -7310,13 +7286,20 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: pulumi.set(__self__, "webhook_notifications", webhook_notifications) + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> str: + """ + string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: + """ + return pulumi.get(self, "task_key") + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional['outputs.JobTaskConditionTask']: @@ -7488,15 +7471,6 @@ def spark_submit_task(self) -> Optional['outputs.JobTaskSparkSubmitTask']: def sql_task(self) -> Optional['outputs.JobTaskSqlTask']: return pulumi.get(self, "sql_task") - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[str]: - """ - string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: - """ - return pulumi.get(self, "task_key") - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[int]: @@ -7862,7 +7836,9 @@ class JobTaskForEachTaskTask(dict): @staticmethod def __key_warning(key: str): suggest = None - if key == "conditionTask": + if key == "taskKey": + suggest = "task_key" + elif key == "conditionTask": suggest = "condition_task" elif key == "dbtTask": suggest = "dbt_task" @@ -7906,8 +7882,6 @@ def __key_warning(key: str): suggest = "spark_submit_task" elif key == "sqlTask": suggest = "sql_task" - elif key == "taskKey": - suggest = "task_key" elif key == "timeoutSeconds": suggest = "timeout_seconds" elif key == "webhookNotifications": @@ -7925,6 +7899,7 @@ def get(self, key: str, default = None) -> Any: return super().get(key, default) def __init__(__self__, *, + task_key: str, condition_task: Optional['outputs.JobTaskForEachTaskTaskConditionTask'] = None, dbt_task: Optional['outputs.JobTaskForEachTaskTaskDbtTask'] = None, depends_ons: Optional[Sequence['outputs.JobTaskForEachTaskTaskDependsOn']] = None, @@ -7950,10 +7925,11 @@ def __init__(__self__, *, spark_python_task: Optional['outputs.JobTaskForEachTaskTaskSparkPythonTask'] = None, spark_submit_task: Optional['outputs.JobTaskForEachTaskTaskSparkSubmitTask'] = None, sql_task: Optional['outputs.JobTaskForEachTaskTaskSqlTask'] = None, - task_key: Optional[str] = None, timeout_seconds: Optional[int] = None, webhook_notifications: Optional['outputs.JobTaskForEachTaskTaskWebhookNotifications'] = None): """ + :param str task_key: string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: :param Sequence['JobTaskForEachTaskTaskDependsOnArgs'] depends_ons: block specifying dependency(-ies) for a given task. :param str description: An optional description for the job. The maximum length is 1024 characters in UTF-8 encoding. :param 'JobTaskForEachTaskTaskEmailNotificationsArgs' email_notifications: (List) An optional set of email addresses notified when this task begins, completes or fails. The default behavior is to not send any emails. This field is a block and is documented below. @@ -7969,11 +7945,10 @@ def __init__(__self__, *, :param 'JobTaskForEachTaskTaskNotificationSettingsArgs' notification_settings: An optional block controlling the notification settings on the job level documented below. :param bool retry_on_timeout: (Bool) An optional policy to specify whether to retry a job when it times out. The default behavior is to not retry on timeout. :param str run_if: An optional value indicating the condition that determines whether the task should be run once its dependencies have been completed. One of `ALL_SUCCESS`, `AT_LEAST_ONE_SUCCESS`, `NONE_FAILED`, `ALL_DONE`, `AT_LEAST_ONE_FAILED` or `ALL_FAILED`. When omitted, defaults to `ALL_SUCCESS`. - :param str task_key: string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: :param int timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. :param 'JobTaskForEachTaskTaskWebhookNotificationsArgs' webhook_notifications: (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this task begins, completes or fails. The default behavior is to not send any notifications. This field is a block and is documented below. """ + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -8024,13 +7999,20 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: pulumi.set(__self__, "webhook_notifications", webhook_notifications) + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> str: + """ + string specifying an unique key for a given task. + * `*_task` - (Required) one of the specific task blocks described below: + """ + return pulumi.get(self, "task_key") + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional['outputs.JobTaskForEachTaskTaskConditionTask']: @@ -8197,15 +8179,6 @@ def spark_submit_task(self) -> Optional['outputs.JobTaskForEachTaskTaskSparkSubm def sql_task(self) -> Optional['outputs.JobTaskForEachTaskTaskSqlTask']: return pulumi.get(self, "sql_task") - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[str]: - """ - string specifying an unique key for a given task. - * `*_task` - (Required) one of the specific task blocks described below: - """ - return pulumi.get(self, "task_key") - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[int]: @@ -8546,24 +8519,21 @@ def rules(self) -> Sequence['outputs.JobTaskForEachTaskTaskHealthRule']: @pulumi.output_type class JobTaskForEachTaskTaskHealthRule(dict): def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): + metric: str, + op: str, + value: int): """ :param str metric: string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). :param str op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param int value: integer value used to compare to the given metric. """ - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: """ string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). """ @@ -8571,7 +8541,7 @@ def metric(self) -> Optional[str]: @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: """ string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ @@ -8579,7 +8549,7 @@ def op(self) -> Optional[str]: @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: """ integer value used to compare to the given metric. """ @@ -8724,8 +8694,6 @@ def __key_warning(key: str): suggest = "spark_version" elif key == "applyPolicyDefaultValues": suggest = "apply_policy_default_values" - elif key == "autoterminationMinutes": - suggest = "autotermination_minutes" elif key == "awsAttributes": suggest = "aws_attributes" elif key == "azureAttributes": @@ -8794,7 +8762,6 @@ def __init__(__self__, *, spark_version: str, apply_policy_default_values: Optional[bool] = None, autoscale: Optional['outputs.JobTaskForEachTaskTaskNewClusterAutoscale'] = None, - autotermination_minutes: Optional[int] = None, aws_attributes: Optional['outputs.JobTaskForEachTaskTaskNewClusterAwsAttributes'] = None, azure_attributes: Optional['outputs.JobTaskForEachTaskTaskNewClusterAzureAttributes'] = None, cluster_id: Optional[str] = None, @@ -8830,8 +8797,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -8902,11 +8867,6 @@ def apply_policy_default_values(self) -> Optional[bool]: def autoscale(self) -> Optional['outputs.JobTaskForEachTaskTaskNewClusterAutoscale']: return pulumi.get(self, "autoscale") - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[int]: - return pulumi.get(self, "autotermination_minutes") - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional['outputs.JobTaskForEachTaskTaskNewClusterAwsAttributes']: @@ -10613,20 +10573,21 @@ def get(self, key: str, default = None) -> Any: return super().get(key, default) def __init__(__self__, *, + warehouse_id: str, alert: Optional['outputs.JobTaskForEachTaskTaskSqlTaskAlert'] = None, dashboard: Optional['outputs.JobTaskForEachTaskTaskSqlTaskDashboard'] = None, file: Optional['outputs.JobTaskForEachTaskTaskSqlTaskFile'] = None, parameters: Optional[Mapping[str, Any]] = None, - query: Optional['outputs.JobTaskForEachTaskTaskSqlTaskQuery'] = None, - warehouse_id: Optional[str] = None): + query: Optional['outputs.JobTaskForEachTaskTaskSqlTaskQuery'] = None): """ + :param str warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. :param 'JobTaskForEachTaskTaskSqlTaskAlertArgs' alert: block consisting of following fields: :param 'JobTaskForEachTaskTaskSqlTaskDashboardArgs' dashboard: block consisting of following fields: :param 'JobTaskForEachTaskTaskSqlTaskFileArgs' file: block consisting of single string fields: :param Mapping[str, Any] parameters: (Map) parameters to be used for each run of this task. The SQL alert task does not support custom parameters. :param 'JobTaskForEachTaskTaskSqlTaskQueryArgs' query: block consisting of single string field: `query_id` - identifier of the Databricks SQL Query (databricks_sql_query). - :param str warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. """ + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -10637,8 +10598,14 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> str: + """ + ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. + """ + return pulumi.get(self, "warehouse_id") @property @pulumi.getter @@ -10680,14 +10647,6 @@ def query(self) -> Optional['outputs.JobTaskForEachTaskTaskSqlTaskQuery']: """ return pulumi.get(self, "query") - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[str]: - """ - ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - """ - return pulumi.get(self, "warehouse_id") - @pulumi.output_type class JobTaskForEachTaskTaskSqlTaskAlert(dict): @@ -11232,24 +11191,21 @@ def rules(self) -> Sequence['outputs.JobTaskHealthRule']: @pulumi.output_type class JobTaskHealthRule(dict): def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): + metric: str, + op: str, + value: int): """ :param str metric: string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). :param str op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param int value: integer value used to compare to the given metric. """ - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: """ string specifying the metric to check. The only supported metric is `RUN_DURATION_SECONDS` (check [Jobs REST API documentation](https://docs.databricks.com/api/workspace/jobs/create) for the latest information). """ @@ -11257,7 +11213,7 @@ def metric(self) -> Optional[str]: @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: """ string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ @@ -11265,7 +11221,7 @@ def op(self) -> Optional[str]: @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: """ integer value used to compare to the given metric. """ @@ -11410,8 +11366,6 @@ def __key_warning(key: str): suggest = "spark_version" elif key == "applyPolicyDefaultValues": suggest = "apply_policy_default_values" - elif key == "autoterminationMinutes": - suggest = "autotermination_minutes" elif key == "awsAttributes": suggest = "aws_attributes" elif key == "azureAttributes": @@ -11480,7 +11434,6 @@ def __init__(__self__, *, spark_version: str, apply_policy_default_values: Optional[bool] = None, autoscale: Optional['outputs.JobTaskNewClusterAutoscale'] = None, - autotermination_minutes: Optional[int] = None, aws_attributes: Optional['outputs.JobTaskNewClusterAwsAttributes'] = None, azure_attributes: Optional['outputs.JobTaskNewClusterAzureAttributes'] = None, cluster_id: Optional[str] = None, @@ -11516,8 +11469,6 @@ def __init__(__self__, *, pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) if autoscale is not None: pulumi.set(__self__, "autoscale", autoscale) - if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) if aws_attributes is not None: pulumi.set(__self__, "aws_attributes", aws_attributes) if azure_attributes is not None: @@ -11588,11 +11539,6 @@ def apply_policy_default_values(self) -> Optional[bool]: def autoscale(self) -> Optional['outputs.JobTaskNewClusterAutoscale']: return pulumi.get(self, "autoscale") - @property - @pulumi.getter(name="autoterminationMinutes") - def autotermination_minutes(self) -> Optional[int]: - return pulumi.get(self, "autotermination_minutes") - @property @pulumi.getter(name="awsAttributes") def aws_attributes(self) -> Optional['outputs.JobTaskNewClusterAwsAttributes']: @@ -13299,20 +13245,21 @@ def get(self, key: str, default = None) -> Any: return super().get(key, default) def __init__(__self__, *, + warehouse_id: str, alert: Optional['outputs.JobTaskSqlTaskAlert'] = None, dashboard: Optional['outputs.JobTaskSqlTaskDashboard'] = None, file: Optional['outputs.JobTaskSqlTaskFile'] = None, parameters: Optional[Mapping[str, Any]] = None, - query: Optional['outputs.JobTaskSqlTaskQuery'] = None, - warehouse_id: Optional[str] = None): + query: Optional['outputs.JobTaskSqlTaskQuery'] = None): """ + :param str warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. :param 'JobTaskSqlTaskAlertArgs' alert: block consisting of following fields: :param 'JobTaskSqlTaskDashboardArgs' dashboard: block consisting of following fields: :param 'JobTaskSqlTaskFileArgs' file: block consisting of single string fields: :param Mapping[str, Any] parameters: (Map) parameters to be used for each run of this task. The SQL alert task does not support custom parameters. :param 'JobTaskSqlTaskQueryArgs' query: block consisting of single string field: `query_id` - identifier of the Databricks SQL Query (databricks_sql_query). - :param str warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. """ + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -13323,8 +13270,14 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> str: + """ + ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. + """ + return pulumi.get(self, "warehouse_id") @property @pulumi.getter @@ -13366,14 +13319,6 @@ def query(self) -> Optional['outputs.JobTaskSqlTaskQuery']: """ return pulumi.get(self, "query") - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[str]: - """ - ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. - """ - return pulumi.get(self, "warehouse_id") - @pulumi.output_type class JobTaskSqlTaskAlert(dict): @@ -13928,7 +13873,6 @@ def __init__(__self__, *, """ :param 'JobTriggerFileArrivalArgs' file_arrival: configuration block to define a trigger for [File Arrival events](https://learn.microsoft.com/en-us/azure/databricks/workflows/jobs/file-arrival-triggers) consisting of following attributes: :param str pause_status: Indicate whether this trigger is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pause_status`. - :param 'JobTriggerTableUpdateArgs' table_update: configuration block to define a trigger for Table Update events consisting of following attributes: """ if file_arrival is not None: pulumi.set(__self__, "file_arrival", file_arrival) @@ -13963,9 +13907,6 @@ def table(self) -> Optional['outputs.JobTriggerTable']: @property @pulumi.getter(name="tableUpdate") def table_update(self) -> Optional['outputs.JobTriggerTableUpdate']: - """ - configuration block to define a trigger for Table Update events consisting of following attributes: - """ return pulumi.get(self, "table_update") @@ -14116,12 +14057,6 @@ def __init__(__self__, *, condition: Optional[str] = None, min_time_between_triggers_seconds: Optional[int] = None, wait_after_last_change_seconds: Optional[int] = None): - """ - :param Sequence[str] table_names: A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - :param str condition: The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - :param int min_time_between_triggers_seconds: If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - :param int wait_after_last_change_seconds: If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - """ pulumi.set(__self__, "table_names", table_names) if condition is not None: pulumi.set(__self__, "condition", condition) @@ -14133,33 +14068,21 @@ def __init__(__self__, *, @property @pulumi.getter(name="tableNames") def table_names(self) -> Sequence[str]: - """ - A list of Delta tables to monitor for changes. The table name must be in the format `catalog_name.schema_name.table_name`. - """ return pulumi.get(self, "table_names") @property @pulumi.getter def condition(self) -> Optional[str]: - """ - The table(s) condition based on which to trigger a job run. Valid values are `ANY_UPDATED` or `ALL_UPDATED`. - """ return pulumi.get(self, "condition") @property @pulumi.getter(name="minTimeBetweenTriggersSeconds") def min_time_between_triggers_seconds(self) -> Optional[int]: - """ - If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. - """ return pulumi.get(self, "min_time_between_triggers_seconds") @property @pulumi.getter(name="waitAfterLastChangeSeconds") def wait_after_last_change_seconds(self) -> Optional[int]: - """ - If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. - """ return pulumi.get(self, "wait_after_last_change_seconds") @@ -25128,50 +25051,45 @@ def rules(self) -> Sequence['outputs.GetJobJobSettingsSettingsHealthRuleResult'] @pulumi.output_type class GetJobJobSettingsSettingsHealthRuleResult(dict): def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + metric: str, + op: str, + value: int): + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: return pulumi.get(self, "metric") @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: return pulumi.get(self, "op") @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: return pulumi.get(self, "value") @pulumi.output_type class GetJobJobSettingsSettingsJobClusterResult(dict): def __init__(__self__, *, - job_cluster_key: Optional[str] = None, - new_cluster: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult'] = None): - if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) - if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + job_cluster_key: str, + new_cluster: 'outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult'): + pulumi.set(__self__, "job_cluster_key", job_cluster_key) + pulumi.set(__self__, "new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") - def job_cluster_key(self) -> Optional[str]: + def job_cluster_key(self) -> str: return pulumi.get(self, "job_cluster_key") @property @pulumi.getter(name="newCluster") - def new_cluster(self) -> Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult']: + def new_cluster(self) -> 'outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult': return pulumi.get(self, "new_cluster") @@ -27227,6 +27145,7 @@ def parameters(self) -> Optional[Sequence[str]]: class GetJobJobSettingsSettingsTaskResult(dict): def __init__(__self__, *, retry_on_timeout: bool, + task_key: str, condition_task: Optional['outputs.GetJobJobSettingsSettingsTaskConditionTaskResult'] = None, dbt_task: Optional['outputs.GetJobJobSettingsSettingsTaskDbtTaskResult'] = None, depends_ons: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskDependsOnResult']] = None, @@ -27251,10 +27170,10 @@ def __init__(__self__, *, spark_python_task: Optional['outputs.GetJobJobSettingsSettingsTaskSparkPythonTaskResult'] = None, spark_submit_task: Optional['outputs.GetJobJobSettingsSettingsTaskSparkSubmitTaskResult'] = None, sql_task: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskResult'] = None, - task_key: Optional[str] = None, timeout_seconds: Optional[int] = None, webhook_notifications: Optional['outputs.GetJobJobSettingsSettingsTaskWebhookNotificationsResult'] = None): pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -27303,8 +27222,6 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: @@ -27315,6 +27232,11 @@ def __init__(__self__, *, def retry_on_timeout(self) -> bool: return pulumi.get(self, "retry_on_timeout") + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> str: + return pulumi.get(self, "task_key") + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskConditionTaskResult']: @@ -27435,11 +27357,6 @@ def spark_submit_task(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskSp def sql_task(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskResult']: return pulumi.get(self, "sql_task") - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[str]: - return pulumi.get(self, "task_key") - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[int]: @@ -27633,6 +27550,7 @@ def concurrency(self) -> Optional[int]: class GetJobJobSettingsSettingsTaskForEachTaskTaskResult(dict): def __init__(__self__, *, retry_on_timeout: bool, + task_key: str, condition_task: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTaskResult'] = None, dbt_task: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTaskResult'] = None, depends_ons: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDependsOnResult']] = None, @@ -27656,10 +27574,10 @@ def __init__(__self__, *, spark_python_task: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkPythonTaskResult'] = None, spark_submit_task: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSparkSubmitTaskResult'] = None, sql_task: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult'] = None, - task_key: Optional[str] = None, timeout_seconds: Optional[int] = None, webhook_notifications: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskWebhookNotificationsResult'] = None): pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + pulumi.set(__self__, "task_key", task_key) if condition_task is not None: pulumi.set(__self__, "condition_task", condition_task) if dbt_task is not None: @@ -27706,8 +27624,6 @@ def __init__(__self__, *, pulumi.set(__self__, "spark_submit_task", spark_submit_task) if sql_task is not None: pulumi.set(__self__, "sql_task", sql_task) - if task_key is not None: - pulumi.set(__self__, "task_key", task_key) if timeout_seconds is not None: pulumi.set(__self__, "timeout_seconds", timeout_seconds) if webhook_notifications is not None: @@ -27718,6 +27634,11 @@ def __init__(__self__, *, def retry_on_timeout(self) -> bool: return pulumi.get(self, "retry_on_timeout") + @property + @pulumi.getter(name="taskKey") + def task_key(self) -> str: + return pulumi.get(self, "task_key") + @property @pulumi.getter(name="conditionTask") def condition_task(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTaskResult']: @@ -27833,11 +27754,6 @@ def spark_submit_task(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskFo def sql_task(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult']: return pulumi.get(self, "sql_task") - @property - @pulumi.getter(name="taskKey") - def task_key(self) -> Optional[str]: - return pulumi.get(self, "task_key") - @property @pulumi.getter(name="timeoutSeconds") def timeout_seconds(self) -> Optional[int]: @@ -28015,29 +27931,26 @@ def rules(self) -> Sequence['outputs.GetJobJobSettingsSettingsTaskForEachTaskTas @pulumi.output_type class GetJobJobSettingsSettingsTaskForEachTaskTaskHealthRuleResult(dict): def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + metric: str, + op: str, + value: int): + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: return pulumi.get(self, "metric") @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: return pulumi.get(self, "op") @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: return pulumi.get(self, "value") @@ -29200,12 +29113,13 @@ def parameters(self) -> Optional[Sequence[str]]: @pulumi.output_type class GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskResult(dict): def __init__(__self__, *, + warehouse_id: str, alert: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlertResult'] = None, dashboard: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskDashboardResult'] = None, file: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskFileResult'] = None, parameters: Optional[Mapping[str, Any]] = None, - query: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryResult'] = None, - warehouse_id: Optional[str] = None): + query: Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryResult'] = None): + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -29216,8 +29130,11 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> str: + return pulumi.get(self, "warehouse_id") @property @pulumi.getter @@ -29244,11 +29161,6 @@ def parameters(self) -> Optional[Mapping[str, Any]]: def query(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskQueryResult']: return pulumi.get(self, "query") - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[str]: - return pulumi.get(self, "warehouse_id") - @pulumi.output_type class GetJobJobSettingsSettingsTaskForEachTaskTaskSqlTaskAlertResult(dict): @@ -29511,29 +29423,26 @@ def rules(self) -> Sequence['outputs.GetJobJobSettingsSettingsTaskHealthRuleResu @pulumi.output_type class GetJobJobSettingsSettingsTaskHealthRuleResult(dict): def __init__(__self__, *, - metric: Optional[str] = None, - op: Optional[str] = None, - value: Optional[int] = None): - if metric is not None: - pulumi.set(__self__, "metric", metric) - if op is not None: - pulumi.set(__self__, "op", op) - if value is not None: - pulumi.set(__self__, "value", value) + metric: str, + op: str, + value: int): + pulumi.set(__self__, "metric", metric) + pulumi.set(__self__, "op", op) + pulumi.set(__self__, "value", value) @property @pulumi.getter - def metric(self) -> Optional[str]: + def metric(self) -> str: return pulumi.get(self, "metric") @property @pulumi.getter - def op(self) -> Optional[str]: + def op(self) -> str: return pulumi.get(self, "op") @property @pulumi.getter - def value(self) -> Optional[int]: + def value(self) -> int: return pulumi.get(self, "value") @@ -30696,12 +30605,13 @@ def parameters(self) -> Optional[Sequence[str]]: @pulumi.output_type class GetJobJobSettingsSettingsTaskSqlTaskResult(dict): def __init__(__self__, *, + warehouse_id: str, alert: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskAlertResult'] = None, dashboard: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskDashboardResult'] = None, file: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskFileResult'] = None, parameters: Optional[Mapping[str, Any]] = None, - query: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskQueryResult'] = None, - warehouse_id: Optional[str] = None): + query: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskQueryResult'] = None): + pulumi.set(__self__, "warehouse_id", warehouse_id) if alert is not None: pulumi.set(__self__, "alert", alert) if dashboard is not None: @@ -30712,8 +30622,11 @@ def __init__(__self__, *, pulumi.set(__self__, "parameters", parameters) if query is not None: pulumi.set(__self__, "query", query) - if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + + @property + @pulumi.getter(name="warehouseId") + def warehouse_id(self) -> str: + return pulumi.get(self, "warehouse_id") @property @pulumi.getter @@ -30740,11 +30653,6 @@ def parameters(self) -> Optional[Mapping[str, Any]]: def query(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskQueryResult']: return pulumi.get(self, "query") - @property - @pulumi.getter(name="warehouseId") - def warehouse_id(self) -> Optional[str]: - return pulumi.get(self, "warehouse_id") - @pulumi.output_type class GetJobJobSettingsSettingsTaskSqlTaskAlertResult(dict):