From b0a2887712568a9232150f2953ad6f0303f8a724 Mon Sep 17 00:00:00 2001 From: Pulumi Bot <30351955+pulumi-bot@users.noreply.github.com> Date: Tue, 10 Oct 2023 12:30:34 -0700 Subject: [PATCH] Upgrade pulumi-terraform-bridge to v3.61.0 (#217) This PR was generated via `$ upgrade-provider pulumi/pulumi-databricks --kind=bridge`. --- - Upgrading pulumi-terraform-bridge from v3.60.1 to v3.61.0. Fixes https://github.com/pulumi/pulumi-databricks/issues/209 --- examples/go.mod | 6 +- examples/go.sum | 19 +- .../pulumi-resource-databricks/schema.json | 23 +- provider/go.mod | 8 +- provider/go.sum | 23 +- sdk/dotnet/Config/Config.cs | 2 +- .../JobJobClusterNewClusterDockerImageArgs.cs | 2 +- ...bJobClusterNewClusterDockerImageGetArgs.cs | 2 +- .../Inputs/JobNewClusterDockerImageArgs.cs | 2 +- .../Inputs/JobNewClusterDockerImageGetArgs.cs | 2 +- .../JobTaskNewClusterDockerImageArgs.cs | 2 +- .../JobTaskNewClusterDockerImageGetArgs.cs | 2 +- .../Inputs/JobTriggerFileArrivalArgs.cs | 2 +- .../Inputs/JobTriggerFileArrivalGetArgs.cs | 2 +- ...sOnDurationWarningThresholdExceededArgs.cs | 4 +- ...DurationWarningThresholdExceededGetArgs.cs | 4 +- .../JobWebhookNotificationsOnFailureArgs.cs | 4 +- ...JobWebhookNotificationsOnFailureGetArgs.cs | 4 +- .../JobWebhookNotificationsOnStartArgs.cs | 4 +- .../JobWebhookNotificationsOnStartGetArgs.cs | 4 +- .../JobWebhookNotificationsOnSuccessArgs.cs | 4 +- ...JobWebhookNotificationsOnSuccessGetArgs.cs | 4 +- sdk/dotnet/Inputs/ShareObjectPartitionArgs.cs | 4 + .../Inputs/ShareObjectPartitionGetArgs.cs | 4 + sdk/dotnet/Job.cs | 4 +- .../JobJobClusterNewClusterDockerImage.cs | 2 +- .../Outputs/JobNewClusterDockerImage.cs | 2 +- .../Outputs/JobTaskNewClusterDockerImage.cs | 2 +- sdk/dotnet/Outputs/JobTriggerFileArrival.cs | 2 +- ...tionsOnDurationWarningThresholdExceeded.cs | 4 +- .../JobWebhookNotificationsOnFailure.cs | 4 +- .../Outputs/JobWebhookNotificationsOnStart.cs | 4 +- .../JobWebhookNotificationsOnSuccess.cs | 4 +- sdk/dotnet/Outputs/ShareObjectPartition.cs | 3 + sdk/go.mod | 2 +- sdk/go.sum | 4 +- sdk/go/databricks/job.go | 8 +- sdk/go/databricks/pulumiTypes.go | 83 +- .../main/java/com/pulumi/databricks/Job.java | 4 +- ...obJobClusterNewClusterDockerImageArgs.java | 8 +- .../inputs/JobNewClusterDockerImageArgs.java | 8 +- .../pulumi/databricks/inputs/JobState.java | 8 +- .../JobTaskNewClusterDockerImageArgs.java | 8 +- .../inputs/JobTriggerFileArrivalArgs.java | 8 +- ...nDurationWarningThresholdExceededArgs.java | 16 +- .../JobWebhookNotificationsOnFailureArgs.java | 16 +- .../JobWebhookNotificationsOnStartArgs.java | 16 +- .../JobWebhookNotificationsOnSuccessArgs.java | 16 +- .../inputs/ShareObjectPartitionArgs.java | 26 + .../JobJobClusterNewClusterDockerImage.java | 4 +- .../outputs/JobNewClusterDockerImage.java | 4 +- .../outputs/JobTaskNewClusterDockerImage.java | 4 +- .../outputs/JobTriggerFileArrival.java | 4 +- ...onsOnDurationWarningThresholdExceeded.java | 8 +- .../JobWebhookNotificationsOnFailure.java | 8 +- .../JobWebhookNotificationsOnStart.java | 8 +- .../JobWebhookNotificationsOnSuccess.java | 8 +- .../outputs/ShareObjectPartition.java | 8 + sdk/nodejs/job.ts | 4 +- sdk/nodejs/types/input.ts | 27 +- sdk/nodejs/types/output.ts | 27 +- sdk/python/pulumi_databricks/_inputs.py | 9451 +++++++++++++--- .../access_control_rule_set.py | 40 +- sdk/python/pulumi_databricks/catalog.py | 116 +- .../catalog_workspace_binding.py | 36 +- sdk/python/pulumi_databricks/cluster.py | 326 +- .../pulumi_databricks/cluster_policy.py | 72 +- .../pulumi_databricks/config/__init__.pyi | 2 +- sdk/python/pulumi_databricks/config/vars.py | 2 +- sdk/python/pulumi_databricks/connection.py | 84 +- sdk/python/pulumi_databricks/dbfs_file.py | 60 +- sdk/python/pulumi_databricks/directory.py | 44 +- sdk/python/pulumi_databricks/entitlements.py | 76 +- .../pulumi_databricks/external_location.py | 121 +- .../get_aws_assume_role_policy.py | 2 +- .../get_aws_bucket_policy.py | 2 +- .../get_aws_cross_account_policy.py | 2 +- sdk/python/pulumi_databricks/get_catalogs.py | 2 +- sdk/python/pulumi_databricks/get_cluster.py | 2 +- .../pulumi_databricks/get_cluster_policy.py | 2 +- sdk/python/pulumi_databricks/get_clusters.py | 2 +- .../pulumi_databricks/get_current_user.py | 31 +- sdk/python/pulumi_databricks/get_dbfs_file.py | 2 +- .../pulumi_databricks/get_dbfs_file_paths.py | 2 +- sdk/python/pulumi_databricks/get_directory.py | 2 +- sdk/python/pulumi_databricks/get_group.py | 2 +- .../pulumi_databricks/get_instance_pool.py | 2 +- sdk/python/pulumi_databricks/get_job.py | 2 +- sdk/python/pulumi_databricks/get_jobs.py | 2 +- sdk/python/pulumi_databricks/get_metastore.py | 2 +- .../pulumi_databricks/get_metastores.py | 2 +- .../pulumi_databricks/get_mws_credentials.py | 2 +- .../pulumi_databricks/get_mws_workspaces.py | 2 +- sdk/python/pulumi_databricks/get_node_type.py | 2 +- sdk/python/pulumi_databricks/get_notebook.py | 2 +- .../pulumi_databricks/get_notebook_paths.py | 2 +- sdk/python/pulumi_databricks/get_pipelines.py | 2 +- sdk/python/pulumi_databricks/get_schemas.py | 2 +- .../get_service_principal.py | 2 +- .../get_service_principals.py | 2 +- sdk/python/pulumi_databricks/get_share.py | 2 +- sdk/python/pulumi_databricks/get_shares.py | 2 +- .../pulumi_databricks/get_spark_version.py | 2 +- .../pulumi_databricks/get_sql_warehouse.py | 2 +- .../pulumi_databricks/get_sql_warehouses.py | 2 +- sdk/python/pulumi_databricks/get_tables.py | 2 +- sdk/python/pulumi_databricks/get_user.py | 2 +- sdk/python/pulumi_databricks/get_views.py | 2 +- sdk/python/pulumi_databricks/get_zones.py | 22 +- .../pulumi_databricks/git_credential.py | 52 +- .../pulumi_databricks/global_init_script.py | 68 +- sdk/python/pulumi_databricks/grants.py | 124 +- sdk/python/pulumi_databricks/group.py | 92 +- .../group_instance_profile.py | 36 +- sdk/python/pulumi_databricks/group_member.py | 36 +- sdk/python/pulumi_databricks/group_role.py | 36 +- sdk/python/pulumi_databricks/instance_pool.py | 165 +- .../pulumi_databricks/instance_profile.py | 52 +- .../pulumi_databricks/ip_access_list.py | 52 +- sdk/python/pulumi_databricks/job.py | 410 +- sdk/python/pulumi_databricks/library.py | 91 +- sdk/python/pulumi_databricks/metastore.py | 156 +- .../pulumi_databricks/metastore_assignment.py | 44 +- .../metastore_data_access.py | 141 +- .../pulumi_databricks/metastore_provider.py | 52 +- .../pulumi_databricks/mlflow_experiment.py | 76 +- sdk/python/pulumi_databricks/mlflow_model.py | 72 +- .../pulumi_databricks/mlflow_webhook.py | 78 +- sdk/python/pulumi_databricks/model_serving.py | 53 +- sdk/python/pulumi_databricks/mount.py | 137 +- .../pulumi_databricks/mws_credentials.py | 56 +- .../mws_customer_managed_keys.py | 78 +- .../pulumi_databricks/mws_log_delivery.py | 108 +- sdk/python/pulumi_databricks/mws_networks.py | 126 +- .../mws_permission_assignment.py | 44 +- .../mws_private_access_settings.py | 84 +- .../mws_storage_configurations.py | 52 +- .../pulumi_databricks/mws_vpc_endpoint.py | 105 +- .../pulumi_databricks/mws_workspaces.py | 248 +- sdk/python/pulumi_databricks/notebook.py | 91 +- sdk/python/pulumi_databricks/obo_token.py | 48 +- sdk/python/pulumi_databricks/outputs.py | 9475 ++++++++++++++--- .../permission_assignment.py | 36 +- sdk/python/pulumi_databricks/permissions.py | 204 +- sdk/python/pulumi_databricks/pipeline.py | 157 +- sdk/python/pulumi_databricks/provider.py | 129 +- sdk/python/pulumi_databricks/recipient.py | 81 +- sdk/python/pulumi_databricks/repo.py | 81 +- sdk/python/pulumi_databricks/schema.py | 84 +- sdk/python/pulumi_databricks/secret.py | 52 +- sdk/python/pulumi_databricks/secret_acl.py | 44 +- sdk/python/pulumi_databricks/secret_scope.py | 57 +- .../pulumi_databricks/service_principal.py | 140 +- .../service_principal_role.py | 36 +- .../service_principal_secret.py | 44 +- sdk/python/pulumi_databricks/share.py | 60 +- sdk/python/pulumi_databricks/sql_alert.py | 81 +- sdk/python/pulumi_databricks/sql_dashboard.py | 60 +- sdk/python/pulumi_databricks/sql_endpoint.py | 171 +- .../pulumi_databricks/sql_global_config.py | 71 +- .../pulumi_databricks/sql_permissions.py | 84 +- sdk/python/pulumi_databricks/sql_query.py | 116 +- sdk/python/pulumi_databricks/sql_table.py | 116 +- .../pulumi_databricks/sql_visualization.py | 76 +- sdk/python/pulumi_databricks/sql_widget.py | 89 +- .../pulumi_databricks/storage_credential.py | 133 +- sdk/python/pulumi_databricks/table.py | 116 +- sdk/python/pulumi_databricks/token.py | 64 +- sdk/python/pulumi_databricks/user.py | 140 +- .../user_instance_profile.py | 36 +- sdk/python/pulumi_databricks/user_role.py | 36 +- sdk/python/pulumi_databricks/volume.py | 76 +- .../pulumi_databricks/workspace_conf.py | 28 +- .../pulumi_databricks/workspace_file.py | 64 +- 174 files changed, 21458 insertions(+), 4904 deletions(-) diff --git a/examples/go.mod b/examples/go.mod index daab96e3..55e66012 100644 --- a/examples/go.mod +++ b/examples/go.mod @@ -2,7 +2,7 @@ module github.com/pulumi/pulumi-databricks/examples go 1.21 -require github.com/pulumi/pulumi/pkg/v3 v3.81.0 +require github.com/pulumi/pulumi/pkg/v3 v3.86.0 require ( cloud.google.com/go v0.110.4 // indirect @@ -13,7 +13,7 @@ require ( cloud.google.com/go/logging v1.7.0 // indirect cloud.google.com/go/longrunning v0.5.1 // indirect cloud.google.com/go/storage v1.30.1 // indirect - github.com/AlecAivazis/survey/v2 v2.0.5 // indirect + github.com/AlecAivazis/survey/v2 v2.3.7 // indirect github.com/Azure/azure-sdk-for-go v66.0.0+incompatible // indirect github.com/Azure/azure-sdk-for-go/sdk/azcore v1.1.1 // indirect github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.0.0 // indirect @@ -154,7 +154,7 @@ require ( github.com/pkg/errors v0.9.1 // indirect github.com/pkg/term v1.1.0 // indirect github.com/pmezard/go-difflib v1.0.0 // indirect - github.com/pulumi/pulumi/sdk/v3 v3.81.0 // indirect + github.com/pulumi/pulumi/sdk/v3 v3.86.0 // indirect github.com/rivo/uniseg v0.4.4 // indirect github.com/rogpeppe/go-internal v1.9.0 // indirect github.com/ryanuber/go-glob v1.0.0 // indirect diff --git a/examples/go.sum b/examples/go.sum index 18ee8c80..ab745422 100644 --- a/examples/go.sum +++ b/examples/go.sum @@ -93,8 +93,8 @@ contrib.go.opencensus.io/exporter/stackdriver v0.13.13/go.mod h1:5pSSGY0Bhuk7waT contrib.go.opencensus.io/integrations/ocsql v0.1.7/go.mod h1:8DsSdjz3F+APR+0z0WkU1aRorQCFfRxvqjUUPMbF3fE= dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7gZCb22OMCxBHrMx7a5I7Hp++hsVxbQ4BYO7hU= github.com/AdaLogics/go-fuzz-headers v0.0.0-20210715213245-6c3934b029d8/go.mod h1:CzsSbkDixRphAF5hS6wbMKq0eI6ccJRb7/A0M6JBnwg= -github.com/AlecAivazis/survey/v2 v2.0.5 h1:xpZp+Q55wi5C7Iaze+40onHnEkex1jSc34CltJjOoPM= -github.com/AlecAivazis/survey/v2 v2.0.5/go.mod h1:WYBhg6f0y/fNYUuesWQc0PKbJcEliGcYHB9sNT3Bg74= +github.com/AlecAivazis/survey/v2 v2.3.7 h1:6I/u8FvytdGsgonrYsVn2t8t4QiRnh6QSTqkkhIiSjQ= +github.com/AlecAivazis/survey/v2 v2.3.7/go.mod h1:xUTIdE4KCOIjsBAE1JYsUPoCqYdZ1reCfTwbto0Fduo= github.com/Azure/azure-amqp-common-go/v3 v3.2.3/go.mod h1:7rPmbSfszeovxGfc5fSAXE4ehlXQZHpMja2OtxC2Tas= github.com/Azure/azure-sdk-for-go v16.2.1+incompatible/go.mod h1:9XXNKU+eRnpl9moKnB4QOLf1HestfXbmab5FXxiDBjc= github.com/Azure/azure-sdk-for-go v63.0.0+incompatible/go.mod h1:9XXNKU+eRnpl9moKnB4QOLf1HestfXbmab5FXxiDBjc= @@ -192,7 +192,6 @@ github.com/Microsoft/hcsshim/test v0.0.0-20201218223536-d3e5debf77da/go.mod h1:5 github.com/Microsoft/hcsshim/test v0.0.0-20210227013316-43a75bb4edd3/go.mod h1:mw7qgWloBUl75W/gVH3cQszUg1+gUITj7D6NY7ywVnY= github.com/NYTimes/gziphandler v0.0.0-20170623195520-56545f4a5d46/go.mod h1:3wb06e3pkSAbeQ52E9H9iFoQsEEwGN64994WTCIhntQ= github.com/NYTimes/gziphandler v1.1.1/go.mod h1:n/CVRwUEOgIxrgPvAQhUUr9oeUtvrhMomdKFjzJNB0c= -github.com/Netflix/go-expect v0.0.0-20180615182759-c93bf25de8e8/go.mod h1:oX5x61PbNXchhh0oikYAH+4Pcfw5LKv21+Jnpr6r6Pc= github.com/Netflix/go-expect v0.0.0-20220104043353-73e0943537d2 h1:+vx7roKuyA63nhn5WAunQHLTznkw5W8b1Xc0dNjp83s= github.com/Netflix/go-expect v0.0.0-20220104043353-73e0943537d2/go.mod h1:HBCaDeC1lPdgDeDbhX8XFpy1jqjK0IBG8W5K+xYqA0w= github.com/OneOfOne/xxhash v1.2.2/go.mod h1:HSdplMjZKSmBqAxg5vPj2TmRDmfkzw+cTzAElWljhcU= @@ -1017,7 +1016,7 @@ github.com/hashicorp/yamux v0.1.1 h1:yrQxtgseBDrq9Y652vSRDvsKCJKOUD+GzTS4Y0Y8pvE github.com/hashicorp/yamux v0.1.1/go.mod h1:CtWFDAQgb7dxtzFs4tWbplKIe2jSi3+5vKbgIO0SLnQ= github.com/hetznercloud/hcloud-go v1.33.1/go.mod h1:XX/TQub3ge0yWR2yHWmnDVIrB+MQbda1pHxkUmDlUME= github.com/hetznercloud/hcloud-go v1.35.0/go.mod h1:mepQwR6va27S3UQthaEPGS86jtzSY9xWL1e9dyxXpgA= -github.com/hinshun/vt10x v0.0.0-20180616224451-1954e6464174/go.mod h1:DqJ97dSdRW1W22yXSB90986pcOyQ7r45iio1KN2ez1A= +github.com/hinshun/vt10x v0.0.0-20220119200601-820417d04eec/go.mod h1:Q48J4R4DvxnHolD5P8pOtXigYlRuPLGl6moFx3ulM68= github.com/hinshun/vt10x v0.0.0-20220301184237-5011da428d02 h1:AgcIVYPa6XJnU3phs104wLj8l5GEththEw6+F79YsIY= github.com/hinshun/vt10x v0.0.0-20220301184237-5011da428d02/go.mod h1:Q48J4R4DvxnHolD5P8pOtXigYlRuPLGl6moFx3ulM68= github.com/hpcloud/tail v1.0.0/go.mod h1:ab1qPbhIpdTxEkNHXyeSf5vhxWSCs/tWer42PpOxQnU= @@ -1142,7 +1141,6 @@ github.com/kr/pretty v0.3.0/go.mod h1:640gp4NfQd8pI5XOwp5fnNeVWj67G7CFk/SaSQn7NB github.com/kr/pretty v0.3.1 h1:flRD4NNwYAUpkphVc1HcthR4KEIFJ65n8Mw5qdRn3LE= github.com/kr/pretty v0.3.1/go.mod h1:hoEshYVHaxMs3cyo3Yncou5ZscifuDolrwPKZanG3xk= github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= -github.com/kr/pty v1.1.4/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= github.com/kr/pty v1.1.5/go.mod h1:9r2w37qlBe7rQ6e1fg1S/9xpWHSnaqNdHD3WcMdbPDA= github.com/kr/pty v1.1.8/go.mod h1:O1sed60cT9XZ5uDucP5qwvh+TE3NnUj51EiZO/lmSfw= github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI= @@ -1464,10 +1462,10 @@ github.com/prometheus/procfs v0.7.3/go.mod h1:cz+aTbrPOrUb4q7XlbU9ygM+/jj0fzG6c1 github.com/prometheus/prometheus v0.35.0/go.mod h1:7HaLx5kEPKJ0GDgbODG0fZgXbQ8K/XjZNJXQmbmgQlY= github.com/prometheus/prometheus v0.37.0/go.mod h1:egARUgz+K93zwqsVIAneFlLZefyGOON44WyAp4Xqbbk= github.com/prometheus/tsdb v0.7.1/go.mod h1:qhTCs0VvXwvX/y3TZrWD7rabWM+ijKTux40TwIPHuXU= -github.com/pulumi/pulumi/pkg/v3 v3.81.0 h1:6rf2farQLszi8inHCu9YdJtDvK0fqNguix51b3FEDRQ= -github.com/pulumi/pulumi/pkg/v3 v3.81.0/go.mod h1:URXozJB/abhxb2yOwa5ywHjT2dqt9SmAX2iLkdcOOl0= -github.com/pulumi/pulumi/sdk/v3 v3.81.0 h1:d711GUpjZK1beNpIKFReC1LBmHXYE684sYWntJWQ4Q0= -github.com/pulumi/pulumi/sdk/v3 v3.81.0/go.mod h1:RMilNNVMlmK1h4Nl/qylb9vzbgh4F3mufZoUOnPy98o= +github.com/pulumi/pulumi/pkg/v3 v3.86.0 h1:G4spuT89ZN8lSxT9WkMF/JaP7n+wu7ubEly7Yy8uza0= +github.com/pulumi/pulumi/pkg/v3 v3.86.0/go.mod h1:Qs55gPhUwM/Dua3VRtHXLLlpY8uEe+llDBIZc+1pvHM= +github.com/pulumi/pulumi/sdk/v3 v3.86.0 h1:Cxg0rGdvMt9GqGvesFTj8+WaO/ihmALYlQf4zm1GzFw= +github.com/pulumi/pulumi/sdk/v3 v3.86.0/go.mod h1:vexSGJ5L834l3T7Fo/KpdywjGAgZ2CvChuqGFl+ebS0= github.com/rakyll/embedmd v0.0.0-20171029212350-c8060a0752a2/go.mod h1:7jOTMgqac46PZcF54q6l2hkLEG8op93fZu61KmxWDV4= github.com/rcrowley/go-metrics v0.0.0-20181016184325-3113b8401b8a/go.mod h1:bCqnVzQkZxMG4s8nGwiZ5l3QUCyqpo9Y+/ZMZ9VjZe4= github.com/rivo/uniseg v0.1.0/go.mod h1:J6wj4VEh+S6ZtnVlnTBMWIodfgj8LQOQFoIToxlJtxc= @@ -1570,7 +1568,6 @@ github.com/stretchr/objx v0.4.0/go.mod h1:YvHI0jy2hoMjB+UWwv71VJQ9isScKT/TqJzVSS github.com/stretchr/objx v0.5.0 h1:1zr/of2m5FGMsad5YfcqgdqdWrIhu+EBEJRhR1U7z/c= github.com/stretchr/objx v0.5.0/go.mod h1:Yh+to48EsGEfYuaHDzXPcE3xhTkx73EhmCGUpEOglKo= github.com/stretchr/testify v0.0.0-20180303142811-b89eecf5ca5d/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= -github.com/stretchr/testify v1.2.1/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= github.com/stretchr/testify v1.2.2/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI= github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4= @@ -1759,7 +1756,6 @@ golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACk golang.org/x/crypto v0.0.0-20190411191339-88737f569e3a/go.mod h1:WFFai1msRO1wXaEeE5yQxYXgSfI8pQAWXbQop6sCtWE= golang.org/x/crypto v0.0.0-20190422162423-af44ce270edf/go.mod h1:WFFai1msRO1wXaEeE5yQxYXgSfI8pQAWXbQop6sCtWE= golang.org/x/crypto v0.0.0-20190510104115-cbcb75029529/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= -golang.org/x/crypto v0.0.0-20190530122614-20be4c3c3ed5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190605123033-f99c8df09eb5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190611184440-5c40567a22f8/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190701094942-4def268fd1a4/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= @@ -1980,7 +1976,6 @@ golang.org/x/sys v0.0.0-20190502145724-3ef323f4f1fd/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20190507160741-ecd444e8653b/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190514135907-3a4b5fb9f71f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190522044717-8097e1b27ff5/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20190530182044-ad28b68e88f1/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190531175056-4c3a928424d2/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190602015325-4c4f7f33c9ed/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190606165138-5da285871e9c/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= diff --git a/provider/cmd/pulumi-resource-databricks/schema.json b/provider/cmd/pulumi-resource-databricks/schema.json index cfc32608..84bdd7a2 100644 --- a/provider/cmd/pulumi-resource-databricks/schema.json +++ b/provider/cmd/pulumi-resource-databricks/schema.json @@ -1256,7 +1256,7 @@ }, "url": { "type": "string", - "description": "string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`).\n" + "description": "URL of the job on the given workspace\n" } }, "type": "object", @@ -1777,7 +1777,7 @@ }, "url": { "type": "string", - "description": "string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`).\n" + "description": "URL of the job on the given workspace\n" } }, "type": "object", @@ -2747,7 +2747,7 @@ }, "url": { "type": "string", - "description": "string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`).\n" + "description": "URL of the job on the given workspace\n" } }, "type": "object", @@ -3233,7 +3233,7 @@ }, "url": { "type": "string", - "description": "string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`).\n" + "description": "URL of the job on the given workspace\n" }, "waitAfterLastChangeSeconds": { "type": "integer", @@ -3282,7 +3282,7 @@ "properties": { "id": { "type": "string", - "description": "ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered.\n\n\u003e **Note** The following configuration blocks can be standalone or nested inside a `task` block\n" + "description": "ID of the job\n" } }, "type": "object", @@ -3294,7 +3294,7 @@ "properties": { "id": { "type": "string", - "description": "ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered.\n\n\u003e **Note** The following configuration blocks can be standalone or nested inside a `task` block\n" + "description": "ID of the job\n" } }, "type": "object", @@ -3306,7 +3306,7 @@ "properties": { "id": { "type": "string", - "description": "ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered.\n\n\u003e **Note** The following configuration blocks can be standalone or nested inside a `task` block\n" + "description": "ID of the job\n" } }, "type": "object", @@ -3318,7 +3318,7 @@ "properties": { "id": { "type": "string", - "description": "ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered.\n\n\u003e **Note** The following configuration blocks can be standalone or nested inside a `task` block\n" + "description": "ID of the job\n" } }, "type": "object", @@ -4724,7 +4724,8 @@ "type": "array", "items": { "$ref": "#/types/databricks:index/ShareObjectPartitionValue:ShareObjectPartitionValue" - } + }, + "description": "The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set.\n" } }, "type": "object", @@ -11789,7 +11790,7 @@ }, "url": { "type": "string", - "description": "string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`).\n" + "description": "URL of the job on the given workspace\n" }, "webhookNotifications": { "$ref": "#/types/databricks:index/JobWebhookNotifications:JobWebhookNotifications", @@ -12084,7 +12085,7 @@ }, "url": { "type": "string", - "description": "string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`).\n" + "description": "URL of the job on the given workspace\n" }, "webhookNotifications": { "$ref": "#/types/databricks:index/JobWebhookNotifications:JobWebhookNotifications", diff --git a/provider/go.mod b/provider/go.mod index 7eb8c5eb..4aec9705 100644 --- a/provider/go.mod +++ b/provider/go.mod @@ -7,7 +7,7 @@ replace github.com/hashicorp/terraform-plugin-sdk/v2 => github.com/pulumi/terraf require ( github.com/databricks/databricks-sdk-go v0.20.0 github.com/databricks/terraform-provider-databricks v1.27.0 - github.com/pulumi/pulumi-terraform-bridge/v3 v3.60.1 + github.com/pulumi/pulumi-terraform-bridge/v3 v3.61.0 ) require ( @@ -20,7 +20,7 @@ require ( cloud.google.com/go/longrunning v0.5.1 // indirect cloud.google.com/go/storage v1.30.1 // indirect dario.cat/mergo v1.0.0 // indirect - github.com/AlecAivazis/survey/v2 v2.0.5 // indirect + github.com/AlecAivazis/survey/v2 v2.3.7 // indirect github.com/Azure/azure-sdk-for-go/sdk/azcore v1.3.1 // indirect github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.2.1 // indirect github.com/Azure/azure-sdk-for-go/sdk/internal v1.1.2 // indirect @@ -187,8 +187,8 @@ require ( github.com/pulumi/pulumi-java/pkg v0.9.6 // indirect github.com/pulumi/pulumi-terraform-bridge/x/muxer v0.0.4 // indirect github.com/pulumi/pulumi-yaml v1.2.2 // indirect - github.com/pulumi/pulumi/pkg/v3 v3.81.0 // indirect - github.com/pulumi/pulumi/sdk/v3 v3.81.0 // indirect + github.com/pulumi/pulumi/pkg/v3 v3.86.0 // indirect + github.com/pulumi/pulumi/sdk/v3 v3.86.0 // indirect github.com/pulumi/schema-tools v0.1.2 // indirect github.com/pulumi/terraform-diff-reader v0.0.2 // indirect github.com/rivo/uniseg v0.4.4 // indirect diff --git a/provider/go.sum b/provider/go.sum index b92a8b6b..5d727f0c 100644 --- a/provider/go.sum +++ b/provider/go.sum @@ -634,8 +634,8 @@ dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7 gioui.org v0.0.0-20210308172011-57750fc8a0a6/go.mod h1:RSH6KIUZ0p2xy5zHDxgAM4zumjgTw83q2ge/PI+yyw8= git.sr.ht/~sbinet/gg v0.3.1/go.mod h1:KGYtlADtqsqANL9ueOFkWymvzUvLMQllU5Ixo+8v3pc= github.com/AdaLogics/go-fuzz-headers v0.0.0-20210715213245-6c3934b029d8/go.mod h1:CzsSbkDixRphAF5hS6wbMKq0eI6ccJRb7/A0M6JBnwg= -github.com/AlecAivazis/survey/v2 v2.0.5 h1:xpZp+Q55wi5C7Iaze+40onHnEkex1jSc34CltJjOoPM= -github.com/AlecAivazis/survey/v2 v2.0.5/go.mod h1:WYBhg6f0y/fNYUuesWQc0PKbJcEliGcYHB9sNT3Bg74= +github.com/AlecAivazis/survey/v2 v2.3.7 h1:6I/u8FvytdGsgonrYsVn2t8t4QiRnh6QSTqkkhIiSjQ= +github.com/AlecAivazis/survey/v2 v2.3.7/go.mod h1:xUTIdE4KCOIjsBAE1JYsUPoCqYdZ1reCfTwbto0Fduo= github.com/Azure/azure-amqp-common-go/v3 v3.2.3/go.mod h1:7rPmbSfszeovxGfc5fSAXE4ehlXQZHpMja2OtxC2Tas= github.com/Azure/azure-sdk-for-go v16.2.1+incompatible/go.mod h1:9XXNKU+eRnpl9moKnB4QOLf1HestfXbmab5FXxiDBjc= github.com/Azure/azure-sdk-for-go v63.0.0+incompatible/go.mod h1:9XXNKU+eRnpl9moKnB4QOLf1HestfXbmab5FXxiDBjc= @@ -755,7 +755,6 @@ github.com/Microsoft/hcsshim/test v0.0.0-20201218223536-d3e5debf77da/go.mod h1:5 github.com/Microsoft/hcsshim/test v0.0.0-20210227013316-43a75bb4edd3/go.mod h1:mw7qgWloBUl75W/gVH3cQszUg1+gUITj7D6NY7ywVnY= github.com/NYTimes/gziphandler v0.0.0-20170623195520-56545f4a5d46/go.mod h1:3wb06e3pkSAbeQ52E9H9iFoQsEEwGN64994WTCIhntQ= github.com/NYTimes/gziphandler v1.1.1/go.mod h1:n/CVRwUEOgIxrgPvAQhUUr9oeUtvrhMomdKFjzJNB0c= -github.com/Netflix/go-expect v0.0.0-20180615182759-c93bf25de8e8/go.mod h1:oX5x61PbNXchhh0oikYAH+4Pcfw5LKv21+Jnpr6r6Pc= github.com/Netflix/go-expect v0.0.0-20220104043353-73e0943537d2 h1:+vx7roKuyA63nhn5WAunQHLTznkw5W8b1Xc0dNjp83s= github.com/Netflix/go-expect v0.0.0-20220104043353-73e0943537d2/go.mod h1:HBCaDeC1lPdgDeDbhX8XFpy1jqjK0IBG8W5K+xYqA0w= github.com/OneOfOne/xxhash v1.2.2/go.mod h1:HSdplMjZKSmBqAxg5vPj2TmRDmfkzw+cTzAElWljhcU= @@ -1788,7 +1787,7 @@ github.com/hexops/gotextdiff v1.0.3 h1:gitA9+qJrrTCsiCl7+kh75nPqQt1cx4ZkudSTLoUq github.com/hexops/gotextdiff v1.0.3/go.mod h1:pSWU5MAI3yDq+fZBTazCSJysOMbxWL1BSow5/V2vxeg= github.com/hexops/valast v1.4.0 h1:sFzyxPDP0riFQUzSBXTCCrAbbIndHPWMndxuEjXdZlc= github.com/hexops/valast v1.4.0/go.mod h1:uVjKZ0smVuYlgCSPz9NRi5A04sl7lp6GtFWsROKDgEs= -github.com/hinshun/vt10x v0.0.0-20180616224451-1954e6464174/go.mod h1:DqJ97dSdRW1W22yXSB90986pcOyQ7r45iio1KN2ez1A= +github.com/hinshun/vt10x v0.0.0-20220119200601-820417d04eec/go.mod h1:Q48J4R4DvxnHolD5P8pOtXigYlRuPLGl6moFx3ulM68= github.com/hinshun/vt10x v0.0.0-20220301184237-5011da428d02 h1:AgcIVYPa6XJnU3phs104wLj8l5GEththEw6+F79YsIY= github.com/hinshun/vt10x v0.0.0-20220301184237-5011da428d02/go.mod h1:Q48J4R4DvxnHolD5P8pOtXigYlRuPLGl6moFx3ulM68= github.com/hpcloud/tail v1.0.0/go.mod h1:ab1qPbhIpdTxEkNHXyeSf5vhxWSCs/tWer42PpOxQnU= @@ -1947,7 +1946,6 @@ github.com/kr/pretty v0.3.0/go.mod h1:640gp4NfQd8pI5XOwp5fnNeVWj67G7CFk/SaSQn7NB github.com/kr/pretty v0.3.1 h1:flRD4NNwYAUpkphVc1HcthR4KEIFJ65n8Mw5qdRn3LE= github.com/kr/pretty v0.3.1/go.mod h1:hoEshYVHaxMs3cyo3Yncou5ZscifuDolrwPKZanG3xk= github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= -github.com/kr/pty v1.1.4/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= github.com/kr/pty v1.1.5/go.mod h1:9r2w37qlBe7rQ6e1fg1S/9xpWHSnaqNdHD3WcMdbPDA= github.com/kr/pty v1.1.8/go.mod h1:O1sed60cT9XZ5uDucP5qwvh+TE3NnUj51EiZO/lmSfw= github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI= @@ -2343,16 +2341,16 @@ github.com/pulumi/pulumi-java/pkg v0.9.6 h1:UJrOAsYHRchwb4QlfI9Q224qg1TOI3rIsI6D github.com/pulumi/pulumi-java/pkg v0.9.6/go.mod h1:c6rSw/+q4O0IImgJ9axxoC6QesbPYWBaG5gimbHouUQ= github.com/pulumi/pulumi-terraform-bridge/testing v0.0.1 h1:SCg1gjfY9N4yn8U8peIUYATifjoDABkyR7H9lmefsfc= github.com/pulumi/pulumi-terraform-bridge/testing v0.0.1/go.mod h1:7OeUPH8rpt5ipyj9EFcnXpuzQ8SHL0dyqdfa8nOacdk= -github.com/pulumi/pulumi-terraform-bridge/v3 v3.60.1 h1:+W2JHLi4y+G57jLPLJbDLv1xvm/9L2NO0gWXrtR8MDM= -github.com/pulumi/pulumi-terraform-bridge/v3 v3.60.1/go.mod h1:o0Vfch2UXtHOnGYpNElzGg4htT6B8X8hS9fa5AguP7g= +github.com/pulumi/pulumi-terraform-bridge/v3 v3.61.0 h1:ryRfPhwm78Ni+U1uaAsITAI6dD2tLtb3suhJXGWb9cM= +github.com/pulumi/pulumi-terraform-bridge/v3 v3.61.0/go.mod h1:k6iVhOWqTkMWxMdd7njoSkPXBJC/T2Rs44Iif3yDgXw= github.com/pulumi/pulumi-terraform-bridge/x/muxer v0.0.4 h1:rIzMmtcVpPX8ynaz6/nW5AHNY63DiNfCohqmxWvMpM4= github.com/pulumi/pulumi-terraform-bridge/x/muxer v0.0.4/go.mod h1:Kt8RIZWa/N8rW3+0g6NrqCBmF3o+HuIhFaZpssEkG6w= github.com/pulumi/pulumi-yaml v1.2.2 h1:W6BeUBLhDrJ2GSU0em1AUVelG9PBI4ABY61DdhJOO3E= github.com/pulumi/pulumi-yaml v1.2.2/go.mod h1:EgakC7b/4+VBNnlgM1RZIea2gUstV8s/7bdFJZt0P64= -github.com/pulumi/pulumi/pkg/v3 v3.81.0 h1:6rf2farQLszi8inHCu9YdJtDvK0fqNguix51b3FEDRQ= -github.com/pulumi/pulumi/pkg/v3 v3.81.0/go.mod h1:URXozJB/abhxb2yOwa5ywHjT2dqt9SmAX2iLkdcOOl0= -github.com/pulumi/pulumi/sdk/v3 v3.81.0 h1:d711GUpjZK1beNpIKFReC1LBmHXYE684sYWntJWQ4Q0= -github.com/pulumi/pulumi/sdk/v3 v3.81.0/go.mod h1:RMilNNVMlmK1h4Nl/qylb9vzbgh4F3mufZoUOnPy98o= +github.com/pulumi/pulumi/pkg/v3 v3.86.0 h1:G4spuT89ZN8lSxT9WkMF/JaP7n+wu7ubEly7Yy8uza0= +github.com/pulumi/pulumi/pkg/v3 v3.86.0/go.mod h1:Qs55gPhUwM/Dua3VRtHXLLlpY8uEe+llDBIZc+1pvHM= +github.com/pulumi/pulumi/sdk/v3 v3.86.0 h1:Cxg0rGdvMt9GqGvesFTj8+WaO/ihmALYlQf4zm1GzFw= +github.com/pulumi/pulumi/sdk/v3 v3.86.0/go.mod h1:vexSGJ5L834l3T7Fo/KpdywjGAgZ2CvChuqGFl+ebS0= github.com/pulumi/schema-tools v0.1.2 h1:Fd9xvUjgck4NA+7/jSk7InqCUT4Kj940+EcnbQKpfZo= github.com/pulumi/schema-tools v0.1.2/go.mod h1:62lgj52Tzq11eqWTIaKd+EVyYAu5dEcDJxMhTjvMO/k= github.com/pulumi/terraform-diff-reader v0.0.2 h1:kTE4nEXU3/SYXESvAIem+wyHMI3abqkI3OhJ0G04LLI= @@ -2488,7 +2486,6 @@ github.com/stretchr/objx v0.4.0/go.mod h1:YvHI0jy2hoMjB+UWwv71VJQ9isScKT/TqJzVSS github.com/stretchr/objx v0.5.0 h1:1zr/of2m5FGMsad5YfcqgdqdWrIhu+EBEJRhR1U7z/c= github.com/stretchr/objx v0.5.0/go.mod h1:Yh+to48EsGEfYuaHDzXPcE3xhTkx73EhmCGUpEOglKo= github.com/stretchr/testify v0.0.0-20180303142811-b89eecf5ca5d/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= -github.com/stretchr/testify v1.2.1/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= github.com/stretchr/testify v1.2.2/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI= github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4= @@ -2733,7 +2730,6 @@ golang.org/x/crypto v0.0.0-20190411191339-88737f569e3a/go.mod h1:WFFai1msRO1wXaE golang.org/x/crypto v0.0.0-20190422162423-af44ce270edf/go.mod h1:WFFai1msRO1wXaEeE5yQxYXgSfI8pQAWXbQop6sCtWE= golang.org/x/crypto v0.0.0-20190426145343-a29dc8fdc734/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190510104115-cbcb75029529/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= -golang.org/x/crypto v0.0.0-20190530122614-20be4c3c3ed5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190605123033-f99c8df09eb5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190611184440-5c40567a22f8/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190701094942-4def268fd1a4/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= @@ -3029,7 +3025,6 @@ golang.org/x/sys v0.0.0-20190502175342-a43fa875dd82/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20190507160741-ecd444e8653b/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190514135907-3a4b5fb9f71f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190522044717-8097e1b27ff5/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20190530182044-ad28b68e88f1/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190531175056-4c3a928424d2/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190602015325-4c4f7f33c9ed/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190606165138-5da285871e9c/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= diff --git a/sdk/dotnet/Config/Config.cs b/sdk/dotnet/Config/Config.cs index 7dcd7bab..fb6083a6 100644 --- a/sdk/dotnet/Config/Config.cs +++ b/sdk/dotnet/Config/Config.cs @@ -8,7 +8,7 @@ namespace Pulumi.Databricks { public static class Config { - [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "IDE1006", Justification = + [global::System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "IDE1006", Justification = "Double underscore prefix used to avoid conflicts with variable names.")] private sealed class __Value { diff --git a/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageArgs.cs b/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageArgs.cs index c8a8f446..a258ad52 100644 --- a/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageArgs.cs +++ b/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageArgs.cs @@ -16,7 +16,7 @@ public sealed class JobJobClusterNewClusterDockerImageArgs : global::Pulumi.Reso public Input? BasicAuth { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageGetArgs.cs b/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageGetArgs.cs index 802d6fda..a7cd3360 100644 --- a/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageGetArgs.cs +++ b/sdk/dotnet/Inputs/JobJobClusterNewClusterDockerImageGetArgs.cs @@ -16,7 +16,7 @@ public sealed class JobJobClusterNewClusterDockerImageGetArgs : global::Pulumi.R public Input? BasicAuth { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobNewClusterDockerImageArgs.cs b/sdk/dotnet/Inputs/JobNewClusterDockerImageArgs.cs index 1b4477eb..4e1590d0 100644 --- a/sdk/dotnet/Inputs/JobNewClusterDockerImageArgs.cs +++ b/sdk/dotnet/Inputs/JobNewClusterDockerImageArgs.cs @@ -16,7 +16,7 @@ public sealed class JobNewClusterDockerImageArgs : global::Pulumi.ResourceArgs public Input? BasicAuth { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobNewClusterDockerImageGetArgs.cs b/sdk/dotnet/Inputs/JobNewClusterDockerImageGetArgs.cs index f73f5ad1..6194659c 100644 --- a/sdk/dotnet/Inputs/JobNewClusterDockerImageGetArgs.cs +++ b/sdk/dotnet/Inputs/JobNewClusterDockerImageGetArgs.cs @@ -16,7 +16,7 @@ public sealed class JobNewClusterDockerImageGetArgs : global::Pulumi.ResourceArg public Input? BasicAuth { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageArgs.cs b/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageArgs.cs index 204c934c..34ab5e99 100644 --- a/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageArgs.cs @@ -16,7 +16,7 @@ public sealed class JobTaskNewClusterDockerImageArgs : global::Pulumi.ResourceAr public Input? BasicAuth { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageGetArgs.cs b/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageGetArgs.cs index 5415036f..1ca2b2b9 100644 --- a/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTaskNewClusterDockerImageGetArgs.cs @@ -16,7 +16,7 @@ public sealed class JobTaskNewClusterDockerImageGetArgs : global::Pulumi.Resourc public Input? BasicAuth { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobTriggerFileArrivalArgs.cs b/sdk/dotnet/Inputs/JobTriggerFileArrivalArgs.cs index b458ad7f..5557df11 100644 --- a/sdk/dotnet/Inputs/JobTriggerFileArrivalArgs.cs +++ b/sdk/dotnet/Inputs/JobTriggerFileArrivalArgs.cs @@ -19,7 +19,7 @@ public sealed class JobTriggerFileArrivalArgs : global::Pulumi.ResourceArgs public Input? MinTimeBetweenTriggersSeconds { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobTriggerFileArrivalGetArgs.cs b/sdk/dotnet/Inputs/JobTriggerFileArrivalGetArgs.cs index b6012b6c..de1443f5 100644 --- a/sdk/dotnet/Inputs/JobTriggerFileArrivalGetArgs.cs +++ b/sdk/dotnet/Inputs/JobTriggerFileArrivalGetArgs.cs @@ -19,7 +19,7 @@ public sealed class JobTriggerFileArrivalGetArgs : global::Pulumi.ResourceArgs public Input? MinTimeBetweenTriggersSeconds { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url", required: true)] public Input Url { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.cs index 12e35696..2acbd2c4 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnDurationWarningThresholdExceededArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededGetArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededGetArgs.cs index 74d782f5..1c46ea3b 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededGetArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededGetArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnDurationWarningThresholdExceededGetArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureArgs.cs index 4f8598a6..5e7d1cb7 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnFailureArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureGetArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureGetArgs.cs index 66b72263..3cec4c61 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureGetArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnFailureGetArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnFailureGetArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartArgs.cs index 10e377d7..110e7429 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnStartArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartGetArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartGetArgs.cs index 7a5a9bca..2b223f2e 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartGetArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnStartGetArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnStartGetArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessArgs.cs index 08bef572..385c2487 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnSuccessArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessGetArgs.cs b/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessGetArgs.cs index 3232c08b..7d87fcc3 100644 --- a/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessGetArgs.cs +++ b/sdk/dotnet/Inputs/JobWebhookNotificationsOnSuccessGetArgs.cs @@ -13,9 +13,7 @@ namespace Pulumi.Databricks.Inputs public sealed class JobWebhookNotificationsOnSuccessGetArgs : global::Pulumi.ResourceArgs { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// [Input("id", required: true)] public Input Id { get; set; } = null!; diff --git a/sdk/dotnet/Inputs/ShareObjectPartitionArgs.cs b/sdk/dotnet/Inputs/ShareObjectPartitionArgs.cs index 63c505c2..dbb26395 100644 --- a/sdk/dotnet/Inputs/ShareObjectPartitionArgs.cs +++ b/sdk/dotnet/Inputs/ShareObjectPartitionArgs.cs @@ -14,6 +14,10 @@ public sealed class ShareObjectPartitionArgs : global::Pulumi.ResourceArgs { [Input("values", required: true)] private InputList? _values; + + /// + /// The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + /// public InputList Values { get => _values ?? (_values = new InputList()); diff --git a/sdk/dotnet/Inputs/ShareObjectPartitionGetArgs.cs b/sdk/dotnet/Inputs/ShareObjectPartitionGetArgs.cs index c9b33af7..027a9228 100644 --- a/sdk/dotnet/Inputs/ShareObjectPartitionGetArgs.cs +++ b/sdk/dotnet/Inputs/ShareObjectPartitionGetArgs.cs @@ -14,6 +14,10 @@ public sealed class ShareObjectPartitionGetArgs : global::Pulumi.ResourceArgs { [Input("values", required: true)] private InputList? _values; + + /// + /// The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + /// public InputList Values { get => _values ?? (_values = new InputList()); diff --git a/sdk/dotnet/Job.cs b/sdk/dotnet/Job.cs index 85b7ebb5..dfb49d2b 100644 --- a/sdk/dotnet/Job.cs +++ b/sdk/dotnet/Job.cs @@ -180,7 +180,7 @@ public partial class Job : global::Pulumi.CustomResource public Output Trigger { get; private set; } = null!; /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Output("url")] public Output Url { get; private set; } = null!; @@ -632,7 +632,7 @@ public InputList Tasks public Input? Trigger { get; set; } /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// [Input("url")] public Input? Url { get; set; } diff --git a/sdk/dotnet/Outputs/JobJobClusterNewClusterDockerImage.cs b/sdk/dotnet/Outputs/JobJobClusterNewClusterDockerImage.cs index 4f594d37..7562887d 100644 --- a/sdk/dotnet/Outputs/JobJobClusterNewClusterDockerImage.cs +++ b/sdk/dotnet/Outputs/JobJobClusterNewClusterDockerImage.cs @@ -15,7 +15,7 @@ public sealed class JobJobClusterNewClusterDockerImage { public readonly Outputs.JobJobClusterNewClusterDockerImageBasicAuth? BasicAuth; /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// public readonly string Url; diff --git a/sdk/dotnet/Outputs/JobNewClusterDockerImage.cs b/sdk/dotnet/Outputs/JobNewClusterDockerImage.cs index 619e5525..ece6d527 100644 --- a/sdk/dotnet/Outputs/JobNewClusterDockerImage.cs +++ b/sdk/dotnet/Outputs/JobNewClusterDockerImage.cs @@ -15,7 +15,7 @@ public sealed class JobNewClusterDockerImage { public readonly Outputs.JobNewClusterDockerImageBasicAuth? BasicAuth; /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// public readonly string Url; diff --git a/sdk/dotnet/Outputs/JobTaskNewClusterDockerImage.cs b/sdk/dotnet/Outputs/JobTaskNewClusterDockerImage.cs index d1618206..9d77855a 100644 --- a/sdk/dotnet/Outputs/JobTaskNewClusterDockerImage.cs +++ b/sdk/dotnet/Outputs/JobTaskNewClusterDockerImage.cs @@ -15,7 +15,7 @@ public sealed class JobTaskNewClusterDockerImage { public readonly Outputs.JobTaskNewClusterDockerImageBasicAuth? BasicAuth; /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// public readonly string Url; diff --git a/sdk/dotnet/Outputs/JobTriggerFileArrival.cs b/sdk/dotnet/Outputs/JobTriggerFileArrival.cs index af34d95c..65343270 100644 --- a/sdk/dotnet/Outputs/JobTriggerFileArrival.cs +++ b/sdk/dotnet/Outputs/JobTriggerFileArrival.cs @@ -18,7 +18,7 @@ public sealed class JobTriggerFileArrival /// public readonly int? MinTimeBetweenTriggersSeconds; /// - /// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + /// URL of the job on the given workspace /// public readonly string Url; /// diff --git a/sdk/dotnet/Outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.cs b/sdk/dotnet/Outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.cs index cff76767..e91412d7 100644 --- a/sdk/dotnet/Outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.cs +++ b/sdk/dotnet/Outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.cs @@ -14,9 +14,7 @@ namespace Pulumi.Databricks.Outputs public sealed class JobWebhookNotificationsOnDurationWarningThresholdExceeded { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// public readonly string Id; diff --git a/sdk/dotnet/Outputs/JobWebhookNotificationsOnFailure.cs b/sdk/dotnet/Outputs/JobWebhookNotificationsOnFailure.cs index 352bc05d..437e426d 100644 --- a/sdk/dotnet/Outputs/JobWebhookNotificationsOnFailure.cs +++ b/sdk/dotnet/Outputs/JobWebhookNotificationsOnFailure.cs @@ -14,9 +14,7 @@ namespace Pulumi.Databricks.Outputs public sealed class JobWebhookNotificationsOnFailure { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// public readonly string Id; diff --git a/sdk/dotnet/Outputs/JobWebhookNotificationsOnStart.cs b/sdk/dotnet/Outputs/JobWebhookNotificationsOnStart.cs index eece8f12..99c02255 100644 --- a/sdk/dotnet/Outputs/JobWebhookNotificationsOnStart.cs +++ b/sdk/dotnet/Outputs/JobWebhookNotificationsOnStart.cs @@ -14,9 +14,7 @@ namespace Pulumi.Databricks.Outputs public sealed class JobWebhookNotificationsOnStart { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// public readonly string Id; diff --git a/sdk/dotnet/Outputs/JobWebhookNotificationsOnSuccess.cs b/sdk/dotnet/Outputs/JobWebhookNotificationsOnSuccess.cs index d0938d29..aed518eb 100644 --- a/sdk/dotnet/Outputs/JobWebhookNotificationsOnSuccess.cs +++ b/sdk/dotnet/Outputs/JobWebhookNotificationsOnSuccess.cs @@ -14,9 +14,7 @@ namespace Pulumi.Databricks.Outputs public sealed class JobWebhookNotificationsOnSuccess { /// - /// ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - /// - /// > **Note** The following configuration blocks can be standalone or nested inside a `task` block + /// ID of the job /// public readonly string Id; diff --git a/sdk/dotnet/Outputs/ShareObjectPartition.cs b/sdk/dotnet/Outputs/ShareObjectPartition.cs index ba41cf18..92f826da 100644 --- a/sdk/dotnet/Outputs/ShareObjectPartition.cs +++ b/sdk/dotnet/Outputs/ShareObjectPartition.cs @@ -13,6 +13,9 @@ namespace Pulumi.Databricks.Outputs [OutputType] public sealed class ShareObjectPartition { + /// + /// The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + /// public readonly ImmutableArray Values; [OutputConstructor] diff --git a/sdk/go.mod b/sdk/go.mod index 694b407b..46489901 100644 --- a/sdk/go.mod +++ b/sdk/go.mod @@ -4,7 +4,7 @@ go 1.21 require ( github.com/blang/semver v3.5.1+incompatible - github.com/pulumi/pulumi/sdk/v3 v3.81.0 + github.com/pulumi/pulumi/sdk/v3 v3.86.0 ) require ( diff --git a/sdk/go.sum b/sdk/go.sum index 22c9c047..8b6acb1f 100644 --- a/sdk/go.sum +++ b/sdk/go.sum @@ -143,8 +143,8 @@ github.com/pkg/term v1.1.0 h1:xIAAdCMh3QIAy+5FrE8Ad8XoDhEU4ufwbaSozViP9kk= github.com/pkg/term v1.1.0/go.mod h1:E25nymQcrSllhX42Ok8MRm1+hyBdHY0dCeiKZ9jpNGw= github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= -github.com/pulumi/pulumi/sdk/v3 v3.81.0 h1:d711GUpjZK1beNpIKFReC1LBmHXYE684sYWntJWQ4Q0= -github.com/pulumi/pulumi/sdk/v3 v3.81.0/go.mod h1:RMilNNVMlmK1h4Nl/qylb9vzbgh4F3mufZoUOnPy98o= +github.com/pulumi/pulumi/sdk/v3 v3.86.0 h1:Cxg0rGdvMt9GqGvesFTj8+WaO/ihmALYlQf4zm1GzFw= +github.com/pulumi/pulumi/sdk/v3 v3.86.0/go.mod h1:vexSGJ5L834l3T7Fo/KpdywjGAgZ2CvChuqGFl+ebS0= github.com/rivo/uniseg v0.1.0/go.mod h1:J6wj4VEh+S6ZtnVlnTBMWIodfgj8LQOQFoIToxlJtxc= github.com/rivo/uniseg v0.2.0/go.mod h1:J6wj4VEh+S6ZtnVlnTBMWIodfgj8LQOQFoIToxlJtxc= github.com/rivo/uniseg v0.4.4 h1:8TfxU8dW6PdqD27gjM8MVNuicgxIjxpm4K7x4jp8sis= diff --git a/sdk/go/databricks/job.go b/sdk/go/databricks/job.go index e4eb87ee..5fe7de55 100644 --- a/sdk/go/databricks/job.go +++ b/sdk/go/databricks/job.go @@ -90,7 +90,7 @@ type Job struct { // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. TimeoutSeconds pulumi.IntPtrOutput `pulumi:"timeoutSeconds"` Trigger JobTriggerPtrOutput `pulumi:"trigger"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url pulumi.StringOutput `pulumi:"url"` // (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this job begins, completes and fails. The default behavior is to not send any notifications. This field is a block and is documented below. WebhookNotifications JobWebhookNotificationsPtrOutput `pulumi:"webhookNotifications"` @@ -192,7 +192,7 @@ type jobState struct { // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. TimeoutSeconds *int `pulumi:"timeoutSeconds"` Trigger *JobTrigger `pulumi:"trigger"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url *string `pulumi:"url"` // (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this job begins, completes and fails. The default behavior is to not send any notifications. This field is a block and is documented below. WebhookNotifications *JobWebhookNotifications `pulumi:"webhookNotifications"` @@ -265,7 +265,7 @@ type JobState struct { // (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. TimeoutSeconds pulumi.IntPtrInput Trigger JobTriggerPtrInput - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url pulumi.StringPtrInput // (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this job begins, completes and fails. The default behavior is to not send any notifications. This field is a block and is documented below. WebhookNotifications JobWebhookNotificationsPtrInput @@ -700,7 +700,7 @@ func (o JobOutput) Trigger() JobTriggerPtrOutput { return o.ApplyT(func(v *Job) JobTriggerPtrOutput { return v.Trigger }).(JobTriggerPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobOutput) Url() pulumi.StringOutput { return o.ApplyT(func(v *Job) pulumi.StringOutput { return v.Url }).(pulumi.StringOutput) } diff --git a/sdk/go/databricks/pulumiTypes.go b/sdk/go/databricks/pulumiTypes.go index 7c6c658e..f6f30e28 100644 --- a/sdk/go/databricks/pulumiTypes.go +++ b/sdk/go/databricks/pulumiTypes.go @@ -10662,7 +10662,7 @@ func (o JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoOutput) Serv type JobJobClusterNewClusterDockerImage struct { BasicAuth *JobJobClusterNewClusterDockerImageBasicAuth `pulumi:"basicAuth"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url string `pulumi:"url"` } @@ -10679,7 +10679,7 @@ type JobJobClusterNewClusterDockerImageInput interface { type JobJobClusterNewClusterDockerImageArgs struct { BasicAuth JobJobClusterNewClusterDockerImageBasicAuthPtrInput `pulumi:"basicAuth"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url pulumi.StringInput `pulumi:"url"` } @@ -10784,7 +10784,7 @@ func (o JobJobClusterNewClusterDockerImageOutput) BasicAuth() JobJobClusterNewCl }).(JobJobClusterNewClusterDockerImageBasicAuthPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobJobClusterNewClusterDockerImageOutput) Url() pulumi.StringOutput { return o.ApplyT(func(v JobJobClusterNewClusterDockerImage) string { return v.Url }).(pulumi.StringOutput) } @@ -10828,7 +10828,7 @@ func (o JobJobClusterNewClusterDockerImagePtrOutput) BasicAuth() JobJobClusterNe }).(JobJobClusterNewClusterDockerImageBasicAuthPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobJobClusterNewClusterDockerImagePtrOutput) Url() pulumi.StringPtrOutput { return o.ApplyT(func(v *JobJobClusterNewClusterDockerImage) *string { if v == nil { @@ -15779,7 +15779,7 @@ func (o JobNewClusterClusterMountInfoNetworkFilesystemInfoOutput) ServerAddress( type JobNewClusterDockerImage struct { BasicAuth *JobNewClusterDockerImageBasicAuth `pulumi:"basicAuth"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url string `pulumi:"url"` } @@ -15796,7 +15796,7 @@ type JobNewClusterDockerImageInput interface { type JobNewClusterDockerImageArgs struct { BasicAuth JobNewClusterDockerImageBasicAuthPtrInput `pulumi:"basicAuth"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url pulumi.StringInput `pulumi:"url"` } @@ -15899,7 +15899,7 @@ func (o JobNewClusterDockerImageOutput) BasicAuth() JobNewClusterDockerImageBasi return o.ApplyT(func(v JobNewClusterDockerImage) *JobNewClusterDockerImageBasicAuth { return v.BasicAuth }).(JobNewClusterDockerImageBasicAuthPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobNewClusterDockerImageOutput) Url() pulumi.StringOutput { return o.ApplyT(func(v JobNewClusterDockerImage) string { return v.Url }).(pulumi.StringOutput) } @@ -15943,7 +15943,7 @@ func (o JobNewClusterDockerImagePtrOutput) BasicAuth() JobNewClusterDockerImageB }).(JobNewClusterDockerImageBasicAuthPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobNewClusterDockerImagePtrOutput) Url() pulumi.StringPtrOutput { return o.ApplyT(func(v *JobNewClusterDockerImage) *string { if v == nil { @@ -24615,7 +24615,7 @@ func (o JobTaskNewClusterClusterMountInfoNetworkFilesystemInfoOutput) ServerAddr type JobTaskNewClusterDockerImage struct { BasicAuth *JobTaskNewClusterDockerImageBasicAuth `pulumi:"basicAuth"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url string `pulumi:"url"` } @@ -24632,7 +24632,7 @@ type JobTaskNewClusterDockerImageInput interface { type JobTaskNewClusterDockerImageArgs struct { BasicAuth JobTaskNewClusterDockerImageBasicAuthPtrInput `pulumi:"basicAuth"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url pulumi.StringInput `pulumi:"url"` } @@ -24735,7 +24735,7 @@ func (o JobTaskNewClusterDockerImageOutput) BasicAuth() JobTaskNewClusterDockerI return o.ApplyT(func(v JobTaskNewClusterDockerImage) *JobTaskNewClusterDockerImageBasicAuth { return v.BasicAuth }).(JobTaskNewClusterDockerImageBasicAuthPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobTaskNewClusterDockerImageOutput) Url() pulumi.StringOutput { return o.ApplyT(func(v JobTaskNewClusterDockerImage) string { return v.Url }).(pulumi.StringOutput) } @@ -24779,7 +24779,7 @@ func (o JobTaskNewClusterDockerImagePtrOutput) BasicAuth() JobTaskNewClusterDock }).(JobTaskNewClusterDockerImageBasicAuthPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobTaskNewClusterDockerImagePtrOutput) Url() pulumi.StringPtrOutput { return o.ApplyT(func(v *JobTaskNewClusterDockerImage) *string { if v == nil { @@ -30270,7 +30270,7 @@ func (o JobTriggerPtrOutput) PauseStatus() pulumi.StringPtrOutput { type JobTriggerFileArrival struct { // If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. MinTimeBetweenTriggersSeconds *int `pulumi:"minTimeBetweenTriggersSeconds"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url string `pulumi:"url"` // If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. WaitAfterLastChangeSeconds *int `pulumi:"waitAfterLastChangeSeconds"` @@ -30290,7 +30290,7 @@ type JobTriggerFileArrivalInput interface { type JobTriggerFileArrivalArgs struct { // If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. MinTimeBetweenTriggersSeconds pulumi.IntPtrInput `pulumi:"minTimeBetweenTriggersSeconds"` - // string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + // URL of the job on the given workspace Url pulumi.StringInput `pulumi:"url"` // If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. WaitAfterLastChangeSeconds pulumi.IntPtrInput `pulumi:"waitAfterLastChangeSeconds"` @@ -30396,7 +30396,7 @@ func (o JobTriggerFileArrivalOutput) MinTimeBetweenTriggersSeconds() pulumi.IntP return o.ApplyT(func(v JobTriggerFileArrival) *int { return v.MinTimeBetweenTriggersSeconds }).(pulumi.IntPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobTriggerFileArrivalOutput) Url() pulumi.StringOutput { return o.ApplyT(func(v JobTriggerFileArrival) string { return v.Url }).(pulumi.StringOutput) } @@ -30446,7 +30446,7 @@ func (o JobTriggerFileArrivalPtrOutput) MinTimeBetweenTriggersSeconds() pulumi.I }).(pulumi.IntPtrOutput) } -// string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). +// URL of the job on the given workspace func (o JobTriggerFileArrivalPtrOutput) Url() pulumi.StringPtrOutput { return o.ApplyT(func(v *JobTriggerFileArrival) *string { if v == nil { @@ -30765,9 +30765,7 @@ func (o JobWebhookNotificationsPtrOutput) OnSuccesses() JobWebhookNotificationsO } type JobWebhookNotificationsOnDurationWarningThresholdExceeded struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id string `pulumi:"id"` } @@ -30783,9 +30781,7 @@ type JobWebhookNotificationsOnDurationWarningThresholdExceededInput interface { } type JobWebhookNotificationsOnDurationWarningThresholdExceededArgs struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id pulumi.StringInput `pulumi:"id"` } @@ -30858,9 +30854,7 @@ func (o JobWebhookNotificationsOnDurationWarningThresholdExceededOutput) ToOutpu } } -// ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. -// -// > **Note** The following configuration blocks can be standalone or nested inside a `task` block +// ID of the job func (o JobWebhookNotificationsOnDurationWarningThresholdExceededOutput) Id() pulumi.StringOutput { return o.ApplyT(func(v JobWebhookNotificationsOnDurationWarningThresholdExceeded) string { return v.Id }).(pulumi.StringOutput) } @@ -30892,9 +30886,7 @@ func (o JobWebhookNotificationsOnDurationWarningThresholdExceededArrayOutput) In } type JobWebhookNotificationsOnFailure struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id string `pulumi:"id"` } @@ -30910,9 +30902,7 @@ type JobWebhookNotificationsOnFailureInput interface { } type JobWebhookNotificationsOnFailureArgs struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id pulumi.StringInput `pulumi:"id"` } @@ -30985,9 +30975,7 @@ func (o JobWebhookNotificationsOnFailureOutput) ToOutput(ctx context.Context) pu } } -// ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. -// -// > **Note** The following configuration blocks can be standalone or nested inside a `task` block +// ID of the job func (o JobWebhookNotificationsOnFailureOutput) Id() pulumi.StringOutput { return o.ApplyT(func(v JobWebhookNotificationsOnFailure) string { return v.Id }).(pulumi.StringOutput) } @@ -31019,9 +31007,7 @@ func (o JobWebhookNotificationsOnFailureArrayOutput) Index(i pulumi.IntInput) Jo } type JobWebhookNotificationsOnStart struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id string `pulumi:"id"` } @@ -31037,9 +31023,7 @@ type JobWebhookNotificationsOnStartInput interface { } type JobWebhookNotificationsOnStartArgs struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id pulumi.StringInput `pulumi:"id"` } @@ -31112,9 +31096,7 @@ func (o JobWebhookNotificationsOnStartOutput) ToOutput(ctx context.Context) pulu } } -// ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. -// -// > **Note** The following configuration blocks can be standalone or nested inside a `task` block +// ID of the job func (o JobWebhookNotificationsOnStartOutput) Id() pulumi.StringOutput { return o.ApplyT(func(v JobWebhookNotificationsOnStart) string { return v.Id }).(pulumi.StringOutput) } @@ -31146,9 +31128,7 @@ func (o JobWebhookNotificationsOnStartArrayOutput) Index(i pulumi.IntInput) JobW } type JobWebhookNotificationsOnSuccess struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id string `pulumi:"id"` } @@ -31164,9 +31144,7 @@ type JobWebhookNotificationsOnSuccessInput interface { } type JobWebhookNotificationsOnSuccessArgs struct { - // ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - // - // > **Note** The following configuration blocks can be standalone or nested inside a `task` block + // ID of the job Id pulumi.StringInput `pulumi:"id"` } @@ -31239,9 +31217,7 @@ func (o JobWebhookNotificationsOnSuccessOutput) ToOutput(ctx context.Context) pu } } -// ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. -// -// > **Note** The following configuration blocks can be standalone or nested inside a `task` block +// ID of the job func (o JobWebhookNotificationsOnSuccessOutput) Id() pulumi.StringOutput { return o.ApplyT(func(v JobWebhookNotificationsOnSuccess) string { return v.Id }).(pulumi.StringOutput) } @@ -42355,6 +42331,7 @@ func (o ShareObjectArrayOutput) Index(i pulumi.IntInput) ShareObjectOutput { } type ShareObjectPartition struct { + // The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipientPropertyKey` can not be set. Values []ShareObjectPartitionValue `pulumi:"values"` } @@ -42370,6 +42347,7 @@ type ShareObjectPartitionInput interface { } type ShareObjectPartitionArgs struct { + // The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipientPropertyKey` can not be set. Values ShareObjectPartitionValueArrayInput `pulumi:"values"` } @@ -42442,6 +42420,7 @@ func (o ShareObjectPartitionOutput) ToOutput(ctx context.Context) pulumix.Output } } +// The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipientPropertyKey` can not be set. func (o ShareObjectPartitionOutput) Values() ShareObjectPartitionValueArrayOutput { return o.ApplyT(func(v ShareObjectPartition) []ShareObjectPartitionValue { return v.Values }).(ShareObjectPartitionValueArrayOutput) } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/Job.java b/sdk/java/src/main/java/com/pulumi/databricks/Job.java index e513da03..64f970cc 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/Job.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/Job.java @@ -432,14 +432,14 @@ public Output> trigger() { return Codegen.optional(this.trigger); } /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace * */ @Export(name="url", type=String.class, parameters={}) private Output url; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public Output url() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterDockerImageArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterDockerImageArgs.java index 1b9c3d7a..31205fa2 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterDockerImageArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobJobClusterNewClusterDockerImageArgs.java @@ -24,14 +24,14 @@ public Optional> basicAu } /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace * */ @Import(name="url", required=true) private Output url; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public Output url() { @@ -73,7 +73,7 @@ public Builder basicAuth(JobJobClusterNewClusterDockerImageBasicAuthArgs basicAu } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * @@ -84,7 +84,7 @@ public Builder url(Output url) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterDockerImageArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterDockerImageArgs.java index 862e7672..46d5bf0c 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterDockerImageArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobNewClusterDockerImageArgs.java @@ -24,14 +24,14 @@ public Optional> basicAuth() { } /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace * */ @Import(name="url", required=true) private Output url; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public Output url() { @@ -73,7 +73,7 @@ public Builder basicAuth(JobNewClusterDockerImageBasicAuthArgs basicAuth) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * @@ -84,7 +84,7 @@ public Builder url(Output url) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobState.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobState.java index 947ed0a3..f240be87 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobState.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobState.java @@ -459,14 +459,14 @@ public Optional> trigger() { } /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace * */ @Import(name="url") private @Nullable Output url; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public Optional> url() { @@ -1122,7 +1122,7 @@ public Builder trigger(JobTriggerArgs trigger) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * @@ -1133,7 +1133,7 @@ public Builder url(@Nullable Output url) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterDockerImageArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterDockerImageArgs.java index 0a023e63..8ac647e5 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterDockerImageArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTaskNewClusterDockerImageArgs.java @@ -24,14 +24,14 @@ public Optional> basicAuth() { } /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace * */ @Import(name="url", required=true) private Output url; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public Output url() { @@ -73,7 +73,7 @@ public Builder basicAuth(JobTaskNewClusterDockerImageBasicAuthArgs basicAuth) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * @@ -84,7 +84,7 @@ public Builder url(Output url) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerFileArrivalArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerFileArrivalArgs.java index c7233e68..74d6aced 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerFileArrivalArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobTriggerFileArrivalArgs.java @@ -32,14 +32,14 @@ public Optional> minTimeBetweenTriggersSeconds() { } /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace * */ @Import(name="url", required=true) private Output url; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public Output url() { @@ -109,7 +109,7 @@ public Builder minTimeBetweenTriggersSeconds(Integer minTimeBetweenTriggersSecon } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * @@ -120,7 +120,7 @@ public Builder url(Output url) { } /** - * @param url string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @param url URL of the job on the given workspace * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.java index 8370f9b7..813bc4ba 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnDurationWarningThresholdExceededArgs.java @@ -14,18 +14,14 @@ public final class JobWebhookNotificationsOnDurationWarningThresholdExceededArgs public static final JobWebhookNotificationsOnDurationWarningThresholdExceededArgs Empty = new JobWebhookNotificationsOnDurationWarningThresholdExceededArgs(); /** - * ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job * */ @Import(name="id", required=true) private Output id; /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public Output id() { @@ -57,9 +53,7 @@ public Builder(JobWebhookNotificationsOnDurationWarningThresholdExceededArgs def } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * @@ -70,9 +64,7 @@ public Builder id(Output id) { } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnFailureArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnFailureArgs.java index 67c959d4..b17f6640 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnFailureArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnFailureArgs.java @@ -14,18 +14,14 @@ public final class JobWebhookNotificationsOnFailureArgs extends com.pulumi.resou public static final JobWebhookNotificationsOnFailureArgs Empty = new JobWebhookNotificationsOnFailureArgs(); /** - * ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job * */ @Import(name="id", required=true) private Output id; /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public Output id() { @@ -57,9 +53,7 @@ public Builder(JobWebhookNotificationsOnFailureArgs defaults) { } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * @@ -70,9 +64,7 @@ public Builder id(Output id) { } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnStartArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnStartArgs.java index 3149bf4a..ed6dd8d1 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnStartArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnStartArgs.java @@ -14,18 +14,14 @@ public final class JobWebhookNotificationsOnStartArgs extends com.pulumi.resourc public static final JobWebhookNotificationsOnStartArgs Empty = new JobWebhookNotificationsOnStartArgs(); /** - * ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job * */ @Import(name="id", required=true) private Output id; /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public Output id() { @@ -57,9 +53,7 @@ public Builder(JobWebhookNotificationsOnStartArgs defaults) { } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * @@ -70,9 +64,7 @@ public Builder id(Output id) { } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnSuccessArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnSuccessArgs.java index 7a73c1f6..b52a6180 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnSuccessArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/JobWebhookNotificationsOnSuccessArgs.java @@ -14,18 +14,14 @@ public final class JobWebhookNotificationsOnSuccessArgs extends com.pulumi.resou public static final JobWebhookNotificationsOnSuccessArgs Empty = new JobWebhookNotificationsOnSuccessArgs(); /** - * ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job * */ @Import(name="id", required=true) private Output id; /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public Output id() { @@ -57,9 +53,7 @@ public Builder(JobWebhookNotificationsOnSuccessArgs defaults) { } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * @@ -70,9 +64,7 @@ public Builder id(Output id) { } /** - * @param id ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @param id ID of the job * * @return builder * diff --git a/sdk/java/src/main/java/com/pulumi/databricks/inputs/ShareObjectPartitionArgs.java b/sdk/java/src/main/java/com/pulumi/databricks/inputs/ShareObjectPartitionArgs.java index 257d6b87..9d35a48f 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/inputs/ShareObjectPartitionArgs.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/inputs/ShareObjectPartitionArgs.java @@ -14,9 +14,17 @@ public final class ShareObjectPartitionArgs extends com.pulumi.resources.Resourc public static final ShareObjectPartitionArgs Empty = new ShareObjectPartitionArgs(); + /** + * The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + * + */ @Import(name="values", required=true) private Output> values; + /** + * @return The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + * + */ public Output> values() { return this.values; } @@ -45,15 +53,33 @@ public Builder(ShareObjectPartitionArgs defaults) { $ = new ShareObjectPartitionArgs(Objects.requireNonNull(defaults)); } + /** + * @param values The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + * + * @return builder + * + */ public Builder values(Output> values) { $.values = values; return this; } + /** + * @param values The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + * + * @return builder + * + */ public Builder values(List values) { return values(Output.of(values)); } + /** + * @param values The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + * + * @return builder + * + */ public Builder values(ShareObjectPartitionValueArgs... values) { return values(List.of(values)); } diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewClusterDockerImage.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewClusterDockerImage.java index 05eb500e..b10c7e0c 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewClusterDockerImage.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobJobClusterNewClusterDockerImage.java @@ -14,7 +14,7 @@ public final class JobJobClusterNewClusterDockerImage { private @Nullable JobJobClusterNewClusterDockerImageBasicAuth basicAuth; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ private String url; @@ -24,7 +24,7 @@ public Optional basicAuth() { return Optional.ofNullable(this.basicAuth); } /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public String url() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewClusterDockerImage.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewClusterDockerImage.java index 89c49b00..6ff334bf 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewClusterDockerImage.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobNewClusterDockerImage.java @@ -14,7 +14,7 @@ public final class JobNewClusterDockerImage { private @Nullable JobNewClusterDockerImageBasicAuth basicAuth; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ private String url; @@ -24,7 +24,7 @@ public Optional basicAuth() { return Optional.ofNullable(this.basicAuth); } /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public String url() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewClusterDockerImage.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewClusterDockerImage.java index 3576cafc..db3eda78 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewClusterDockerImage.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTaskNewClusterDockerImage.java @@ -14,7 +14,7 @@ public final class JobTaskNewClusterDockerImage { private @Nullable JobTaskNewClusterDockerImageBasicAuth basicAuth; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ private String url; @@ -24,7 +24,7 @@ public Optional basicAuth() { return Optional.ofNullable(this.basicAuth); } /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public String url() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerFileArrival.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerFileArrival.java index 185903b7..949cfe92 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerFileArrival.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobTriggerFileArrival.java @@ -18,7 +18,7 @@ public final class JobTriggerFileArrival { */ private @Nullable Integer minTimeBetweenTriggersSeconds; /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ private String url; @@ -37,7 +37,7 @@ public Optional minTimeBetweenTriggersSeconds() { return Optional.ofNullable(this.minTimeBetweenTriggersSeconds); } /** - * @return string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * @return URL of the job on the given workspace * */ public String url() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.java index 548c0585..d31266ef 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnDurationWarningThresholdExceeded.java @@ -10,18 +10,14 @@ @CustomType public final class JobWebhookNotificationsOnDurationWarningThresholdExceeded { /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ private String id; private JobWebhookNotificationsOnDurationWarningThresholdExceeded() {} /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public String id() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnFailure.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnFailure.java index 7c27ae6f..bc026f87 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnFailure.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnFailure.java @@ -10,18 +10,14 @@ @CustomType public final class JobWebhookNotificationsOnFailure { /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ private String id; private JobWebhookNotificationsOnFailure() {} /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public String id() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnStart.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnStart.java index bb9fbf90..33810529 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnStart.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnStart.java @@ -10,18 +10,14 @@ @CustomType public final class JobWebhookNotificationsOnStart { /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ private String id; private JobWebhookNotificationsOnStart() {} /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public String id() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnSuccess.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnSuccess.java index f355a675..ef519186 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnSuccess.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/JobWebhookNotificationsOnSuccess.java @@ -10,18 +10,14 @@ @CustomType public final class JobWebhookNotificationsOnSuccess { /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ private String id; private JobWebhookNotificationsOnSuccess() {} /** - * @return ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * @return ID of the job * */ public String id() { diff --git a/sdk/java/src/main/java/com/pulumi/databricks/outputs/ShareObjectPartition.java b/sdk/java/src/main/java/com/pulumi/databricks/outputs/ShareObjectPartition.java index e0f9d861..8cea86c5 100644 --- a/sdk/java/src/main/java/com/pulumi/databricks/outputs/ShareObjectPartition.java +++ b/sdk/java/src/main/java/com/pulumi/databricks/outputs/ShareObjectPartition.java @@ -10,9 +10,17 @@ @CustomType public final class ShareObjectPartition { + /** + * @return The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + * + */ private List values; private ShareObjectPartition() {} + /** + * @return The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + * + */ public List values() { return this.values; } diff --git a/sdk/nodejs/job.ts b/sdk/nodejs/job.ts index abf9870a..c971a9dc 100644 --- a/sdk/nodejs/job.ts +++ b/sdk/nodejs/job.ts @@ -130,7 +130,7 @@ export class Job extends pulumi.CustomResource { public readonly timeoutSeconds!: pulumi.Output; public readonly trigger!: pulumi.Output; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ public /*out*/ readonly url!: pulumi.Output; /** @@ -322,7 +322,7 @@ export interface JobState { timeoutSeconds?: pulumi.Input; trigger?: pulumi.Input; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url?: pulumi.Input; /** diff --git a/sdk/nodejs/types/input.ts b/sdk/nodejs/types/input.ts index 2ca64569..06194b7a 100644 --- a/sdk/nodejs/types/input.ts +++ b/sdk/nodejs/types/input.ts @@ -3152,7 +3152,7 @@ export interface JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfo { export interface JobJobClusterNewClusterDockerImage { basicAuth?: pulumi.Input; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: pulumi.Input; } @@ -3386,7 +3386,7 @@ export interface JobNewClusterClusterMountInfoNetworkFilesystemInfo { export interface JobNewClusterDockerImage { basicAuth?: pulumi.Input; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: pulumi.Input; } @@ -3938,7 +3938,7 @@ export interface JobTaskNewClusterClusterMountInfoNetworkFilesystemInfo { export interface JobTaskNewClusterDockerImage { basicAuth?: pulumi.Input; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: pulumi.Input; } @@ -4311,7 +4311,7 @@ export interface JobTriggerFileArrival { */ minTimeBetweenTriggersSeconds?: pulumi.Input; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: pulumi.Input; /** @@ -4349,36 +4349,28 @@ export interface JobWebhookNotifications { export interface JobWebhookNotificationsOnDurationWarningThresholdExceeded { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: pulumi.Input; } export interface JobWebhookNotificationsOnFailure { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: pulumi.Input; } export interface JobWebhookNotificationsOnStart { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: pulumi.Input; } export interface JobWebhookNotificationsOnSuccess { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: pulumi.Input; } @@ -5000,6 +4992,9 @@ export interface ShareObject { } export interface ShareObjectPartition { + /** + * The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipientPropertyKey` can not be set. + */ values: pulumi.Input[]>; } diff --git a/sdk/nodejs/types/output.ts b/sdk/nodejs/types/output.ts index 0924a15c..57f78db4 100644 --- a/sdk/nodejs/types/output.ts +++ b/sdk/nodejs/types/output.ts @@ -1842,7 +1842,7 @@ export interface JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfo { export interface JobJobClusterNewClusterDockerImage { basicAuth?: outputs.JobJobClusterNewClusterDockerImageBasicAuth; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: string; } @@ -2076,7 +2076,7 @@ export interface JobNewClusterClusterMountInfoNetworkFilesystemInfo { export interface JobNewClusterDockerImage { basicAuth?: outputs.JobNewClusterDockerImageBasicAuth; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: string; } @@ -2628,7 +2628,7 @@ export interface JobTaskNewClusterClusterMountInfoNetworkFilesystemInfo { export interface JobTaskNewClusterDockerImage { basicAuth?: outputs.JobTaskNewClusterDockerImageBasicAuth; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: string; } @@ -3001,7 +3001,7 @@ export interface JobTriggerFileArrival { */ minTimeBetweenTriggersSeconds?: number; /** - * string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + * URL of the job on the given workspace */ url: string; /** @@ -3039,36 +3039,28 @@ export interface JobWebhookNotifications { export interface JobWebhookNotificationsOnDurationWarningThresholdExceeded { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: string; } export interface JobWebhookNotificationsOnFailure { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: string; } export interface JobWebhookNotificationsOnStart { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: string; } export interface JobWebhookNotificationsOnSuccess { /** - * ID of the system notification that is notified when an event defined in `webhookNotifications` is triggered. - * - * > **Note** The following configuration blocks can be standalone or nested inside a `task` block + * ID of the job */ id: string; } @@ -3690,6 +3682,9 @@ export interface ShareObject { } export interface ShareObjectPartition { + /** + * The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipientPropertyKey` can not be set. + */ values: outputs.ShareObjectPartitionValue[]; } diff --git a/sdk/python/pulumi_databricks/_inputs.py b/sdk/python/pulumi_databricks/_inputs.py index f19db595..b71f0435 100644 --- a/sdk/python/pulumi_databricks/_inputs.py +++ b/sdk/python/pulumi_databricks/_inputs.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ @@ -463,9 +463,20 @@ def __init__(__self__, *, * `groups/{groupname}` (also exposed as `acl_principal_id` attribute of `Group` resource). * `servicePrincipals/{applicationId}` (also exposed as `acl_principal_id` attribute of `ServicePrincipal` resource). """ - pulumi.set(__self__, "role", role) + AccessControlRuleSetGrantRuleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + role=role, + principals=principals, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role: pulumi.Input[str], + principals: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role", role) if principals is not None: - pulumi.set(__self__, "principals", principals) + _setter("principals", principals) @property @pulumi.getter @@ -504,10 +515,21 @@ class ClusterAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[pulumi.Input[int]] = None, min_workers: Optional[pulumi.Input[int]] = None): + ClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[pulumi.Input[int]] = None, + min_workers: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -539,22 +561,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[pulumi.Input[str]] = None, spot_bid_price_percent: Optional[pulumi.Input[int]] = None, zone_id: Optional[pulumi.Input[str]] = None): + ClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + ebs_volume_count: Optional[pulumi.Input[int]] = None, + ebs_volume_size: Optional[pulumi.Input[int]] = None, + ebs_volume_type: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + spot_bid_price_percent: Optional[pulumi.Input[int]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -635,12 +680,25 @@ def __init__(__self__, *, availability: Optional[pulumi.Input[str]] = None, first_on_demand: Optional[pulumi.Input[int]] = None, spot_bid_max_price: Optional[pulumi.Input[float]] = None): + ClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + spot_bid_max_price: Optional[pulumi.Input[float]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -675,10 +733,21 @@ class ClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional[pulumi.Input['ClusterClusterLogConfDbfsArgs']] = None, s3: Optional[pulumi.Input['ClusterClusterLogConfS3Args']] = None): + ClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional[pulumi.Input['ClusterClusterLogConfDbfsArgs']] = None, + s3: Optional[pulumi.Input['ClusterClusterLogConfS3Args']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -703,7 +772,16 @@ def s3(self, value: Optional[pulumi.Input['ClusterClusterLogConfS3Args']]): class ClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + ClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -725,19 +803,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + ClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -809,10 +908,23 @@ def __init__(__self__, *, local_mount_dir_path: pulumi.Input[str], network_filesystem_info: pulumi.Input['ClusterClusterMountInfoNetworkFilesystemInfoArgs'], remote_mount_dir_path: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + ClusterClusterMountInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: pulumi.Input[str], + network_filesystem_info: pulumi.Input['ClusterClusterMountInfoNetworkFilesystemInfoArgs'], + remote_mount_dir_path: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -847,9 +959,20 @@ class ClusterClusterMountInfoNetworkFilesystemInfoArgs: def __init__(__self__, *, server_address: pulumi.Input[str], mount_options: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "server_address", server_address) + ClusterClusterMountInfoNetworkFilesystemInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: pulumi.Input[str], + mount_options: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -875,9 +998,20 @@ class ClusterDockerImageArgs: def __init__(__self__, *, url: pulumi.Input[str], basic_auth: Optional[pulumi.Input['ClusterDockerImageBasicAuthArgs']] = None): - pulumi.set(__self__, "url", url) + ClusterDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + basic_auth: Optional[pulumi.Input['ClusterDockerImageBasicAuthArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -903,8 +1037,19 @@ class ClusterDockerImageBasicAuthArgs: def __init__(__self__, *, password: pulumi.Input[str], username: pulumi.Input[str]): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + ClusterDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: pulumi.Input[str], + username: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -934,21 +1079,40 @@ def __init__(__self__, *, local_ssd_count: Optional[pulumi.Input[int]] = None, use_preemptible_executors: Optional[pulumi.Input[bool]] = None, zone_id: Optional[pulumi.Input[str]] = None): + ClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + boot_disk_size: Optional[pulumi.Input[int]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + local_ssd_count: Optional[pulumi.Input[int]] = None, + use_preemptible_executors: Optional[pulumi.Input[bool]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: warnings.warn("""Please use 'availability' instead.""", DeprecationWarning) pulumi.log.warn("""use_preemptible_executors is deprecated: Please use 'availability' instead.""") if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -1018,23 +1182,44 @@ def __init__(__self__, *, s3: Optional[pulumi.Input['ClusterInitScriptS3Args']] = None, volumes: Optional[pulumi.Input['ClusterInitScriptVolumesArgs']] = None, workspace: Optional[pulumi.Input['ClusterInitScriptWorkspaceArgs']] = None): + ClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional[pulumi.Input['ClusterInitScriptAbfssArgs']] = None, + dbfs: Optional[pulumi.Input['ClusterInitScriptDbfsArgs']] = None, + file: Optional[pulumi.Input['ClusterInitScriptFileArgs']] = None, + gcs: Optional[pulumi.Input['ClusterInitScriptGcsArgs']] = None, + s3: Optional[pulumi.Input['ClusterInitScriptS3Args']] = None, + volumes: Optional[pulumi.Input['ClusterInitScriptVolumesArgs']] = None, + workspace: Optional[pulumi.Input['ClusterInitScriptWorkspaceArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: warnings.warn("""For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""", DeprecationWarning) pulumi.log.warn("""dbfs is deprecated: For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""") if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -1107,8 +1292,17 @@ def workspace(self, value: Optional[pulumi.Input['ClusterInitScriptWorkspaceArgs class ClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + ClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1124,7 +1318,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class ClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + ClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -1140,8 +1343,17 @@ def destination(self, value: pulumi.Input[str]): class ClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + ClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1157,8 +1369,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class ClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + ClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1180,19 +1401,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + ClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -1262,8 +1504,17 @@ def region(self, value: Optional[pulumi.Input[str]]): class ClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + ClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1279,8 +1530,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class ClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + ClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1301,18 +1561,37 @@ def __init__(__self__, *, maven: Optional[pulumi.Input['ClusterLibraryMavenArgs']] = None, pypi: Optional[pulumi.Input['ClusterLibraryPypiArgs']] = None, whl: Optional[pulumi.Input[str]] = None): + ClusterLibraryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional[pulumi.Input['ClusterLibraryCranArgs']] = None, + egg: Optional[pulumi.Input[str]] = None, + jar: Optional[pulumi.Input[str]] = None, + maven: Optional[pulumi.Input['ClusterLibraryMavenArgs']] = None, + pypi: Optional[pulumi.Input['ClusterLibraryPypiArgs']] = None, + whl: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -1374,9 +1653,20 @@ class ClusterLibraryCranArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + ClusterLibraryCranArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -1403,11 +1693,24 @@ def __init__(__self__, *, coordinates: pulumi.Input[str], exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "coordinates", coordinates) + ClusterLibraryMavenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: pulumi.Input[str], + exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -1442,9 +1745,20 @@ class ClusterLibraryPypiArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + ClusterLibraryPypiArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -1469,7 +1783,16 @@ def repo(self, value: Optional[pulumi.Input[str]]): class ClusterWorkloadTypeArgs: def __init__(__self__, *, clients: pulumi.Input['ClusterWorkloadTypeClientsArgs']): - pulumi.set(__self__, "clients", clients) + ClusterWorkloadTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: pulumi.Input['ClusterWorkloadTypeClientsArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -1486,10 +1809,21 @@ class ClusterWorkloadTypeClientsArgs: def __init__(__self__, *, jobs: Optional[pulumi.Input[bool]] = None, notebooks: Optional[pulumi.Input[bool]] = None): + ClusterWorkloadTypeClientsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[pulumi.Input[bool]] = None, + notebooks: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -1514,8 +1848,17 @@ def notebooks(self, value: Optional[pulumi.Input[bool]]): class ExternalLocationEncryptionDetailsArgs: def __init__(__self__, *, sse_encryption_details: Optional[pulumi.Input['ExternalLocationEncryptionDetailsSseEncryptionDetailsArgs']] = None): + ExternalLocationEncryptionDetailsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + sse_encryption_details=sse_encryption_details, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + sse_encryption_details: Optional[pulumi.Input['ExternalLocationEncryptionDetailsSseEncryptionDetailsArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if sse_encryption_details is not None: - pulumi.set(__self__, "sse_encryption_details", sse_encryption_details) + _setter("sse_encryption_details", sse_encryption_details) @property @pulumi.getter(name="sseEncryptionDetails") @@ -1532,10 +1875,21 @@ class ExternalLocationEncryptionDetailsSseEncryptionDetailsArgs: def __init__(__self__, *, algorithm: Optional[pulumi.Input[str]] = None, aws_kms_key_arn: Optional[pulumi.Input[str]] = None): + ExternalLocationEncryptionDetailsSseEncryptionDetailsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + algorithm=algorithm, + aws_kms_key_arn=aws_kms_key_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + algorithm: Optional[pulumi.Input[str]] = None, + aws_kms_key_arn: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if algorithm is not None: - pulumi.set(__self__, "algorithm", algorithm) + _setter("algorithm", algorithm) if aws_kms_key_arn is not None: - pulumi.set(__self__, "aws_kms_key_arn", aws_kms_key_arn) + _setter("aws_kms_key_arn", aws_kms_key_arn) @property @pulumi.getter @@ -1561,8 +1915,19 @@ class GrantsGrantArgs: def __init__(__self__, *, principal: pulumi.Input[str], privileges: pulumi.Input[Sequence[pulumi.Input[str]]]): - pulumi.set(__self__, "principal", principal) - pulumi.set(__self__, "privileges", privileges) + GrantsGrantArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + principal=principal, + privileges=privileges, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + principal: pulumi.Input[str], + privileges: pulumi.Input[Sequence[pulumi.Input[str]]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("principal", principal) + _setter("privileges", privileges) @property @pulumi.getter @@ -1594,12 +1959,25 @@ def __init__(__self__, *, :param pulumi.Input[int] spot_bid_price_percent: (Integer) The max price for AWS spot instances, as a percentage of the corresponding instance type’s on-demand price. For example, if this field is set to 50, and the instance pool needs a new i3.xlarge spot instance, then the max price is half of the price of on-demand i3.xlarge instances. Similarly, if this field is set to 200, the max price is twice the price of on-demand i3.xlarge instances. If not specified, the *default value is 100*. When spot instances are requested for this instance pool, only spot instances whose max price percentage matches this field are considered. *For safety, this field cannot be greater than 10000.* :param pulumi.Input[str] zone_id: (String) Identifier for the availability zone/datacenter in which the instance pool resides. This string is of the form like `"us-west-2a"`. The provided availability zone must be in the same region as the Databricks deployment. For example, `"us-west-2a"` is not a valid zone ID if the Databricks deployment resides in the `"us-east-1"` region. If not specified, a default zone is used. You can find the list of available zones as well as the default value by using the [List Zones API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistavailablezones). """ + InstancePoolAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + spot_bid_price_percent: Optional[pulumi.Input[int]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -1647,10 +2025,21 @@ def __init__(__self__, *, :param pulumi.Input[str] availability: Availability type used for all nodes. Valid values are `SPOT_AZURE` and `ON_DEMAND_AZURE`. :param pulumi.Input[float] spot_bid_max_price: The max price for Azure spot instances. Use `-1` to specify the lowest price. """ + InstancePoolAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + spot_bid_max_price: Optional[pulumi.Input[float]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -1687,12 +2076,25 @@ def __init__(__self__, *, :param pulumi.Input[int] disk_count: (Integer) The number of disks to attach to each instance. This feature is only enabled for supported node types. Users can choose up to the limit of the disks supported by the node type. For node types with no local disk, at least one disk needs to be specified. :param pulumi.Input[int] disk_size: (Integer) The size of each disk (in GiB) to attach. """ + InstancePoolDiskSpecArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + disk_count=disk_count, + disk_size=disk_size, + disk_type=disk_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + disk_count: Optional[pulumi.Input[int]] = None, + disk_size: Optional[pulumi.Input[int]] = None, + disk_type: Optional[pulumi.Input['InstancePoolDiskSpecDiskTypeArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if disk_count is not None: - pulumi.set(__self__, "disk_count", disk_count) + _setter("disk_count", disk_count) if disk_size is not None: - pulumi.set(__self__, "disk_size", disk_size) + _setter("disk_size", disk_size) if disk_type is not None: - pulumi.set(__self__, "disk_type", disk_type) + _setter("disk_type", disk_type) @property @pulumi.getter(name="diskCount") @@ -1733,10 +2135,21 @@ class InstancePoolDiskSpecDiskTypeArgs: def __init__(__self__, *, azure_disk_volume_type: Optional[pulumi.Input[str]] = None, ebs_volume_type: Optional[pulumi.Input[str]] = None): + InstancePoolDiskSpecDiskTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + azure_disk_volume_type=azure_disk_volume_type, + ebs_volume_type=ebs_volume_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + azure_disk_volume_type: Optional[pulumi.Input[str]] = None, + ebs_volume_type: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if azure_disk_volume_type is not None: - pulumi.set(__self__, "azure_disk_volume_type", azure_disk_volume_type) + _setter("azure_disk_volume_type", azure_disk_volume_type) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) @property @pulumi.getter(name="azureDiskVolumeType") @@ -1766,10 +2179,21 @@ def __init__(__self__, *, :param pulumi.Input[str] gcp_availability: Availability type used for all nodes. Valid values are `PREEMPTIBLE_GCP`, `PREEMPTIBLE_WITH_FALLBACK_GCP` and `ON_DEMAND_GCP`, default: `ON_DEMAND_GCP`. :param pulumi.Input[int] local_ssd_count: Number of local SSD disks (each is 375GB in size) that will be attached to each node of the cluster. """ + InstancePoolGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + gcp_availability=gcp_availability, + local_ssd_count=local_ssd_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gcp_availability: Optional[pulumi.Input[str]] = None, + local_ssd_count: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if gcp_availability is not None: - pulumi.set(__self__, "gcp_availability", gcp_availability) + _setter("gcp_availability", gcp_availability) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) @property @pulumi.getter(name="gcpAvailability") @@ -1802,11 +2226,24 @@ def __init__(__self__, *, launch_template_overrides: pulumi.Input[Sequence[pulumi.Input['InstancePoolInstancePoolFleetAttributesLaunchTemplateOverrideArgs']]], fleet_on_demand_option: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesFleetOnDemandOptionArgs']] = None, fleet_spot_option: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesFleetSpotOptionArgs']] = None): - pulumi.set(__self__, "launch_template_overrides", launch_template_overrides) + InstancePoolInstancePoolFleetAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + launch_template_overrides=launch_template_overrides, + fleet_on_demand_option=fleet_on_demand_option, + fleet_spot_option=fleet_spot_option, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + launch_template_overrides: pulumi.Input[Sequence[pulumi.Input['InstancePoolInstancePoolFleetAttributesLaunchTemplateOverrideArgs']]], + fleet_on_demand_option: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesFleetOnDemandOptionArgs']] = None, + fleet_spot_option: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesFleetSpotOptionArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("launch_template_overrides", launch_template_overrides) if fleet_on_demand_option is not None: - pulumi.set(__self__, "fleet_on_demand_option", fleet_on_demand_option) + _setter("fleet_on_demand_option", fleet_on_demand_option) if fleet_spot_option is not None: - pulumi.set(__self__, "fleet_spot_option", fleet_spot_option) + _setter("fleet_spot_option", fleet_spot_option) @property @pulumi.getter(name="launchTemplateOverrides") @@ -1841,9 +2278,20 @@ class InstancePoolInstancePoolFleetAttributesFleetOnDemandOptionArgs: def __init__(__self__, *, allocation_strategy: pulumi.Input[str], instance_pools_to_use_count: Optional[pulumi.Input[int]] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + InstancePoolInstancePoolFleetAttributesFleetOnDemandOptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: pulumi.Input[str], + instance_pools_to_use_count: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -1869,9 +2317,20 @@ class InstancePoolInstancePoolFleetAttributesFleetSpotOptionArgs: def __init__(__self__, *, allocation_strategy: pulumi.Input[str], instance_pools_to_use_count: Optional[pulumi.Input[int]] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + InstancePoolInstancePoolFleetAttributesFleetSpotOptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: pulumi.Input[str], + instance_pools_to_use_count: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -1897,8 +2356,19 @@ class InstancePoolInstancePoolFleetAttributesLaunchTemplateOverrideArgs: def __init__(__self__, *, availability_zone: pulumi.Input[str], instance_type: pulumi.Input[str]): - pulumi.set(__self__, "availability_zone", availability_zone) - pulumi.set(__self__, "instance_type", instance_type) + InstancePoolInstancePoolFleetAttributesLaunchTemplateOverrideArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability_zone=availability_zone, + instance_type=instance_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability_zone: pulumi.Input[str], + instance_type: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("availability_zone", availability_zone) + _setter("instance_type", instance_type) @property @pulumi.getter(name="availabilityZone") @@ -1924,9 +2394,20 @@ class InstancePoolPreloadedDockerImageArgs: def __init__(__self__, *, url: pulumi.Input[str], basic_auth: Optional[pulumi.Input['InstancePoolPreloadedDockerImageBasicAuthArgs']] = None): - pulumi.set(__self__, "url", url) + InstancePoolPreloadedDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + basic_auth: Optional[pulumi.Input['InstancePoolPreloadedDockerImageBasicAuthArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -1952,8 +2433,19 @@ class InstancePoolPreloadedDockerImageBasicAuthArgs: def __init__(__self__, *, password: pulumi.Input[str], username: pulumi.Input[str]): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + InstancePoolPreloadedDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: pulumi.Input[str], + username: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -1979,10 +2471,21 @@ class JobComputeArgs: def __init__(__self__, *, compute_key: Optional[pulumi.Input[str]] = None, spec: Optional[pulumi.Input['JobComputeSpecArgs']] = None): + JobComputeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + compute_key=compute_key, + spec=spec, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + compute_key: Optional[pulumi.Input[str]] = None, + spec: Optional[pulumi.Input['JobComputeSpecArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if spec is not None: - pulumi.set(__self__, "spec", spec) + _setter("spec", spec) @property @pulumi.getter(name="computeKey") @@ -2007,8 +2510,17 @@ def spec(self, value: Optional[pulumi.Input['JobComputeSpecArgs']]): class JobComputeSpecArgs: def __init__(__self__, *, kind: Optional[pulumi.Input[str]] = None): + JobComputeSpecArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + kind=kind, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + kind: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if kind is not None: - pulumi.set(__self__, "kind", kind) + _setter("kind", kind) @property @pulumi.getter @@ -2027,8 +2539,17 @@ def __init__(__self__, *, """ :param pulumi.Input[str] pause_status: Indicate whether this continuous job is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pause_status`. """ + JobContinuousArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pause_status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="pauseStatus") @@ -2062,17 +2583,36 @@ def __init__(__self__, *, You also need to include a `git_source` block to configure the repository that contains the dbt project. """ - pulumi.set(__self__, "commands", commands) + JobDbtTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: pulumi.Input[Sequence[pulumi.Input[str]]], + catalog: Optional[pulumi.Input[str]] = None, + profiles_directory: Optional[pulumi.Input[str]] = None, + project_directory: Optional[pulumi.Input[str]] = None, + schema: Optional[pulumi.Input[str]] = None, + warehouse_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -2166,18 +2706,37 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input[str]]] on_starts: (List) list of emails to notify when the run starts. :param pulumi.Input[Sequence[pulumi.Input[str]]] on_successes: (List) list of emails to notify when the run completes successfully. """ + JobEmailNotificationsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[pulumi.Input[bool]] = None, + no_alert_for_skipped_runs: Optional[pulumi.Input[bool]] = None, + on_duration_warning_threshold_exceededs: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + on_failures: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + on_starts: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + on_successes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -2268,17 +2827,36 @@ def __init__(__self__, *, :param pulumi.Input[str] provider: case insensitive name of the Git provider. Following values are supported right now (could be a subject for change, consult [Repos API documentation](https://docs.databricks.com/dev-tools/api/latest/repos.html)): `gitHub`, `gitHubEnterprise`, `bitbucketCloud`, `bitbucketServer`, `azureDevOpsServices`, `gitLab`, `gitLabEnterpriseEdition`. :param pulumi.Input[str] tag: name of the Git branch to use. Conflicts with `branch` and `commit`. """ - pulumi.set(__self__, "url", url) + JobGitSourceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + branch=branch, + commit=commit, + job_source=job_source, + provider=provider, + tag=tag, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + branch: Optional[pulumi.Input[str]] = None, + commit: Optional[pulumi.Input[str]] = None, + job_source: Optional[pulumi.Input['JobGitSourceJobSourceArgs']] = None, + provider: Optional[pulumi.Input[str]] = None, + tag: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if branch is not None: - pulumi.set(__self__, "branch", branch) + _setter("branch", branch) if commit is not None: - pulumi.set(__self__, "commit", commit) + _setter("commit", commit) if job_source is not None: - pulumi.set(__self__, "job_source", job_source) + _setter("job_source", job_source) if provider is not None: - pulumi.set(__self__, "provider", provider) + _setter("provider", provider) if tag is not None: - pulumi.set(__self__, "tag", tag) + _setter("tag", tag) @property @pulumi.getter @@ -2356,10 +2934,23 @@ def __init__(__self__, *, import_from_git_branch: pulumi.Input[str], job_config_path: pulumi.Input[str], dirty_state: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "import_from_git_branch", import_from_git_branch) - pulumi.set(__self__, "job_config_path", job_config_path) + JobGitSourceJobSourceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + import_from_git_branch=import_from_git_branch, + job_config_path=job_config_path, + dirty_state=dirty_state, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + import_from_git_branch: pulumi.Input[str], + job_config_path: pulumi.Input[str], + dirty_state: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("import_from_git_branch", import_from_git_branch) + _setter("job_config_path", job_config_path) if dirty_state is not None: - pulumi.set(__self__, "dirty_state", dirty_state) + _setter("dirty_state", dirty_state) @property @pulumi.getter(name="importFromGitBranch") @@ -2396,7 +2987,16 @@ def __init__(__self__, *, """ :param pulumi.Input[Sequence[pulumi.Input['JobHealthRuleArgs']]] rules: list of rules that are represented as objects with the following attributes: """ - pulumi.set(__self__, "rules", rules) + JobHealthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: pulumi.Input[Sequence[pulumi.Input['JobHealthRuleArgs']]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -2422,12 +3022,25 @@ def __init__(__self__, *, :param pulumi.Input[str] op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param pulumi.Input[int] value: integer value used to compare to the given metric. """ + JobHealthRuleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[pulumi.Input[str]] = None, + op: Optional[pulumi.Input[str]] = None, + value: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -2475,10 +3088,21 @@ def __init__(__self__, *, :param pulumi.Input[str] job_cluster_key: Identifier that can be referenced in `task` block, so that cluster is shared between tasks :param pulumi.Input['JobJobClusterNewClusterArgs'] new_cluster: Same set of parameters as for Cluster resource. """ + JobJobClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_cluster_key=job_cluster_key, + new_cluster=new_cluster, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_cluster_key: Optional[pulumi.Input[str]] = None, + new_cluster: Optional[pulumi.Input['JobJobClusterNewClusterArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") @@ -2538,65 +3162,132 @@ def __init__(__self__, *, spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, workload_type: Optional[pulumi.Input['JobJobClusterNewClusterWorkloadTypeArgs']] = None): - pulumi.set(__self__, "spark_version", spark_version) + JobJobClusterNewClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + spark_version: pulumi.Input[str], + apply_policy_default_values: Optional[pulumi.Input[bool]] = None, + autoscale: Optional[pulumi.Input['JobJobClusterNewClusterAutoscaleArgs']] = None, + autotermination_minutes: Optional[pulumi.Input[int]] = None, + aws_attributes: Optional[pulumi.Input['JobJobClusterNewClusterAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['JobJobClusterNewClusterAzureAttributesArgs']] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + cluster_log_conf: Optional[pulumi.Input['JobJobClusterNewClusterClusterLogConfArgs']] = None, + cluster_mount_infos: Optional[pulumi.Input[Sequence[pulumi.Input['JobJobClusterNewClusterClusterMountInfoArgs']]]] = None, + cluster_name: Optional[pulumi.Input[str]] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + data_security_mode: Optional[pulumi.Input[str]] = None, + docker_image: Optional[pulumi.Input['JobJobClusterNewClusterDockerImageArgs']] = None, + driver_instance_pool_id: Optional[pulumi.Input[str]] = None, + driver_node_type_id: Optional[pulumi.Input[str]] = None, + enable_elastic_disk: Optional[pulumi.Input[bool]] = None, + enable_local_disk_encryption: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['JobJobClusterNewClusterGcpAttributesArgs']] = None, + idempotency_token: Optional[pulumi.Input[str]] = None, + init_scripts: Optional[pulumi.Input[Sequence[pulumi.Input['JobJobClusterNewClusterInitScriptArgs']]]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + num_workers: Optional[pulumi.Input[int]] = None, + policy_id: Optional[pulumi.Input[str]] = None, + runtime_engine: Optional[pulumi.Input[str]] = None, + single_user_name: Optional[pulumi.Input[str]] = None, + spark_conf: Optional[pulumi.Input[Mapping[str, Any]]] = None, + spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, + ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + workload_type: Optional[pulumi.Input['JobJobClusterNewClusterWorkloadTypeArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="sparkVersion") @@ -2874,10 +3565,21 @@ class JobJobClusterNewClusterAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[pulumi.Input[int]] = None, min_workers: Optional[pulumi.Input[int]] = None): + JobJobClusterNewClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[pulumi.Input[int]] = None, + min_workers: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -2909,22 +3611,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[pulumi.Input[str]] = None, spot_bid_price_percent: Optional[pulumi.Input[int]] = None, zone_id: Optional[pulumi.Input[str]] = None): + JobJobClusterNewClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + ebs_volume_count: Optional[pulumi.Input[int]] = None, + ebs_volume_size: Optional[pulumi.Input[int]] = None, + ebs_volume_type: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + spot_bid_price_percent: Optional[pulumi.Input[int]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -3005,12 +3730,25 @@ def __init__(__self__, *, availability: Optional[pulumi.Input[str]] = None, first_on_demand: Optional[pulumi.Input[int]] = None, spot_bid_max_price: Optional[pulumi.Input[float]] = None): + JobJobClusterNewClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + spot_bid_max_price: Optional[pulumi.Input[float]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -3045,10 +3783,21 @@ class JobJobClusterNewClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional[pulumi.Input['JobJobClusterNewClusterClusterLogConfDbfsArgs']] = None, s3: Optional[pulumi.Input['JobJobClusterNewClusterClusterLogConfS3Args']] = None): + JobJobClusterNewClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional[pulumi.Input['JobJobClusterNewClusterClusterLogConfDbfsArgs']] = None, + s3: Optional[pulumi.Input['JobJobClusterNewClusterClusterLogConfS3Args']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -3073,7 +3822,16 @@ def s3(self, value: Optional[pulumi.Input['JobJobClusterNewClusterClusterLogConf class JobJobClusterNewClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -3095,19 +3853,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -3179,10 +3958,23 @@ def __init__(__self__, *, local_mount_dir_path: pulumi.Input[str], network_filesystem_info: pulumi.Input['JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoArgs'], remote_mount_dir_path: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + JobJobClusterNewClusterClusterMountInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: pulumi.Input[str], + network_filesystem_info: pulumi.Input['JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoArgs'], + remote_mount_dir_path: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -3217,9 +4009,20 @@ class JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoArgs: def __init__(__self__, *, server_address: pulumi.Input[str], mount_options: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "server_address", server_address) + JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: pulumi.Input[str], + mount_options: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -3246,17 +4049,28 @@ def __init__(__self__, *, url: pulumi.Input[str], basic_auth: Optional[pulumi.Input['JobJobClusterNewClusterDockerImageBasicAuthArgs']] = None): """ - :param pulumi.Input[str] url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). - """ - pulumi.set(__self__, "url", url) + :param pulumi.Input[str] url: URL of the job on the given workspace + """ + JobJobClusterNewClusterDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + basic_auth: Optional[pulumi.Input['JobJobClusterNewClusterDockerImageBasicAuthArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter def url(self) -> pulumi.Input[str]: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -3279,8 +4093,19 @@ class JobJobClusterNewClusterDockerImageBasicAuthArgs: def __init__(__self__, *, password: pulumi.Input[str], username: pulumi.Input[str]): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + JobJobClusterNewClusterDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: pulumi.Input[str], + username: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -3310,18 +4135,37 @@ def __init__(__self__, *, local_ssd_count: Optional[pulumi.Input[int]] = None, use_preemptible_executors: Optional[pulumi.Input[bool]] = None, zone_id: Optional[pulumi.Input[str]] = None): + JobJobClusterNewClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + boot_disk_size: Optional[pulumi.Input[int]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + local_ssd_count: Optional[pulumi.Input[int]] = None, + use_preemptible_executors: Optional[pulumi.Input[bool]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -3434,23 +4278,44 @@ def __init__(__self__, *, ]) ``` """ + JobJobClusterNewClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional[pulumi.Input['JobJobClusterNewClusterInitScriptAbfssArgs']] = None, + dbfs: Optional[pulumi.Input['JobJobClusterNewClusterInitScriptDbfsArgs']] = None, + file: Optional[pulumi.Input['JobJobClusterNewClusterInitScriptFileArgs']] = None, + gcs: Optional[pulumi.Input['JobJobClusterNewClusterInitScriptGcsArgs']] = None, + s3: Optional[pulumi.Input['JobJobClusterNewClusterInitScriptS3Args']] = None, + volumes: Optional[pulumi.Input['JobJobClusterNewClusterInitScriptVolumesArgs']] = None, + workspace: Optional[pulumi.Input['JobJobClusterNewClusterInitScriptWorkspaceArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: warnings.warn("""For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""", DeprecationWarning) pulumi.log.warn("""dbfs is deprecated: For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""") if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -3569,8 +4434,17 @@ def workspace(self, value: Optional[pulumi.Input['JobJobClusterNewClusterInitScr class JobJobClusterNewClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobJobClusterNewClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3586,7 +4460,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobJobClusterNewClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -3602,8 +4485,17 @@ def destination(self, value: pulumi.Input[str]): class JobJobClusterNewClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobJobClusterNewClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3619,8 +4511,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobJobClusterNewClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobJobClusterNewClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3642,19 +4543,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -3724,8 +4646,17 @@ def region(self, value: Optional[pulumi.Input[str]]): class JobJobClusterNewClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobJobClusterNewClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3741,8 +4672,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobJobClusterNewClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobJobClusterNewClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3758,7 +4698,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobJobClusterNewClusterWorkloadTypeArgs: def __init__(__self__, *, clients: pulumi.Input['JobJobClusterNewClusterWorkloadTypeClientsArgs']): - pulumi.set(__self__, "clients", clients) + JobJobClusterNewClusterWorkloadTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: pulumi.Input['JobJobClusterNewClusterWorkloadTypeClientsArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -3775,10 +4724,21 @@ class JobJobClusterNewClusterWorkloadTypeClientsArgs: def __init__(__self__, *, jobs: Optional[pulumi.Input[bool]] = None, notebooks: Optional[pulumi.Input[bool]] = None): + JobJobClusterNewClusterWorkloadTypeClientsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[pulumi.Input[bool]] = None, + notebooks: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -3808,18 +4768,37 @@ def __init__(__self__, *, maven: Optional[pulumi.Input['JobLibraryMavenArgs']] = None, pypi: Optional[pulumi.Input['JobLibraryPypiArgs']] = None, whl: Optional[pulumi.Input[str]] = None): + JobLibraryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional[pulumi.Input['JobLibraryCranArgs']] = None, + egg: Optional[pulumi.Input[str]] = None, + jar: Optional[pulumi.Input[str]] = None, + maven: Optional[pulumi.Input['JobLibraryMavenArgs']] = None, + pypi: Optional[pulumi.Input['JobLibraryPypiArgs']] = None, + whl: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -3881,9 +4860,20 @@ class JobLibraryCranArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + JobLibraryCranArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -3910,11 +4900,24 @@ def __init__(__self__, *, coordinates: pulumi.Input[str], exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "coordinates", coordinates) + JobLibraryMavenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: pulumi.Input[str], + exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -3949,9 +4952,20 @@ class JobLibraryPypiArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + JobLibraryPypiArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -4005,65 +5019,132 @@ def __init__(__self__, *, spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, workload_type: Optional[pulumi.Input['JobNewClusterWorkloadTypeArgs']] = None): - pulumi.set(__self__, "spark_version", spark_version) + JobNewClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + spark_version: pulumi.Input[str], + apply_policy_default_values: Optional[pulumi.Input[bool]] = None, + autoscale: Optional[pulumi.Input['JobNewClusterAutoscaleArgs']] = None, + autotermination_minutes: Optional[pulumi.Input[int]] = None, + aws_attributes: Optional[pulumi.Input['JobNewClusterAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['JobNewClusterAzureAttributesArgs']] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + cluster_log_conf: Optional[pulumi.Input['JobNewClusterClusterLogConfArgs']] = None, + cluster_mount_infos: Optional[pulumi.Input[Sequence[pulumi.Input['JobNewClusterClusterMountInfoArgs']]]] = None, + cluster_name: Optional[pulumi.Input[str]] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + data_security_mode: Optional[pulumi.Input[str]] = None, + docker_image: Optional[pulumi.Input['JobNewClusterDockerImageArgs']] = None, + driver_instance_pool_id: Optional[pulumi.Input[str]] = None, + driver_node_type_id: Optional[pulumi.Input[str]] = None, + enable_elastic_disk: Optional[pulumi.Input[bool]] = None, + enable_local_disk_encryption: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['JobNewClusterGcpAttributesArgs']] = None, + idempotency_token: Optional[pulumi.Input[str]] = None, + init_scripts: Optional[pulumi.Input[Sequence[pulumi.Input['JobNewClusterInitScriptArgs']]]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + num_workers: Optional[pulumi.Input[int]] = None, + policy_id: Optional[pulumi.Input[str]] = None, + runtime_engine: Optional[pulumi.Input[str]] = None, + single_user_name: Optional[pulumi.Input[str]] = None, + spark_conf: Optional[pulumi.Input[Mapping[str, Any]]] = None, + spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, + ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + workload_type: Optional[pulumi.Input['JobNewClusterWorkloadTypeArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="sparkVersion") @@ -4341,10 +5422,21 @@ class JobNewClusterAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[pulumi.Input[int]] = None, min_workers: Optional[pulumi.Input[int]] = None): + JobNewClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[pulumi.Input[int]] = None, + min_workers: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -4376,22 +5468,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[pulumi.Input[str]] = None, spot_bid_price_percent: Optional[pulumi.Input[int]] = None, zone_id: Optional[pulumi.Input[str]] = None): + JobNewClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + ebs_volume_count: Optional[pulumi.Input[int]] = None, + ebs_volume_size: Optional[pulumi.Input[int]] = None, + ebs_volume_type: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + spot_bid_price_percent: Optional[pulumi.Input[int]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -4472,12 +5587,25 @@ def __init__(__self__, *, availability: Optional[pulumi.Input[str]] = None, first_on_demand: Optional[pulumi.Input[int]] = None, spot_bid_max_price: Optional[pulumi.Input[float]] = None): + JobNewClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + spot_bid_max_price: Optional[pulumi.Input[float]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -4512,10 +5640,21 @@ class JobNewClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional[pulumi.Input['JobNewClusterClusterLogConfDbfsArgs']] = None, s3: Optional[pulumi.Input['JobNewClusterClusterLogConfS3Args']] = None): + JobNewClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional[pulumi.Input['JobNewClusterClusterLogConfDbfsArgs']] = None, + s3: Optional[pulumi.Input['JobNewClusterClusterLogConfS3Args']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -4540,7 +5679,16 @@ def s3(self, value: Optional[pulumi.Input['JobNewClusterClusterLogConfS3Args']]) class JobNewClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + JobNewClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -4562,19 +5710,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + JobNewClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -4646,10 +5815,23 @@ def __init__(__self__, *, local_mount_dir_path: pulumi.Input[str], network_filesystem_info: pulumi.Input['JobNewClusterClusterMountInfoNetworkFilesystemInfoArgs'], remote_mount_dir_path: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + JobNewClusterClusterMountInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: pulumi.Input[str], + network_filesystem_info: pulumi.Input['JobNewClusterClusterMountInfoNetworkFilesystemInfoArgs'], + remote_mount_dir_path: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -4684,9 +5866,20 @@ class JobNewClusterClusterMountInfoNetworkFilesystemInfoArgs: def __init__(__self__, *, server_address: pulumi.Input[str], mount_options: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "server_address", server_address) + JobNewClusterClusterMountInfoNetworkFilesystemInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: pulumi.Input[str], + mount_options: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -4713,17 +5906,28 @@ def __init__(__self__, *, url: pulumi.Input[str], basic_auth: Optional[pulumi.Input['JobNewClusterDockerImageBasicAuthArgs']] = None): """ - :param pulumi.Input[str] url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). - """ - pulumi.set(__self__, "url", url) + :param pulumi.Input[str] url: URL of the job on the given workspace + """ + JobNewClusterDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + basic_auth: Optional[pulumi.Input['JobNewClusterDockerImageBasicAuthArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter def url(self) -> pulumi.Input[str]: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -4746,8 +5950,19 @@ class JobNewClusterDockerImageBasicAuthArgs: def __init__(__self__, *, password: pulumi.Input[str], username: pulumi.Input[str]): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + JobNewClusterDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: pulumi.Input[str], + username: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -4777,18 +5992,37 @@ def __init__(__self__, *, local_ssd_count: Optional[pulumi.Input[int]] = None, use_preemptible_executors: Optional[pulumi.Input[bool]] = None, zone_id: Optional[pulumi.Input[str]] = None): + JobNewClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + boot_disk_size: Optional[pulumi.Input[int]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + local_ssd_count: Optional[pulumi.Input[int]] = None, + use_preemptible_executors: Optional[pulumi.Input[bool]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -4901,23 +6135,44 @@ def __init__(__self__, *, ]) ``` """ + JobNewClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional[pulumi.Input['JobNewClusterInitScriptAbfssArgs']] = None, + dbfs: Optional[pulumi.Input['JobNewClusterInitScriptDbfsArgs']] = None, + file: Optional[pulumi.Input['JobNewClusterInitScriptFileArgs']] = None, + gcs: Optional[pulumi.Input['JobNewClusterInitScriptGcsArgs']] = None, + s3: Optional[pulumi.Input['JobNewClusterInitScriptS3Args']] = None, + volumes: Optional[pulumi.Input['JobNewClusterInitScriptVolumesArgs']] = None, + workspace: Optional[pulumi.Input['JobNewClusterInitScriptWorkspaceArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: warnings.warn("""For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""", DeprecationWarning) pulumi.log.warn("""dbfs is deprecated: For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""") if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -5036,8 +6291,17 @@ def workspace(self, value: Optional[pulumi.Input['JobNewClusterInitScriptWorkspa class JobNewClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobNewClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -5053,7 +6317,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobNewClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + JobNewClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -5069,8 +6342,17 @@ def destination(self, value: pulumi.Input[str]): class JobNewClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobNewClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -5086,8 +6368,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobNewClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobNewClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -5109,19 +6400,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + JobNewClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -5191,8 +6503,17 @@ def region(self, value: Optional[pulumi.Input[str]]): class JobNewClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobNewClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -5208,8 +6529,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobNewClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobNewClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -5225,7 +6555,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobNewClusterWorkloadTypeArgs: def __init__(__self__, *, clients: pulumi.Input['JobNewClusterWorkloadTypeClientsArgs']): - pulumi.set(__self__, "clients", clients) + JobNewClusterWorkloadTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: pulumi.Input['JobNewClusterWorkloadTypeClientsArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -5242,10 +6581,21 @@ class JobNewClusterWorkloadTypeClientsArgs: def __init__(__self__, *, jobs: Optional[pulumi.Input[bool]] = None, notebooks: Optional[pulumi.Input[bool]] = None): + JobNewClusterWorkloadTypeClientsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[pulumi.Input[bool]] = None, + notebooks: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -5277,11 +6627,24 @@ def __init__(__self__, *, :param pulumi.Input[Mapping[str, Any]] base_parameters: (Map) Base parameters to be used for each run of this job. If the run is initiated by a call to run-now with parameters specified, the two parameters maps will be merged. If the same key is specified in base_parameters and in run-now, the value from run-now will be used. If the notebook takes a parameter that is not specified in the job’s base_parameters or the run-now override parameters, the default value from the notebook will be used. Retrieve these parameters in a notebook using `dbutils.widgets.get`. :param pulumi.Input[str] source: Location type of the notebook, can only be `WORKSPACE` or `GIT`. When set to `WORKSPACE`, the notebook will be retrieved from the local Databricks workspace. When set to `GIT`, the notebook will be retrieved from a Git repository defined in `git_source`. If the value is empty, the task will use `GIT` if `git_source` is defined and `WORKSPACE` otherwise. """ - pulumi.set(__self__, "notebook_path", notebook_path) + JobNotebookTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: pulumi.Input[str], + base_parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -5329,10 +6692,21 @@ def __init__(__self__, *, :param pulumi.Input[bool] no_alert_for_canceled_runs: (Bool) don't send alert for cancelled runs. :param pulumi.Input[bool] no_alert_for_skipped_runs: (Bool) don't send alert for skipped runs. """ + JobNotificationSettingsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + no_alert_for_canceled_runs: Optional[pulumi.Input[bool]] = None, + no_alert_for_skipped_runs: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="noAlertForCanceledRuns") @@ -5368,10 +6742,21 @@ def __init__(__self__, *, :param pulumi.Input[str] default: Default value of the parameter. :param pulumi.Input[str] name: An optional name for the job. The default value is Untitled. """ + JobParameterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + default=default, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if default is not None: - pulumi.set(__self__, "default", default) + _setter("default", default) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -5409,9 +6794,20 @@ def __init__(__self__, *, > **Note** The following configuration blocks are only supported inside a `task` block """ - pulumi.set(__self__, "pipeline_id", pipeline_id) + JobPipelineTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: pulumi.Input[str], + full_refresh: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -5453,14 +6849,29 @@ def __init__(__self__, *, :param pulumi.Input[str] package_name: Name of Python package :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: Parameters for the task """ + JobPythonWheelTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[pulumi.Input[str]] = None, + named_parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, + package_name: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -5515,6 +6926,11 @@ def parameters(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]) class JobQueueArgs: def __init__(__self__): pass + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + opts: Optional[pulumi.ResourceOptions]=None): + pass @pulumi.input_type @@ -5537,10 +6953,21 @@ def __init__(__self__, *, ``` :param pulumi.Input[str] user_name: The email of an active workspace user. Non-admin users can only set this field to their own email. """ + JobRunAsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + service_principal_name=service_principal_name, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + service_principal_name: Optional[pulumi.Input[str]] = None, + user_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if service_principal_name is not None: - pulumi.set(__self__, "service_principal_name", service_principal_name) + _setter("service_principal_name", service_principal_name) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="servicePrincipalName") @@ -5587,9 +7014,20 @@ def __init__(__self__, *, :param pulumi.Input[int] job_id: (String) ID of the job :param pulumi.Input[Mapping[str, Any]] job_parameters: (Map) Job parameters for the task """ - pulumi.set(__self__, "job_id", job_id) + JobRunJobTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: pulumi.Input[int], + job_parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -5627,10 +7065,23 @@ def __init__(__self__, *, :param pulumi.Input[str] timezone_id: A Java timezone ID. The schedule for a job will be resolved with respect to this timezone. See Java TimeZone for details. This field is required. :param pulumi.Input[str] pause_status: Indicate whether this schedule is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted and a schedule is provided, the server will default to using `UNPAUSED` as a value for `pause_status`. """ - pulumi.set(__self__, "quartz_cron_expression", quartz_cron_expression) - pulumi.set(__self__, "timezone_id", timezone_id) + JobScheduleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + quartz_cron_expression=quartz_cron_expression, + timezone_id=timezone_id, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + quartz_cron_expression: pulumi.Input[str], + timezone_id: pulumi.Input[str], + pause_status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("quartz_cron_expression", quartz_cron_expression) + _setter("timezone_id", timezone_id) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="quartzCronExpression") @@ -5679,12 +7130,25 @@ def __init__(__self__, *, :param pulumi.Input[str] main_class_name: The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. The code should use `SparkContext.getOrCreate` to obtain a Spark context; otherwise, runs of the job will fail. :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: (List) Parameters passed to the main method. """ + JobSparkJarTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[pulumi.Input[str]] = None, + main_class_name: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -5731,11 +7195,24 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: (List) Command line parameters passed to the Python file. :param pulumi.Input[str] source: Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `git_source`. """ - pulumi.set(__self__, "python_file", python_file) + JobSparkPythonTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: pulumi.Input[str], + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -5781,8 +7258,17 @@ def __init__(__self__, *, """ :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: (List) Command-line parameters passed to spark submit. """ + JobSparkSubmitTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -5842,58 +7328,117 @@ def __init__(__self__, *, * `*_task` - (Required) one of the specific task blocks described below: :param pulumi.Input[int] timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. """ + JobTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + compute_key=compute_key, + condition_task=condition_task, + dbt_task=dbt_task, + depends_ons=depends_ons, + description=description, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + health=health, + job_cluster_key=job_cluster_key, + libraries=libraries, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + retry_on_timeout=retry_on_timeout, + run_if=run_if, + run_job_task=run_job_task, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + sql_task=sql_task, + task_key=task_key, + timeout_seconds=timeout_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + compute_key: Optional[pulumi.Input[str]] = None, + condition_task: Optional[pulumi.Input['JobTaskConditionTaskArgs']] = None, + dbt_task: Optional[pulumi.Input['JobTaskDbtTaskArgs']] = None, + depends_ons: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskDependsOnArgs']]]] = None, + description: Optional[pulumi.Input[str]] = None, + email_notifications: Optional[pulumi.Input['JobTaskEmailNotificationsArgs']] = None, + existing_cluster_id: Optional[pulumi.Input[str]] = None, + health: Optional[pulumi.Input['JobTaskHealthArgs']] = None, + job_cluster_key: Optional[pulumi.Input[str]] = None, + libraries: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskLibraryArgs']]]] = None, + max_retries: Optional[pulumi.Input[int]] = None, + min_retry_interval_millis: Optional[pulumi.Input[int]] = None, + new_cluster: Optional[pulumi.Input['JobTaskNewClusterArgs']] = None, + notebook_task: Optional[pulumi.Input['JobTaskNotebookTaskArgs']] = None, + notification_settings: Optional[pulumi.Input['JobTaskNotificationSettingsArgs']] = None, + pipeline_task: Optional[pulumi.Input['JobTaskPipelineTaskArgs']] = None, + python_wheel_task: Optional[pulumi.Input['JobTaskPythonWheelTaskArgs']] = None, + retry_on_timeout: Optional[pulumi.Input[bool]] = None, + run_if: Optional[pulumi.Input[str]] = None, + run_job_task: Optional[pulumi.Input['JobTaskRunJobTaskArgs']] = None, + spark_jar_task: Optional[pulumi.Input['JobTaskSparkJarTaskArgs']] = None, + spark_python_task: Optional[pulumi.Input['JobTaskSparkPythonTaskArgs']] = None, + spark_submit_task: Optional[pulumi.Input['JobTaskSparkSubmitTaskArgs']] = None, + sql_task: Optional[pulumi.Input['JobTaskSqlTaskArgs']] = None, + task_key: Optional[pulumi.Input[str]] = None, + timeout_seconds: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if condition_task is not None: - pulumi.set(__self__, "condition_task", condition_task) + _setter("condition_task", condition_task) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if depends_ons is not None: - pulumi.set(__self__, "depends_ons", depends_ons) + _setter("depends_ons", depends_ons) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if retry_on_timeout is not None: - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + _setter("retry_on_timeout", retry_on_timeout) if run_if is not None: - pulumi.set(__self__, "run_if", run_if) + _setter("run_if", run_if) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if sql_task is not None: - pulumi.set(__self__, "sql_task", sql_task) + _setter("sql_task", sql_task) if task_key is not None: - pulumi.set(__self__, "task_key", task_key) + _setter("task_key", task_key) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) @property @pulumi.getter(name="computeKey") @@ -6179,12 +7724,25 @@ def __init__(__self__, *, """ :param pulumi.Input[str] op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ + JobTaskConditionTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + left=left, + op=op, + right=right, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + left: Optional[pulumi.Input[str]] = None, + op: Optional[pulumi.Input[str]] = None, + right: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if left is not None: - pulumi.set(__self__, "left", left) + _setter("left", left) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if right is not None: - pulumi.set(__self__, "right", right) + _setter("right", right) @property @pulumi.getter @@ -6236,17 +7794,36 @@ def __init__(__self__, *, You also need to include a `git_source` block to configure the repository that contains the dbt project. """ - pulumi.set(__self__, "commands", commands) + JobTaskDbtTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: pulumi.Input[Sequence[pulumi.Input[str]]], + catalog: Optional[pulumi.Input[str]] = None, + profiles_directory: Optional[pulumi.Input[str]] = None, + project_directory: Optional[pulumi.Input[str]] = None, + schema: Optional[pulumi.Input[str]] = None, + warehouse_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -6331,9 +7908,20 @@ def __init__(__self__, *, """ :param pulumi.Input[str] task_key: The name of the task this task depends on. """ - pulumi.set(__self__, "task_key", task_key) + JobTaskDependsOnArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + task_key=task_key, + outcome=outcome, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + task_key: pulumi.Input[str], + outcome: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("task_key", task_key) if outcome is not None: - pulumi.set(__self__, "outcome", outcome) + _setter("outcome", outcome) @property @pulumi.getter(name="taskKey") @@ -6374,18 +7962,37 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input[str]]] on_starts: (List) list of emails to notify when the run starts. :param pulumi.Input[Sequence[pulumi.Input[str]]] on_successes: (List) list of emails to notify when the run completes successfully. """ + JobTaskEmailNotificationsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[pulumi.Input[bool]] = None, + no_alert_for_skipped_runs: Optional[pulumi.Input[bool]] = None, + on_duration_warning_threshold_exceededs: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + on_failures: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + on_starts: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + on_successes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -6467,7 +8074,16 @@ def __init__(__self__, *, """ :param pulumi.Input[Sequence[pulumi.Input['JobTaskHealthRuleArgs']]] rules: list of rules that are represented as objects with the following attributes: """ - pulumi.set(__self__, "rules", rules) + JobTaskHealthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: pulumi.Input[Sequence[pulumi.Input['JobTaskHealthRuleArgs']]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -6493,12 +8109,25 @@ def __init__(__self__, *, :param pulumi.Input[str] op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param pulumi.Input[int] value: integer value used to compare to the given metric. """ + JobTaskHealthRuleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[pulumi.Input[str]] = None, + op: Optional[pulumi.Input[str]] = None, + value: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -6546,18 +8175,37 @@ def __init__(__self__, *, maven: Optional[pulumi.Input['JobTaskLibraryMavenArgs']] = None, pypi: Optional[pulumi.Input['JobTaskLibraryPypiArgs']] = None, whl: Optional[pulumi.Input[str]] = None): + JobTaskLibraryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional[pulumi.Input['JobTaskLibraryCranArgs']] = None, + egg: Optional[pulumi.Input[str]] = None, + jar: Optional[pulumi.Input[str]] = None, + maven: Optional[pulumi.Input['JobTaskLibraryMavenArgs']] = None, + pypi: Optional[pulumi.Input['JobTaskLibraryPypiArgs']] = None, + whl: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -6619,9 +8267,20 @@ class JobTaskLibraryCranArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + JobTaskLibraryCranArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -6648,11 +8307,24 @@ def __init__(__self__, *, coordinates: pulumi.Input[str], exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "coordinates", coordinates) + JobTaskLibraryMavenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: pulumi.Input[str], + exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -6687,9 +8359,20 @@ class JobTaskLibraryPypiArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + JobTaskLibraryPypiArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -6743,65 +8426,132 @@ def __init__(__self__, *, spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, workload_type: Optional[pulumi.Input['JobTaskNewClusterWorkloadTypeArgs']] = None): - pulumi.set(__self__, "spark_version", spark_version) + JobTaskNewClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + spark_version: pulumi.Input[str], + apply_policy_default_values: Optional[pulumi.Input[bool]] = None, + autoscale: Optional[pulumi.Input['JobTaskNewClusterAutoscaleArgs']] = None, + autotermination_minutes: Optional[pulumi.Input[int]] = None, + aws_attributes: Optional[pulumi.Input['JobTaskNewClusterAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['JobTaskNewClusterAzureAttributesArgs']] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + cluster_log_conf: Optional[pulumi.Input['JobTaskNewClusterClusterLogConfArgs']] = None, + cluster_mount_infos: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskNewClusterClusterMountInfoArgs']]]] = None, + cluster_name: Optional[pulumi.Input[str]] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + data_security_mode: Optional[pulumi.Input[str]] = None, + docker_image: Optional[pulumi.Input['JobTaskNewClusterDockerImageArgs']] = None, + driver_instance_pool_id: Optional[pulumi.Input[str]] = None, + driver_node_type_id: Optional[pulumi.Input[str]] = None, + enable_elastic_disk: Optional[pulumi.Input[bool]] = None, + enable_local_disk_encryption: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['JobTaskNewClusterGcpAttributesArgs']] = None, + idempotency_token: Optional[pulumi.Input[str]] = None, + init_scripts: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskNewClusterInitScriptArgs']]]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + num_workers: Optional[pulumi.Input[int]] = None, + policy_id: Optional[pulumi.Input[str]] = None, + runtime_engine: Optional[pulumi.Input[str]] = None, + single_user_name: Optional[pulumi.Input[str]] = None, + spark_conf: Optional[pulumi.Input[Mapping[str, Any]]] = None, + spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, + ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + workload_type: Optional[pulumi.Input['JobTaskNewClusterWorkloadTypeArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="sparkVersion") @@ -7079,10 +8829,21 @@ class JobTaskNewClusterAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[pulumi.Input[int]] = None, min_workers: Optional[pulumi.Input[int]] = None): + JobTaskNewClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[pulumi.Input[int]] = None, + min_workers: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -7114,22 +8875,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[pulumi.Input[str]] = None, spot_bid_price_percent: Optional[pulumi.Input[int]] = None, zone_id: Optional[pulumi.Input[str]] = None): + JobTaskNewClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + ebs_volume_count: Optional[pulumi.Input[int]] = None, + ebs_volume_size: Optional[pulumi.Input[int]] = None, + ebs_volume_type: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + spot_bid_price_percent: Optional[pulumi.Input[int]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -7210,12 +8994,25 @@ def __init__(__self__, *, availability: Optional[pulumi.Input[str]] = None, first_on_demand: Optional[pulumi.Input[int]] = None, spot_bid_max_price: Optional[pulumi.Input[float]] = None): + JobTaskNewClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + spot_bid_max_price: Optional[pulumi.Input[float]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -7250,10 +9047,21 @@ class JobTaskNewClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional[pulumi.Input['JobTaskNewClusterClusterLogConfDbfsArgs']] = None, s3: Optional[pulumi.Input['JobTaskNewClusterClusterLogConfS3Args']] = None): + JobTaskNewClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional[pulumi.Input['JobTaskNewClusterClusterLogConfDbfsArgs']] = None, + s3: Optional[pulumi.Input['JobTaskNewClusterClusterLogConfS3Args']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -7278,7 +9086,16 @@ def s3(self, value: Optional[pulumi.Input['JobTaskNewClusterClusterLogConfS3Args class JobTaskNewClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -7300,19 +9117,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -7384,10 +9222,23 @@ def __init__(__self__, *, local_mount_dir_path: pulumi.Input[str], network_filesystem_info: pulumi.Input['JobTaskNewClusterClusterMountInfoNetworkFilesystemInfoArgs'], remote_mount_dir_path: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + JobTaskNewClusterClusterMountInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: pulumi.Input[str], + network_filesystem_info: pulumi.Input['JobTaskNewClusterClusterMountInfoNetworkFilesystemInfoArgs'], + remote_mount_dir_path: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -7422,9 +9273,20 @@ class JobTaskNewClusterClusterMountInfoNetworkFilesystemInfoArgs: def __init__(__self__, *, server_address: pulumi.Input[str], mount_options: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "server_address", server_address) + JobTaskNewClusterClusterMountInfoNetworkFilesystemInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: pulumi.Input[str], + mount_options: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -7451,17 +9313,28 @@ def __init__(__self__, *, url: pulumi.Input[str], basic_auth: Optional[pulumi.Input['JobTaskNewClusterDockerImageBasicAuthArgs']] = None): """ - :param pulumi.Input[str] url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). - """ - pulumi.set(__self__, "url", url) + :param pulumi.Input[str] url: URL of the job on the given workspace + """ + JobTaskNewClusterDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + basic_auth: Optional[pulumi.Input['JobTaskNewClusterDockerImageBasicAuthArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter def url(self) -> pulumi.Input[str]: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -7484,8 +9357,19 @@ class JobTaskNewClusterDockerImageBasicAuthArgs: def __init__(__self__, *, password: pulumi.Input[str], username: pulumi.Input[str]): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + JobTaskNewClusterDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: pulumi.Input[str], + username: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -7515,18 +9399,37 @@ def __init__(__self__, *, local_ssd_count: Optional[pulumi.Input[int]] = None, use_preemptible_executors: Optional[pulumi.Input[bool]] = None, zone_id: Optional[pulumi.Input[str]] = None): + JobTaskNewClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + boot_disk_size: Optional[pulumi.Input[int]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + local_ssd_count: Optional[pulumi.Input[int]] = None, + use_preemptible_executors: Optional[pulumi.Input[bool]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -7639,23 +9542,44 @@ def __init__(__self__, *, ]) ``` """ + JobTaskNewClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional[pulumi.Input['JobTaskNewClusterInitScriptAbfssArgs']] = None, + dbfs: Optional[pulumi.Input['JobTaskNewClusterInitScriptDbfsArgs']] = None, + file: Optional[pulumi.Input['JobTaskNewClusterInitScriptFileArgs']] = None, + gcs: Optional[pulumi.Input['JobTaskNewClusterInitScriptGcsArgs']] = None, + s3: Optional[pulumi.Input['JobTaskNewClusterInitScriptS3Args']] = None, + volumes: Optional[pulumi.Input['JobTaskNewClusterInitScriptVolumesArgs']] = None, + workspace: Optional[pulumi.Input['JobTaskNewClusterInitScriptWorkspaceArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: warnings.warn("""For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""", DeprecationWarning) pulumi.log.warn("""dbfs is deprecated: For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""") if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -7774,8 +9698,17 @@ def workspace(self, value: Optional[pulumi.Input['JobTaskNewClusterInitScriptWor class JobTaskNewClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobTaskNewClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7791,7 +9724,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobTaskNewClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -7807,8 +9749,17 @@ def destination(self, value: pulumi.Input[str]): class JobTaskNewClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobTaskNewClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7824,8 +9775,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobTaskNewClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobTaskNewClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7847,19 +9807,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -7929,8 +9910,17 @@ def region(self, value: Optional[pulumi.Input[str]]): class JobTaskNewClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobTaskNewClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7946,8 +9936,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobTaskNewClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + JobTaskNewClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7963,7 +9962,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class JobTaskNewClusterWorkloadTypeArgs: def __init__(__self__, *, clients: pulumi.Input['JobTaskNewClusterWorkloadTypeClientsArgs']): - pulumi.set(__self__, "clients", clients) + JobTaskNewClusterWorkloadTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: pulumi.Input['JobTaskNewClusterWorkloadTypeClientsArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -7980,10 +9988,21 @@ class JobTaskNewClusterWorkloadTypeClientsArgs: def __init__(__self__, *, jobs: Optional[pulumi.Input[bool]] = None, notebooks: Optional[pulumi.Input[bool]] = None): + JobTaskNewClusterWorkloadTypeClientsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[pulumi.Input[bool]] = None, + notebooks: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -8015,11 +10034,24 @@ def __init__(__self__, *, :param pulumi.Input[Mapping[str, Any]] base_parameters: (Map) Base parameters to be used for each run of this job. If the run is initiated by a call to run-now with parameters specified, the two parameters maps will be merged. If the same key is specified in base_parameters and in run-now, the value from run-now will be used. If the notebook takes a parameter that is not specified in the job’s base_parameters or the run-now override parameters, the default value from the notebook will be used. Retrieve these parameters in a notebook using `dbutils.widgets.get`. :param pulumi.Input[str] source: Location type of the notebook, can only be `WORKSPACE` or `GIT`. When set to `WORKSPACE`, the notebook will be retrieved from the local Databricks workspace. When set to `GIT`, the notebook will be retrieved from a Git repository defined in `git_source`. If the value is empty, the task will use `GIT` if `git_source` is defined and `WORKSPACE` otherwise. """ - pulumi.set(__self__, "notebook_path", notebook_path) + JobTaskNotebookTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: pulumi.Input[str], + base_parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -8069,12 +10101,25 @@ def __init__(__self__, *, :param pulumi.Input[bool] no_alert_for_canceled_runs: (Bool) don't send alert for cancelled runs. :param pulumi.Input[bool] no_alert_for_skipped_runs: (Bool) don't send alert for skipped runs. """ + JobTaskNotificationSettingsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[pulumi.Input[bool]] = None, + no_alert_for_canceled_runs: Optional[pulumi.Input[bool]] = None, + no_alert_for_skipped_runs: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="alertOnLastAttempt") @@ -8124,9 +10169,20 @@ def __init__(__self__, *, > **Note** The following configuration blocks are only supported inside a `task` block """ - pulumi.set(__self__, "pipeline_id", pipeline_id) + JobTaskPipelineTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: pulumi.Input[str], + full_refresh: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -8168,14 +10224,29 @@ def __init__(__self__, *, :param pulumi.Input[str] package_name: Name of Python package :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: Parameters for the task """ + JobTaskPythonWheelTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[pulumi.Input[str]] = None, + named_parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, + package_name: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -8235,9 +10306,20 @@ def __init__(__self__, *, :param pulumi.Input[int] job_id: (String) ID of the job :param pulumi.Input[Mapping[str, Any]] job_parameters: (Map) Job parameters for the task """ - pulumi.set(__self__, "job_id", job_id) + JobTaskRunJobTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: pulumi.Input[int], + job_parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -8274,12 +10356,25 @@ def __init__(__self__, *, :param pulumi.Input[str] main_class_name: The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. The code should use `SparkContext.getOrCreate` to obtain a Spark context; otherwise, runs of the job will fail. :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: (List) Parameters passed to the main method. """ + JobTaskSparkJarTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[pulumi.Input[str]] = None, + main_class_name: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -8326,11 +10421,24 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: (List) Command line parameters passed to the Python file. :param pulumi.Input[str] source: Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `git_source`. """ - pulumi.set(__self__, "python_file", python_file) + JobTaskSparkPythonTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: pulumi.Input[str], + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -8376,8 +10484,17 @@ def __init__(__self__, *, """ :param pulumi.Input[Sequence[pulumi.Input[str]]] parameters: (List) Command-line parameters passed to spark submit. """ + JobTaskSparkSubmitTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -8452,18 +10569,37 @@ def __init__(__self__, *, :param pulumi.Input['JobTaskSqlTaskQueryArgs'] query: block consisting of single string field: `query_id` - identifier of the Databricks SQL Query (databricks_sql_query). :param pulumi.Input[str] warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. """ + JobTaskSqlTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert=alert, + dashboard=dashboard, + file=file, + parameters=parameters, + query=query, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert: Optional[pulumi.Input['JobTaskSqlTaskAlertArgs']] = None, + dashboard: Optional[pulumi.Input['JobTaskSqlTaskDashboardArgs']] = None, + file: Optional[pulumi.Input['JobTaskSqlTaskFileArgs']] = None, + parameters: Optional[pulumi.Input[Mapping[str, Any]]] = None, + query: Optional[pulumi.Input['JobTaskSqlTaskQueryArgs']] = None, + warehouse_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert is not None: - pulumi.set(__self__, "alert", alert) + _setter("alert", alert) if dashboard is not None: - pulumi.set(__self__, "dashboard", dashboard) + _setter("dashboard", dashboard) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if query is not None: - pulumi.set(__self__, "query", query) + _setter("query", query) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -8592,10 +10728,23 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input['JobTaskSqlTaskAlertSubscriptionArgs']]] subscriptions: a list of subscription blocks consisting out of one of the required fields: `user_name` for user emails or `destination_id` - for Alert destination's identifier. :param pulumi.Input[bool] pause_subscriptions: flag that specifies if subscriptions are paused or not. """ - pulumi.set(__self__, "alert_id", alert_id) - pulumi.set(__self__, "subscriptions", subscriptions) + JobTaskSqlTaskAlertArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_id=alert_id, + subscriptions=subscriptions, + pause_subscriptions=pause_subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_id: pulumi.Input[str], + subscriptions: pulumi.Input[Sequence[pulumi.Input['JobTaskSqlTaskAlertSubscriptionArgs']]], + pause_subscriptions: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("alert_id", alert_id) + _setter("subscriptions", subscriptions) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) @property @pulumi.getter(name="alertId") @@ -8642,10 +10791,21 @@ def __init__(__self__, *, """ :param pulumi.Input[str] user_name: The email of an active workspace user. Non-admin users can only set this field to their own email. """ + JobTaskSqlTaskAlertSubscriptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[pulumi.Input[str]] = None, + user_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -8682,13 +10842,28 @@ def __init__(__self__, *, :param pulumi.Input[bool] pause_subscriptions: flag that specifies if subscriptions are paused or not. :param pulumi.Input[Sequence[pulumi.Input['JobTaskSqlTaskDashboardSubscriptionArgs']]] subscriptions: a list of subscription blocks consisting out of one of the required fields: `user_name` for user emails or `destination_id` - for Alert destination's identifier. """ - pulumi.set(__self__, "dashboard_id", dashboard_id) + JobTaskSqlTaskDashboardArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dashboard_id=dashboard_id, + custom_subject=custom_subject, + pause_subscriptions=pause_subscriptions, + subscriptions=subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dashboard_id: pulumi.Input[str], + custom_subject: Optional[pulumi.Input[str]] = None, + pause_subscriptions: Optional[pulumi.Input[bool]] = None, + subscriptions: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskSqlTaskDashboardSubscriptionArgs']]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dashboard_id", dashboard_id) if custom_subject is not None: - pulumi.set(__self__, "custom_subject", custom_subject) + _setter("custom_subject", custom_subject) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) if subscriptions is not None: - pulumi.set(__self__, "subscriptions", subscriptions) + _setter("subscriptions", subscriptions) @property @pulumi.getter(name="dashboardId") @@ -8747,10 +10922,21 @@ def __init__(__self__, *, """ :param pulumi.Input[str] user_name: The email of an active workspace user. Non-admin users can only set this field to their own email. """ + JobTaskSqlTaskDashboardSubscriptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[pulumi.Input[str]] = None, + user_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -8778,7 +10964,16 @@ def user_name(self, value: Optional[pulumi.Input[str]]): class JobTaskSqlTaskFileArgs: def __init__(__self__, *, path: pulumi.Input[str]): - pulumi.set(__self__, "path", path) + JobTaskSqlTaskFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -8794,7 +10989,16 @@ def path(self, value: pulumi.Input[str]): class JobTaskSqlTaskQueryArgs: def __init__(__self__, *, query_id: pulumi.Input[str]): - pulumi.set(__self__, "query_id", query_id) + JobTaskSqlTaskQueryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + query_id=query_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + query_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("query_id", query_id) @property @pulumi.getter(name="queryId") @@ -8815,9 +11019,20 @@ def __init__(__self__, *, :param pulumi.Input['JobTriggerFileArrivalArgs'] file_arrival: configuration block to define a trigger for [File Arrival events](https://learn.microsoft.com/en-us/azure/databricks/workflows/jobs/file-arrival-triggers) consisting of following attributes: :param pulumi.Input[str] pause_status: Indicate whether this trigger is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pause_status`. """ - pulumi.set(__self__, "file_arrival", file_arrival) + JobTriggerArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + file_arrival=file_arrival, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + file_arrival: pulumi.Input['JobTriggerFileArrivalArgs'], + pause_status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("file_arrival", file_arrival) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="fileArrival") @@ -8851,21 +11066,34 @@ def __init__(__self__, *, min_time_between_triggers_seconds: Optional[pulumi.Input[int]] = None, wait_after_last_change_seconds: Optional[pulumi.Input[int]] = None): """ - :param pulumi.Input[str] url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + :param pulumi.Input[str] url: URL of the job on the given workspace :param pulumi.Input[int] min_time_between_triggers_seconds: If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. :param pulumi.Input[int] wait_after_last_change_seconds: If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. """ - pulumi.set(__self__, "url", url) + JobTriggerFileArrivalArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + min_time_between_triggers_seconds=min_time_between_triggers_seconds, + wait_after_last_change_seconds=wait_after_last_change_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + min_time_between_triggers_seconds: Optional[pulumi.Input[int]] = None, + wait_after_last_change_seconds: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if min_time_between_triggers_seconds is not None: - pulumi.set(__self__, "min_time_between_triggers_seconds", min_time_between_triggers_seconds) + _setter("min_time_between_triggers_seconds", min_time_between_triggers_seconds) if wait_after_last_change_seconds is not None: - pulumi.set(__self__, "wait_after_last_change_seconds", wait_after_last_change_seconds) + _setter("wait_after_last_change_seconds", wait_after_last_change_seconds) @property @pulumi.getter def url(self) -> pulumi.Input[str]: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -8919,14 +11147,29 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input['JobWebhookNotificationsOnStartArgs']]] on_starts: (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified. :param pulumi.Input[Sequence[pulumi.Input['JobWebhookNotificationsOnSuccessArgs']]] on_successes: (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified. """ + JobWebhookNotificationsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + on_duration_warning_threshold_exceededs: Optional[pulumi.Input[Sequence[pulumi.Input['JobWebhookNotificationsOnDurationWarningThresholdExceededArgs']]]] = None, + on_failures: Optional[pulumi.Input[Sequence[pulumi.Input['JobWebhookNotificationsOnFailureArgs']]]] = None, + on_starts: Optional[pulumi.Input[Sequence[pulumi.Input['JobWebhookNotificationsOnStartArgs']]]] = None, + on_successes: Optional[pulumi.Input[Sequence[pulumi.Input['JobWebhookNotificationsOnSuccessArgs']]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="onDurationWarningThresholdExceededs") @@ -8990,19 +11233,24 @@ class JobWebhookNotificationsOnDurationWarningThresholdExceededArgs: def __init__(__self__, *, id: pulumi.Input[str]): """ - :param pulumi.Input[str] id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param pulumi.Input[str] id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnDurationWarningThresholdExceededArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> pulumi.Input[str]: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -9016,19 +11264,24 @@ class JobWebhookNotificationsOnFailureArgs: def __init__(__self__, *, id: pulumi.Input[str]): """ - :param pulumi.Input[str] id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param pulumi.Input[str] id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnFailureArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> pulumi.Input[str]: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -9042,19 +11295,24 @@ class JobWebhookNotificationsOnStartArgs: def __init__(__self__, *, id: pulumi.Input[str]): """ - :param pulumi.Input[str] id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param pulumi.Input[str] id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnStartArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> pulumi.Input[str]: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -9068,19 +11326,24 @@ class JobWebhookNotificationsOnSuccessArgs: def __init__(__self__, *, id: pulumi.Input[str]): """ - :param pulumi.Input[str] id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param pulumi.Input[str] id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnSuccessArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> pulumi.Input[str]: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -9094,9 +11357,20 @@ class LibraryCranArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + LibraryCranArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -9123,11 +11397,24 @@ def __init__(__self__, *, coordinates: pulumi.Input[str], exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "coordinates", coordinates) + LibraryMavenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: pulumi.Input[str], + exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -9162,9 +11449,20 @@ class LibraryPypiArgs: def __init__(__self__, *, package: pulumi.Input[str], repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "package", package) + LibraryPypiArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: pulumi.Input[str], + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -9194,7 +11492,16 @@ def __init__(__self__, *, `azure_managed_identity` optional configuration block for using managed identity as credential details for Azure (Recommended): """ - pulumi.set(__self__, "role_arn", role_arn) + MetastoreDataAccessAwsIamRoleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + role_arn=role_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role_arn: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role_arn", role_arn) @property @pulumi.getter(name="roleArn") @@ -9223,11 +11530,24 @@ def __init__(__self__, *, `databricks_gcp_service_account` optional configuration block for creating a Databricks-managed GCP Service Account: """ - pulumi.set(__self__, "access_connector_id", access_connector_id) + MetastoreDataAccessAzureManagedIdentityArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_connector_id=access_connector_id, + credential_id=credential_id, + managed_identity_id=managed_identity_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_connector_id: pulumi.Input[str], + credential_id: Optional[pulumi.Input[str]] = None, + managed_identity_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("access_connector_id", access_connector_id) if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if managed_identity_id is not None: - pulumi.set(__self__, "managed_identity_id", managed_identity_id) + _setter("managed_identity_id", managed_identity_id) @property @pulumi.getter(name="accessConnectorId") @@ -9276,9 +11596,22 @@ def __init__(__self__, *, :param pulumi.Input[str] client_secret: The client secret generated for the above app ID in AAD. **This field is redacted on output** :param pulumi.Input[str] directory_id: The directory ID corresponding to the Azure Active Directory (AAD) tenant of the application """ - pulumi.set(__self__, "application_id", application_id) - pulumi.set(__self__, "client_secret", client_secret) - pulumi.set(__self__, "directory_id", directory_id) + MetastoreDataAccessAzureServicePrincipalArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + application_id=application_id, + client_secret=client_secret, + directory_id=directory_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + application_id: pulumi.Input[str], + client_secret: pulumi.Input[str], + directory_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("application_id", application_id) + _setter("client_secret", client_secret) + _setter("directory_id", directory_id) @property @pulumi.getter(name="applicationId") @@ -9327,10 +11660,21 @@ def __init__(__self__, *, `azure_service_principal` optional configuration block for credential details for Azure (Legacy): """ + MetastoreDataAccessDatabricksGcpServiceAccountArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + credential_id=credential_id, + email=email, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + credential_id: Optional[pulumi.Input[str]] = None, + email: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if email is not None: - pulumi.set(__self__, "email", email) + _setter("email", email) @property @pulumi.getter(name="credentialId") @@ -9367,9 +11711,22 @@ def __init__(__self__, *, `azure_service_principal` optional configuration block for credential details for Azure (Legacy): """ - pulumi.set(__self__, "email", email) - pulumi.set(__self__, "private_key", private_key) - pulumi.set(__self__, "private_key_id", private_key_id) + MetastoreDataAccessGcpServiceAccountKeyArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + email=email, + private_key=private_key, + private_key_id=private_key_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + email: pulumi.Input[str], + private_key: pulumi.Input[str], + private_key_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("email", email) + _setter("private_key", private_key) + _setter("private_key_id", private_key_id) @property @pulumi.getter @@ -9409,10 +11766,21 @@ class MlflowModelTagArgs: def __init__(__self__, *, key: Optional[pulumi.Input[str]] = None, value: Optional[pulumi.Input[str]] = None): + MlflowModelTagArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: Optional[pulumi.Input[str]] = None, + value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if key is not None: - pulumi.set(__self__, "key", key) + _setter("key", key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -9446,13 +11814,28 @@ def __init__(__self__, *, :param pulumi.Input[bool] enable_ssl_verification: Enable/disable SSL certificate validation. Default is `true`. For self-signed certificates, this field must be `false` AND the destination server must disable certificate validation as well. For security purposes, it is encouraged to perform secret validation with the HMAC-encoded portion of the payload and acknowledge the risk associated with disabling hostname validation whereby it becomes more likely that requests can be maliciously routed to an unintended host. :param pulumi.Input[str] secret: Shared secret required for HMAC encoding payload. The HMAC-encoded payload will be sent in the header as `X-Databricks-Signature: encoded_payload`. """ - pulumi.set(__self__, "url", url) + MlflowWebhookHttpUrlSpecArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + authorization=authorization, + enable_ssl_verification=enable_ssl_verification, + secret=secret, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + authorization: Optional[pulumi.Input[str]] = None, + enable_ssl_verification: Optional[pulumi.Input[bool]] = None, + secret: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if authorization is not None: - pulumi.set(__self__, "authorization", authorization) + _setter("authorization", authorization) if enable_ssl_verification is not None: - pulumi.set(__self__, "enable_ssl_verification", enable_ssl_verification) + _setter("enable_ssl_verification", enable_ssl_verification) if secret is not None: - pulumi.set(__self__, "secret", secret) + _setter("secret", secret) @property @pulumi.getter @@ -9514,10 +11897,23 @@ def __init__(__self__, *, :param pulumi.Input[str] job_id: ID of the Databricks job that the webhook runs. :param pulumi.Input[str] workspace_url: URL of the workspace containing the job that this webhook runs. If not specified, the job’s workspace URL is assumed to be the same as the workspace where the webhook is created. """ - pulumi.set(__self__, "access_token", access_token) - pulumi.set(__self__, "job_id", job_id) + MlflowWebhookJobSpecArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_token=access_token, + job_id=job_id, + workspace_url=workspace_url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_token: pulumi.Input[str], + job_id: pulumi.Input[str], + workspace_url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("access_token", access_token) + _setter("job_id", job_id) if workspace_url is not None: - pulumi.set(__self__, "workspace_url", workspace_url) + _setter("workspace_url", workspace_url) @property @pulumi.getter(name="accessToken") @@ -9565,9 +11961,20 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input['ModelServingConfigServedModelArgs']]] served_models: Each block represents a served model for the endpoint to serve. A model serving endpoint can have up to 10 served models. :param pulumi.Input['ModelServingConfigTrafficConfigArgs'] traffic_config: A single block represents the traffic split configuration amongst the served models. """ - pulumi.set(__self__, "served_models", served_models) + ModelServingConfigArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + served_models=served_models, + traffic_config=traffic_config, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + served_models: pulumi.Input[Sequence[pulumi.Input['ModelServingConfigServedModelArgs']]], + traffic_config: Optional[pulumi.Input['ModelServingConfigTrafficConfigArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("served_models", served_models) if traffic_config is not None: - pulumi.set(__self__, "traffic_config", traffic_config) + _setter("traffic_config", traffic_config) @property @pulumi.getter(name="servedModels") @@ -9613,17 +12020,38 @@ def __init__(__self__, *, :param pulumi.Input[str] name: The name of a served model. It must be unique across an endpoint. If not specified, this field will default to `modelname-modelversion`. A served model name can consist of alphanumeric characters, dashes, and underscores. :param pulumi.Input[bool] scale_to_zero_enabled: Whether the compute resources for the served model should scale down to zero. If scale-to-zero is enabled, the lower bound of the provisioned concurrency for each workload size will be 0. The default value is `true`. """ - pulumi.set(__self__, "model_name", model_name) - pulumi.set(__self__, "model_version", model_version) - pulumi.set(__self__, "workload_size", workload_size) + ModelServingConfigServedModelArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + model_name=model_name, + model_version=model_version, + workload_size=workload_size, + environment_vars=environment_vars, + instance_profile_arn=instance_profile_arn, + name=name, + scale_to_zero_enabled=scale_to_zero_enabled, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + model_name: pulumi.Input[str], + model_version: pulumi.Input[str], + workload_size: pulumi.Input[str], + environment_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + scale_to_zero_enabled: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("model_name", model_name) + _setter("model_version", model_version) + _setter("workload_size", workload_size) if environment_vars is not None: - pulumi.set(__self__, "environment_vars", environment_vars) + _setter("environment_vars", environment_vars) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if scale_to_zero_enabled is not None: - pulumi.set(__self__, "scale_to_zero_enabled", scale_to_zero_enabled) + _setter("scale_to_zero_enabled", scale_to_zero_enabled) @property @pulumi.getter(name="modelName") @@ -9717,8 +12145,17 @@ def __init__(__self__, *, """ :param pulumi.Input[Sequence[pulumi.Input['ModelServingConfigTrafficConfigRouteArgs']]] routes: Each block represents a route that defines traffic to each served model. Each `served_models` block needs to have a corresponding `routes` block """ + ModelServingConfigTrafficConfigArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + routes=routes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + routes: Optional[pulumi.Input[Sequence[pulumi.Input['ModelServingConfigTrafficConfigRouteArgs']]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if routes is not None: - pulumi.set(__self__, "routes", routes) + _setter("routes", routes) @property @pulumi.getter @@ -9742,8 +12179,19 @@ def __init__(__self__, *, :param pulumi.Input[str] served_model_name: The name of the served model this route configures traffic for. This needs to match the name of a `served_models` block :param pulumi.Input[int] traffic_percentage: The percentage of endpoint traffic to send to this route. It must be an integer between 0 and 100 inclusive. """ - pulumi.set(__self__, "served_model_name", served_model_name) - pulumi.set(__self__, "traffic_percentage", traffic_percentage) + ModelServingConfigTrafficConfigRouteArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + served_model_name=served_model_name, + traffic_percentage=traffic_percentage, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + served_model_name: pulumi.Input[str], + traffic_percentage: pulumi.Input[int], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("served_model_name", served_model_name) + _setter("traffic_percentage", traffic_percentage) @property @pulumi.getter(name="servedModelName") @@ -9775,9 +12223,20 @@ class ModelServingTagArgs: def __init__(__self__, *, key: pulumi.Input[str], value: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "key", key) + ModelServingTagArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: pulumi.Input[str], + value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key", key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -9809,18 +12268,41 @@ def __init__(__self__, *, directory: Optional[pulumi.Input[str]] = None, storage_account_name: Optional[pulumi.Input[str]] = None, tenant_id: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "client_id", client_id) - pulumi.set(__self__, "client_secret_key", client_secret_key) - pulumi.set(__self__, "client_secret_scope", client_secret_scope) - pulumi.set(__self__, "initialize_file_system", initialize_file_system) + MountAbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + client_id=client_id, + client_secret_key=client_secret_key, + client_secret_scope=client_secret_scope, + initialize_file_system=initialize_file_system, + container_name=container_name, + directory=directory, + storage_account_name=storage_account_name, + tenant_id=tenant_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + client_id: pulumi.Input[str], + client_secret_key: pulumi.Input[str], + client_secret_scope: pulumi.Input[str], + initialize_file_system: pulumi.Input[bool], + container_name: Optional[pulumi.Input[str]] = None, + directory: Optional[pulumi.Input[str]] = None, + storage_account_name: Optional[pulumi.Input[str]] = None, + tenant_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("client_id", client_id) + _setter("client_secret_key", client_secret_key) + _setter("client_secret_scope", client_secret_scope) + _setter("initialize_file_system", initialize_file_system) if container_name is not None: - pulumi.set(__self__, "container_name", container_name) + _setter("container_name", container_name) if directory is not None: - pulumi.set(__self__, "directory", directory) + _setter("directory", directory) if storage_account_name is not None: - pulumi.set(__self__, "storage_account_name", storage_account_name) + _setter("storage_account_name", storage_account_name) if tenant_id is not None: - pulumi.set(__self__, "tenant_id", tenant_id) + _setter("tenant_id", tenant_id) @property @pulumi.getter(name="clientId") @@ -9905,17 +12387,38 @@ def __init__(__self__, *, spark_conf_prefix: Optional[pulumi.Input[str]] = None, storage_resource_name: Optional[pulumi.Input[str]] = None, tenant_id: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "client_id", client_id) - pulumi.set(__self__, "client_secret_key", client_secret_key) - pulumi.set(__self__, "client_secret_scope", client_secret_scope) + MountAdlArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + client_id=client_id, + client_secret_key=client_secret_key, + client_secret_scope=client_secret_scope, + directory=directory, + spark_conf_prefix=spark_conf_prefix, + storage_resource_name=storage_resource_name, + tenant_id=tenant_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + client_id: pulumi.Input[str], + client_secret_key: pulumi.Input[str], + client_secret_scope: pulumi.Input[str], + directory: Optional[pulumi.Input[str]] = None, + spark_conf_prefix: Optional[pulumi.Input[str]] = None, + storage_resource_name: Optional[pulumi.Input[str]] = None, + tenant_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("client_id", client_id) + _setter("client_secret_key", client_secret_key) + _setter("client_secret_scope", client_secret_scope) if directory is not None: - pulumi.set(__self__, "directory", directory) + _setter("directory", directory) if spark_conf_prefix is not None: - pulumi.set(__self__, "spark_conf_prefix", spark_conf_prefix) + _setter("spark_conf_prefix", spark_conf_prefix) if storage_resource_name is not None: - pulumi.set(__self__, "storage_resource_name", storage_resource_name) + _setter("storage_resource_name", storage_resource_name) if tenant_id is not None: - pulumi.set(__self__, "tenant_id", tenant_id) + _setter("tenant_id", tenant_id) @property @pulumi.getter(name="clientId") @@ -9986,9 +12489,20 @@ class MountGsArgs: def __init__(__self__, *, bucket_name: pulumi.Input[str], service_account: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "bucket_name", bucket_name) + MountGsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + bucket_name=bucket_name, + service_account=service_account, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + bucket_name: pulumi.Input[str], + service_account: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("bucket_name", bucket_name) if service_account is not None: - pulumi.set(__self__, "service_account", service_account) + _setter("service_account", service_account) @property @pulumi.getter(name="bucketName") @@ -10014,9 +12528,20 @@ class MountS3Args: def __init__(__self__, *, bucket_name: pulumi.Input[str], instance_profile: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "bucket_name", bucket_name) + MountS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + bucket_name=bucket_name, + instance_profile=instance_profile, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + bucket_name: pulumi.Input[str], + instance_profile: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("bucket_name", bucket_name) if instance_profile is not None: - pulumi.set(__self__, "instance_profile", instance_profile) + _setter("instance_profile", instance_profile) @property @pulumi.getter(name="bucketName") @@ -10046,15 +12571,34 @@ def __init__(__self__, *, container_name: Optional[pulumi.Input[str]] = None, directory: Optional[pulumi.Input[str]] = None, storage_account_name: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "auth_type", auth_type) - pulumi.set(__self__, "token_secret_key", token_secret_key) - pulumi.set(__self__, "token_secret_scope", token_secret_scope) + MountWasbArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + auth_type=auth_type, + token_secret_key=token_secret_key, + token_secret_scope=token_secret_scope, + container_name=container_name, + directory=directory, + storage_account_name=storage_account_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + auth_type: pulumi.Input[str], + token_secret_key: pulumi.Input[str], + token_secret_scope: pulumi.Input[str], + container_name: Optional[pulumi.Input[str]] = None, + directory: Optional[pulumi.Input[str]] = None, + storage_account_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("auth_type", auth_type) + _setter("token_secret_key", token_secret_key) + _setter("token_secret_scope", token_secret_scope) if container_name is not None: - pulumi.set(__self__, "container_name", container_name) + _setter("container_name", container_name) if directory is not None: - pulumi.set(__self__, "directory", directory) + _setter("directory", directory) if storage_account_name is not None: - pulumi.set(__self__, "storage_account_name", storage_account_name) + _setter("storage_account_name", storage_account_name) @property @pulumi.getter(name="authType") @@ -10122,10 +12666,23 @@ def __init__(__self__, *, :param pulumi.Input[str] key_arn: The AWS KMS key's Amazon Resource Name (ARN). :param pulumi.Input[str] key_region: (Computed) The AWS region in which KMS key is deployed to. This is not required. """ - pulumi.set(__self__, "key_alias", key_alias) - pulumi.set(__self__, "key_arn", key_arn) + MwsCustomerManagedKeysAwsKeyInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + key_alias=key_alias, + key_arn=key_arn, + key_region=key_region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key_alias: pulumi.Input[str], + key_arn: pulumi.Input[str], + key_region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key_alias", key_alias) + _setter("key_arn", key_arn) if key_region is not None: - pulumi.set(__self__, "key_region", key_region) + _setter("key_region", key_region) @property @pulumi.getter(name="keyAlias") @@ -10171,7 +12728,16 @@ def __init__(__self__, *, """ :param pulumi.Input[str] kms_key_id: The GCP KMS key's resource name. """ - pulumi.set(__self__, "kms_key_id", kms_key_id) + MwsCustomerManagedKeysGcpKeyInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + kms_key_id=kms_key_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + kms_key_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("kms_key_id", kms_key_id) @property @pulumi.getter(name="kmsKeyId") @@ -10191,10 +12757,21 @@ class MwsNetworksErrorMessageArgs: def __init__(__self__, *, error_message: Optional[pulumi.Input[str]] = None, error_type: Optional[pulumi.Input[str]] = None): + MwsNetworksErrorMessageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + error_message=error_message, + error_type=error_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + error_message: Optional[pulumi.Input[str]] = None, + error_type: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if error_message is not None: - pulumi.set(__self__, "error_message", error_message) + _setter("error_message", error_message) if error_type is not None: - pulumi.set(__self__, "error_type", error_type) + _setter("error_type", error_type) @property @pulumi.getter(name="errorMessage") @@ -10232,12 +12809,31 @@ def __init__(__self__, *, :param pulumi.Input[str] subnet_region: The Google Cloud region of the workspace data plane. For example, `us-east4`. :param pulumi.Input[str] vpc_id: The ID of the VPC associated with this network. VPC IDs can be used in multiple network configurations. """ - pulumi.set(__self__, "network_project_id", network_project_id) - pulumi.set(__self__, "pod_ip_range_name", pod_ip_range_name) - pulumi.set(__self__, "service_ip_range_name", service_ip_range_name) - pulumi.set(__self__, "subnet_id", subnet_id) - pulumi.set(__self__, "subnet_region", subnet_region) - pulumi.set(__self__, "vpc_id", vpc_id) + MwsNetworksGcpNetworkInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + network_project_id=network_project_id, + pod_ip_range_name=pod_ip_range_name, + service_ip_range_name=service_ip_range_name, + subnet_id=subnet_id, + subnet_region=subnet_region, + vpc_id=vpc_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + network_project_id: pulumi.Input[str], + pod_ip_range_name: pulumi.Input[str], + service_ip_range_name: pulumi.Input[str], + subnet_id: pulumi.Input[str], + subnet_region: pulumi.Input[str], + vpc_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("network_project_id", network_project_id) + _setter("pod_ip_range_name", pod_ip_range_name) + _setter("service_ip_range_name", service_ip_range_name) + _setter("subnet_id", subnet_id) + _setter("subnet_region", subnet_region) + _setter("vpc_id", vpc_id) @property @pulumi.getter(name="networkProjectId") @@ -10317,8 +12913,19 @@ class MwsNetworksVpcEndpointsArgs: def __init__(__self__, *, dataplane_relays: pulumi.Input[Sequence[pulumi.Input[str]]], rest_apis: pulumi.Input[Sequence[pulumi.Input[str]]]): - pulumi.set(__self__, "dataplane_relays", dataplane_relays) - pulumi.set(__self__, "rest_apis", rest_apis) + MwsNetworksVpcEndpointsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dataplane_relays=dataplane_relays, + rest_apis=rest_apis, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dataplane_relays: pulumi.Input[Sequence[pulumi.Input[str]]], + rest_apis: pulumi.Input[Sequence[pulumi.Input[str]]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dataplane_relays", dataplane_relays) + _setter("rest_apis", rest_apis) @property @pulumi.getter(name="dataplaneRelays") @@ -10354,13 +12961,30 @@ def __init__(__self__, *, :param pulumi.Input[str] psc_connection_id: The unique ID of this PSC connection. :param pulumi.Input[str] service_attachment_id: The service attachment this PSC connection connects to. """ - pulumi.set(__self__, "endpoint_region", endpoint_region) - pulumi.set(__self__, "project_id", project_id) - pulumi.set(__self__, "psc_endpoint_name", psc_endpoint_name) + MwsVpcEndpointGcpVpcEndpointInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + endpoint_region=endpoint_region, + project_id=project_id, + psc_endpoint_name=psc_endpoint_name, + psc_connection_id=psc_connection_id, + service_attachment_id=service_attachment_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + endpoint_region: pulumi.Input[str], + project_id: pulumi.Input[str], + psc_endpoint_name: pulumi.Input[str], + psc_connection_id: Optional[pulumi.Input[str]] = None, + service_attachment_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("endpoint_region", endpoint_region) + _setter("project_id", project_id) + _setter("psc_endpoint_name", psc_endpoint_name) if psc_connection_id is not None: - pulumi.set(__self__, "psc_connection_id", psc_connection_id) + _setter("psc_connection_id", psc_connection_id) if service_attachment_id is not None: - pulumi.set(__self__, "service_attachment_id", service_attachment_id) + _setter("service_attachment_id", service_attachment_id) @property @pulumi.getter(name="endpointRegion") @@ -10430,7 +13054,16 @@ def __init__(__self__, *, """ :param pulumi.Input['MwsWorkspacesCloudResourceContainerGcpArgs'] gcp: A block that consists of the following field: """ - pulumi.set(__self__, "gcp", gcp) + MwsWorkspacesCloudResourceContainerArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + gcp=gcp, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gcp: pulumi.Input['MwsWorkspacesCloudResourceContainerGcpArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("gcp", gcp) @property @pulumi.getter @@ -10452,7 +13085,16 @@ def __init__(__self__, *, """ :param pulumi.Input[str] project_id: The Google Cloud project ID, which the workspace uses to instantiate cloud resources for your workspace. """ - pulumi.set(__self__, "project_id", project_id) + MwsWorkspacesCloudResourceContainerGcpArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + project_id=project_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + project_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("project_id", project_id) @property @pulumi.getter(name="projectId") @@ -10473,9 +13115,22 @@ def __init__(__self__, *, authoritative_user_email: pulumi.Input[str], authoritative_user_full_name: pulumi.Input[str], customer_name: pulumi.Input[str]): - pulumi.set(__self__, "authoritative_user_email", authoritative_user_email) - pulumi.set(__self__, "authoritative_user_full_name", authoritative_user_full_name) - pulumi.set(__self__, "customer_name", customer_name) + MwsWorkspacesExternalCustomerInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + authoritative_user_email=authoritative_user_email, + authoritative_user_full_name=authoritative_user_full_name, + customer_name=customer_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + authoritative_user_email: pulumi.Input[str], + authoritative_user_full_name: pulumi.Input[str], + customer_name: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("authoritative_user_email", authoritative_user_email) + _setter("authoritative_user_full_name", authoritative_user_full_name) + _setter("customer_name", customer_name) @property @pulumi.getter(name="authoritativeUserEmail") @@ -10511,9 +13166,22 @@ def __init__(__self__, *, gke_cluster_pod_ip_range: pulumi.Input[str], gke_cluster_service_ip_range: pulumi.Input[str], subnet_cidr: pulumi.Input[str]): - pulumi.set(__self__, "gke_cluster_pod_ip_range", gke_cluster_pod_ip_range) - pulumi.set(__self__, "gke_cluster_service_ip_range", gke_cluster_service_ip_range) - pulumi.set(__self__, "subnet_cidr", subnet_cidr) + MwsWorkspacesGcpManagedNetworkConfigArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + gke_cluster_pod_ip_range=gke_cluster_pod_ip_range, + gke_cluster_service_ip_range=gke_cluster_service_ip_range, + subnet_cidr=subnet_cidr, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gke_cluster_pod_ip_range: pulumi.Input[str], + gke_cluster_service_ip_range: pulumi.Input[str], + subnet_cidr: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("gke_cluster_pod_ip_range", gke_cluster_pod_ip_range) + _setter("gke_cluster_service_ip_range", gke_cluster_service_ip_range) + _setter("subnet_cidr", subnet_cidr) @property @pulumi.getter(name="gkeClusterPodIpRange") @@ -10552,8 +13220,19 @@ def __init__(__self__, *, :param pulumi.Input[str] connectivity_type: Specifies the network connectivity types for the GKE nodes and the GKE master network. Possible values are: `PRIVATE_NODE_PUBLIC_MASTER`, `PUBLIC_NODE_PUBLIC_MASTER`. :param pulumi.Input[str] master_ip_range: The IP range from which to allocate GKE cluster master resources. This field will be ignored if GKE private cluster is not enabled. It must be exactly as big as `/28`. """ - pulumi.set(__self__, "connectivity_type", connectivity_type) - pulumi.set(__self__, "master_ip_range", master_ip_range) + MwsWorkspacesGkeConfigArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + connectivity_type=connectivity_type, + master_ip_range=master_ip_range, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + connectivity_type: pulumi.Input[str], + master_ip_range: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("connectivity_type", connectivity_type) + _setter("master_ip_range", master_ip_range) @property @pulumi.getter(name="connectivityType") @@ -10590,14 +13269,29 @@ def __init__(__self__, *, """ :param pulumi.Input[int] lifetime_seconds: Token expiry lifetime. By default its 2592000 (30 days). """ + MwsWorkspacesTokenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + comment=comment, + lifetime_seconds=lifetime_seconds, + token_id=token_id, + token_value=token_value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + comment: Optional[pulumi.Input[str]] = None, + lifetime_seconds: Optional[pulumi.Input[int]] = None, + token_id: Optional[pulumi.Input[str]] = None, + token_value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if lifetime_seconds is not None: - pulumi.set(__self__, "lifetime_seconds", lifetime_seconds) + _setter("lifetime_seconds", lifetime_seconds) if token_id is not None: - pulumi.set(__self__, "token_id", token_id) + _setter("token_id", token_id) if token_value is not None: - pulumi.set(__self__, "token_value", token_value) + _setter("token_value", token_value) @property @pulumi.getter @@ -10654,13 +13348,28 @@ def __init__(__self__, *, :param pulumi.Input[str] service_principal_name: Application ID of the service_principal. :param pulumi.Input[str] user_name: name of the user. """ - pulumi.set(__self__, "permission_level", permission_level) + PermissionsAccessControlArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + permission_level=permission_level, + group_name=group_name, + service_principal_name=service_principal_name, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permission_level: pulumi.Input[str], + group_name: Optional[pulumi.Input[str]] = None, + service_principal_name: Optional[pulumi.Input[str]] = None, + user_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("permission_level", permission_level) if group_name is not None: - pulumi.set(__self__, "group_name", group_name) + _setter("group_name", group_name) if service_principal_name is not None: - pulumi.set(__self__, "service_principal_name", service_principal_name) + _setter("service_principal_name", service_principal_name) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="permissionLevel") @@ -10735,44 +13444,89 @@ def __init__(__self__, *, spark_conf: Optional[pulumi.Input[Mapping[str, Any]]] = None, spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): + PipelineClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_log_conf=cluster_log_conf, + custom_tags=custom_tags, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + label=label, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + apply_policy_default_values: Optional[pulumi.Input[bool]] = None, + autoscale: Optional[pulumi.Input['PipelineClusterAutoscaleArgs']] = None, + aws_attributes: Optional[pulumi.Input['PipelineClusterAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['PipelineClusterAzureAttributesArgs']] = None, + cluster_log_conf: Optional[pulumi.Input['PipelineClusterClusterLogConfArgs']] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + driver_instance_pool_id: Optional[pulumi.Input[str]] = None, + driver_node_type_id: Optional[pulumi.Input[str]] = None, + enable_local_disk_encryption: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['PipelineClusterGcpAttributesArgs']] = None, + init_scripts: Optional[pulumi.Input[Sequence[pulumi.Input['PipelineClusterInitScriptArgs']]]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + label: Optional[pulumi.Input[str]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + num_workers: Optional[pulumi.Input[int]] = None, + policy_id: Optional[pulumi.Input[str]] = None, + spark_conf: Optional[pulumi.Input[Mapping[str, Any]]] = None, + spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, + ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if label is not None: - pulumi.set(__self__, "label", label) + _setter("label", label) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) @property @pulumi.getter(name="applyPolicyDefaultValues") @@ -10952,12 +13706,25 @@ def __init__(__self__, *, max_workers: Optional[pulumi.Input[int]] = None, min_workers: Optional[pulumi.Input[int]] = None, mode: Optional[pulumi.Input[str]] = None): + PipelineClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + mode=mode, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[pulumi.Input[int]] = None, + min_workers: Optional[pulumi.Input[int]] = None, + mode: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) if mode is not None: - pulumi.set(__self__, "mode", mode) + _setter("mode", mode) @property @pulumi.getter(name="maxWorkers") @@ -10998,22 +13765,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[pulumi.Input[str]] = None, spot_bid_price_percent: Optional[pulumi.Input[int]] = None, zone_id: Optional[pulumi.Input[str]] = None): + PipelineClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + ebs_volume_count: Optional[pulumi.Input[int]] = None, + ebs_volume_size: Optional[pulumi.Input[int]] = None, + ebs_volume_type: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + spot_bid_price_percent: Optional[pulumi.Input[int]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -11094,12 +13884,25 @@ def __init__(__self__, *, availability: Optional[pulumi.Input[str]] = None, first_on_demand: Optional[pulumi.Input[int]] = None, spot_bid_max_price: Optional[pulumi.Input[float]] = None): + PipelineClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + first_on_demand: Optional[pulumi.Input[int]] = None, + spot_bid_max_price: Optional[pulumi.Input[float]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -11134,10 +13937,21 @@ class PipelineClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional[pulumi.Input['PipelineClusterClusterLogConfDbfsArgs']] = None, s3: Optional[pulumi.Input['PipelineClusterClusterLogConfS3Args']] = None): + PipelineClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional[pulumi.Input['PipelineClusterClusterLogConfDbfsArgs']] = None, + s3: Optional[pulumi.Input['PipelineClusterClusterLogConfS3Args']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -11162,7 +13976,16 @@ def s3(self, value: Optional[pulumi.Input['PipelineClusterClusterLogConfS3Args'] class PipelineClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + PipelineClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -11184,19 +14007,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + PipelineClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -11269,14 +14113,29 @@ def __init__(__self__, *, google_service_account: Optional[pulumi.Input[str]] = None, local_ssd_count: Optional[pulumi.Input[int]] = None, zone_id: Optional[pulumi.Input[str]] = None): + PipelineClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[pulumi.Input[str]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + local_ssd_count: Optional[pulumi.Input[int]] = None, + zone_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -11325,23 +14184,44 @@ def __init__(__self__, *, s3: Optional[pulumi.Input['PipelineClusterInitScriptS3Args']] = None, volumes: Optional[pulumi.Input['PipelineClusterInitScriptVolumesArgs']] = None, workspace: Optional[pulumi.Input['PipelineClusterInitScriptWorkspaceArgs']] = None): + PipelineClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional[pulumi.Input['PipelineClusterInitScriptAbfssArgs']] = None, + dbfs: Optional[pulumi.Input['PipelineClusterInitScriptDbfsArgs']] = None, + file: Optional[pulumi.Input['PipelineClusterInitScriptFileArgs']] = None, + gcs: Optional[pulumi.Input['PipelineClusterInitScriptGcsArgs']] = None, + s3: Optional[pulumi.Input['PipelineClusterInitScriptS3Args']] = None, + volumes: Optional[pulumi.Input['PipelineClusterInitScriptVolumesArgs']] = None, + workspace: Optional[pulumi.Input['PipelineClusterInitScriptWorkspaceArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: warnings.warn("""For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""", DeprecationWarning) pulumi.log.warn("""dbfs is deprecated: For init scripts use 'volumes', 'workspace' or cloud storage location instead of 'dbfs'.""") if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -11414,8 +14294,17 @@ def workspace(self, value: Optional[pulumi.Input['PipelineClusterInitScriptWorks class PipelineClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + PipelineClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11431,7 +14320,16 @@ def destination(self, value: Optional[pulumi.Input[str]]): class PipelineClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: pulumi.Input[str]): - pulumi.set(__self__, "destination", destination) + PipelineClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -11447,8 +14345,17 @@ def destination(self, value: pulumi.Input[str]): class PipelineClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + PipelineClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11464,8 +14371,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class PipelineClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + PipelineClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11487,19 +14403,40 @@ def __init__(__self__, *, endpoint: Optional[pulumi.Input[str]] = None, kms_key: Optional[pulumi.Input[str]] = None, region: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "destination", destination) + PipelineClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: pulumi.Input[str], + canned_acl: Optional[pulumi.Input[str]] = None, + enable_encryption: Optional[pulumi.Input[bool]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + endpoint: Optional[pulumi.Input[str]] = None, + kms_key: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -11569,8 +14506,17 @@ def region(self, value: Optional[pulumi.Input[str]]): class PipelineClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + PipelineClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11586,8 +14532,17 @@ def destination(self, value: Optional[pulumi.Input[str]]): class PipelineClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[pulumi.Input[str]] = None): + PipelineClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11604,10 +14559,21 @@ class PipelineFiltersArgs: def __init__(__self__, *, excludes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, includes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): + PipelineFiltersArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + excludes=excludes, + includes=includes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + excludes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + includes: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if excludes is not None: - pulumi.set(__self__, "excludes", excludes) + _setter("excludes", excludes) if includes is not None: - pulumi.set(__self__, "includes", includes) + _setter("includes", includes) @property @pulumi.getter @@ -11636,16 +14602,33 @@ def __init__(__self__, *, maven: Optional[pulumi.Input['PipelineLibraryMavenArgs']] = None, notebook: Optional[pulumi.Input['PipelineLibraryNotebookArgs']] = None, whl: Optional[pulumi.Input[str]] = None): + PipelineLibraryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + file=file, + jar=jar, + maven=maven, + notebook=notebook, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + file: Optional[pulumi.Input['PipelineLibraryFileArgs']] = None, + jar: Optional[pulumi.Input[str]] = None, + maven: Optional[pulumi.Input['PipelineLibraryMavenArgs']] = None, + notebook: Optional[pulumi.Input['PipelineLibraryNotebookArgs']] = None, + whl: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if notebook is not None: - pulumi.set(__self__, "notebook", notebook) + _setter("notebook", notebook) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -11697,7 +14680,16 @@ def whl(self, value: Optional[pulumi.Input[str]]): class PipelineLibraryFileArgs: def __init__(__self__, *, path: pulumi.Input[str]): - pulumi.set(__self__, "path", path) + PipelineLibraryFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -11715,11 +14707,24 @@ def __init__(__self__, *, coordinates: pulumi.Input[str], exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, repo: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "coordinates", coordinates) + PipelineLibraryMavenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: pulumi.Input[str], + exclusions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + repo: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -11753,7 +14758,16 @@ def repo(self, value: Optional[pulumi.Input[str]]): class PipelineLibraryNotebookArgs: def __init__(__self__, *, path: pulumi.Input[str]): - pulumi.set(__self__, "path", path) + PipelineLibraryNotebookArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -11778,8 +14792,19 @@ def __init__(__self__, *, * `on-flow-failure` - a single data flow fails. :param pulumi.Input[Sequence[pulumi.Input[str]]] email_recipients: non-empty list of emails to notify. """ - pulumi.set(__self__, "alerts", alerts) - pulumi.set(__self__, "email_recipients", email_recipients) + PipelineNotificationArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alerts=alerts, + email_recipients=email_recipients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alerts: pulumi.Input[Sequence[pulumi.Input[str]]], + email_recipients: pulumi.Input[Sequence[pulumi.Input[str]]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("alerts", alerts) + _setter("email_recipients", email_recipients) @property @pulumi.getter @@ -11817,7 +14842,16 @@ def __init__(__self__, *, """ :param pulumi.Input[Sequence[pulumi.Input[str]]] allowed_ip_addresses: Allowed IP Addresses in CIDR notation. Limit of 100. """ - pulumi.set(__self__, "allowed_ip_addresses", allowed_ip_addresses) + RecipientIpAccessListArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + allowed_ip_addresses=allowed_ip_addresses, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allowed_ip_addresses: pulumi.Input[Sequence[pulumi.Input[str]]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allowed_ip_addresses", allowed_ip_addresses) @property @pulumi.getter(name="allowedIpAddresses") @@ -11851,20 +14885,41 @@ def __init__(__self__, *, :param pulumi.Input[int] updated_at: Time at which this recipient Token was updated, in epoch milliseconds. :param pulumi.Input[str] updated_by: Username of recipient Token updater. """ + RecipientTokenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + activation_url=activation_url, + created_at=created_at, + created_by=created_by, + expiration_time=expiration_time, + id=id, + updated_at=updated_at, + updated_by=updated_by, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + activation_url: Optional[pulumi.Input[str]] = None, + created_at: Optional[pulumi.Input[int]] = None, + created_by: Optional[pulumi.Input[str]] = None, + expiration_time: Optional[pulumi.Input[int]] = None, + id: Optional[pulumi.Input[str]] = None, + updated_at: Optional[pulumi.Input[int]] = None, + updated_by: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if activation_url is not None: - pulumi.set(__self__, "activation_url", activation_url) + _setter("activation_url", activation_url) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if expiration_time is not None: - pulumi.set(__self__, "expiration_time", expiration_time) + _setter("expiration_time", expiration_time) if id is not None: - pulumi.set(__self__, "id", id) + _setter("id", id) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) if updated_by is not None: - pulumi.set(__self__, "updated_by", updated_by) + _setter("updated_by", updated_by) @property @pulumi.getter(name="activationUrl") @@ -11955,7 +15010,16 @@ def updated_by(self, value: Optional[pulumi.Input[str]]): class RepoSparseCheckoutArgs: def __init__(__self__, *, patterns: pulumi.Input[Sequence[pulumi.Input[str]]]): - pulumi.set(__self__, "patterns", patterns) + RepoSparseCheckoutArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + patterns=patterns, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + patterns: pulumi.Input[Sequence[pulumi.Input[str]]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("patterns", patterns) @property @pulumi.getter @@ -11972,8 +15036,19 @@ class SecretScopeKeyvaultMetadataArgs: def __init__(__self__, *, dns_name: pulumi.Input[str], resource_id: pulumi.Input[str]): - pulumi.set(__self__, "dns_name", dns_name) - pulumi.set(__self__, "resource_id", resource_id) + SecretScopeKeyvaultMetadataArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dns_name=dns_name, + resource_id=resource_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dns_name: pulumi.Input[str], + resource_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dns_name", dns_name) + _setter("resource_id", resource_id) @property @pulumi.getter(name="dnsName") @@ -12020,26 +15095,55 @@ def __init__(__self__, *, :param pulumi.Input[int] start_version: The start version associated with the object for cdf. This allows data providers to control the lowest object version that is accessible by clients. :param pulumi.Input[str] status: Status of the object, one of: `ACTIVE`, `PERMISSION_DENIED`. """ - pulumi.set(__self__, "data_object_type", data_object_type) - pulumi.set(__self__, "name", name) + ShareObjectArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + data_object_type=data_object_type, + name=name, + added_at=added_at, + added_by=added_by, + cdf_enabled=cdf_enabled, + comment=comment, + history_data_sharing_status=history_data_sharing_status, + partitions=partitions, + shared_as=shared_as, + start_version=start_version, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + data_object_type: pulumi.Input[str], + name: pulumi.Input[str], + added_at: Optional[pulumi.Input[int]] = None, + added_by: Optional[pulumi.Input[str]] = None, + cdf_enabled: Optional[pulumi.Input[bool]] = None, + comment: Optional[pulumi.Input[str]] = None, + history_data_sharing_status: Optional[pulumi.Input[str]] = None, + partitions: Optional[pulumi.Input[Sequence[pulumi.Input['ShareObjectPartitionArgs']]]] = None, + shared_as: Optional[pulumi.Input[str]] = None, + start_version: Optional[pulumi.Input[int]] = None, + status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("data_object_type", data_object_type) + _setter("name", name) if added_at is not None: - pulumi.set(__self__, "added_at", added_at) + _setter("added_at", added_at) if added_by is not None: - pulumi.set(__self__, "added_by", added_by) + _setter("added_by", added_by) if cdf_enabled is not None: - pulumi.set(__self__, "cdf_enabled", cdf_enabled) + _setter("cdf_enabled", cdf_enabled) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if history_data_sharing_status is not None: - pulumi.set(__self__, "history_data_sharing_status", history_data_sharing_status) + _setter("history_data_sharing_status", history_data_sharing_status) if partitions is not None: - pulumi.set(__self__, "partitions", partitions) + _setter("partitions", partitions) if shared_as is not None: - pulumi.set(__self__, "shared_as", shared_as) + _setter("shared_as", shared_as) if start_version is not None: - pulumi.set(__self__, "start_version", start_version) + _setter("start_version", start_version) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter(name="dataObjectType") @@ -12171,11 +15275,26 @@ def status(self, value: Optional[pulumi.Input[str]]): class ShareObjectPartitionArgs: def __init__(__self__, *, values: pulumi.Input[Sequence[pulumi.Input['ShareObjectPartitionValueArgs']]]): - pulumi.set(__self__, "values", values) + """ + :param pulumi.Input[Sequence[pulumi.Input['ShareObjectPartitionValueArgs']]] values: The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + """ + ShareObjectPartitionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + values: pulumi.Input[Sequence[pulumi.Input['ShareObjectPartitionValueArgs']]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("values", values) @property @pulumi.getter def values(self) -> pulumi.Input[Sequence[pulumi.Input['ShareObjectPartitionValueArgs']]]: + """ + The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + """ return pulumi.get(self, "values") @values.setter @@ -12196,12 +15315,27 @@ def __init__(__self__, *, :param pulumi.Input[str] recipient_property_key: The key of a Delta Sharing recipient's property. For example `databricks-account-id`. When this field is set, field `value` can not be set. :param pulumi.Input[str] value: The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. """ - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "op", op) + ShareObjectPartitionValueArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + op=op, + recipient_property_key=recipient_property_key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: pulumi.Input[str], + op: pulumi.Input[str], + recipient_property_key: Optional[pulumi.Input[str]] = None, + value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("op", op) if recipient_property_key is not None: - pulumi.set(__self__, "recipient_property_key", recipient_property_key) + _setter("recipient_property_key", recipient_property_key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12269,15 +15403,34 @@ def __init__(__self__, *, :param pulumi.Input[str] custom_subject: Custom subject of alert notification, if it exists. This includes email subject, Slack notification header, etc. See [Alerts API reference](https://docs.databricks.com/sql/user/alerts/index.html) for custom templating instructions. :param pulumi.Input[bool] muted: Whether or not the alert is muted. If an alert is muted, it will not notify users and alert destinations when triggered. """ - pulumi.set(__self__, "column", column) - pulumi.set(__self__, "op", op) - pulumi.set(__self__, "value", value) + SqlAlertOptionsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + column=column, + op=op, + value=value, + custom_body=custom_body, + custom_subject=custom_subject, + muted=muted, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + column: pulumi.Input[str], + op: pulumi.Input[str], + value: pulumi.Input[str], + custom_body: Optional[pulumi.Input[str]] = None, + custom_subject: Optional[pulumi.Input[str]] = None, + muted: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("column", column) + _setter("op", op) + _setter("value", value) if custom_body is not None: - pulumi.set(__self__, "custom_body", custom_body) + _setter("custom_body", custom_body) if custom_subject is not None: - pulumi.set(__self__, "custom_subject", custom_subject) + _setter("custom_subject", custom_subject) if muted is not None: - pulumi.set(__self__, "muted", muted) + _setter("muted", muted) @property @pulumi.getter @@ -12359,8 +15512,17 @@ def __init__(__self__, *, """ :param pulumi.Input[str] name: Name of the Databricks SQL release channel. Possible values are: `CHANNEL_NAME_PREVIEW` and `CHANNEL_NAME_CURRENT`. Default is `CHANNEL_NAME_CURRENT`. """ + SqlEndpointChannelArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -12383,13 +15545,30 @@ def __init__(__self__, *, protocol: pulumi.Input[str], host: Optional[pulumi.Input[str]] = None, hostname: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "path", path) - pulumi.set(__self__, "port", port) - pulumi.set(__self__, "protocol", protocol) + SqlEndpointOdbcParamsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + port=port, + protocol=protocol, + host=host, + hostname=hostname, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + port: pulumi.Input[int], + protocol: pulumi.Input[str], + host: Optional[pulumi.Input[str]] = None, + hostname: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) + _setter("port", port) + _setter("protocol", protocol) if host is not None: - pulumi.set(__self__, "host", host) + _setter("host", host) if hostname is not None: - pulumi.set(__self__, "hostname", hostname) + _setter("hostname", hostname) @property @pulumi.getter @@ -12441,7 +15620,16 @@ def hostname(self, value: Optional[pulumi.Input[str]]): class SqlEndpointTagsArgs: def __init__(__self__, *, custom_tags: pulumi.Input[Sequence[pulumi.Input['SqlEndpointTagsCustomTagArgs']]]): - pulumi.set(__self__, "custom_tags", custom_tags) + SqlEndpointTagsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + custom_tags=custom_tags, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + custom_tags: pulumi.Input[Sequence[pulumi.Input['SqlEndpointTagsCustomTagArgs']]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("custom_tags", custom_tags) @property @pulumi.getter(name="customTags") @@ -12458,8 +15646,19 @@ class SqlEndpointTagsCustomTagArgs: def __init__(__self__, *, key: pulumi.Input[str], value: pulumi.Input[str]): - pulumi.set(__self__, "key", key) - pulumi.set(__self__, "value", value) + SqlEndpointTagsCustomTagArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: pulumi.Input[str], + value: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key", key) + _setter("value", value) @property @pulumi.getter @@ -12491,8 +15690,19 @@ def __init__(__self__, *, [Available](https://docs.databricks.com/security/access-control/table-acls/object-privileges.html) privilege names are: """ - pulumi.set(__self__, "principal", principal) - pulumi.set(__self__, "privileges", privileges) + SqlPermissionsPrivilegeAssignmentArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + principal=principal, + privileges=privileges, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + principal: pulumi.Input[str], + privileges: pulumi.Input[Sequence[pulumi.Input[str]]], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("principal", principal) + _setter("privileges", privileges) @property @pulumi.getter @@ -12536,29 +15746,60 @@ def __init__(__self__, *, query: Optional[pulumi.Input['SqlQueryParameterQueryArgs']] = None, text: Optional[pulumi.Input['SqlQueryParameterTextArgs']] = None, title: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "name", name) + SqlQueryParameterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + date=date, + date_range=date_range, + datetime=datetime, + datetime_range=datetime_range, + datetimesec=datetimesec, + datetimesec_range=datetimesec_range, + enum=enum, + number=number, + query=query, + text=text, + title=title, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: pulumi.Input[str], + date: Optional[pulumi.Input['SqlQueryParameterDateArgs']] = None, + date_range: Optional[pulumi.Input['SqlQueryParameterDateRangeArgs']] = None, + datetime: Optional[pulumi.Input['SqlQueryParameterDatetimeArgs']] = None, + datetime_range: Optional[pulumi.Input['SqlQueryParameterDatetimeRangeArgs']] = None, + datetimesec: Optional[pulumi.Input['SqlQueryParameterDatetimesecArgs']] = None, + datetimesec_range: Optional[pulumi.Input['SqlQueryParameterDatetimesecRangeArgs']] = None, + enum: Optional[pulumi.Input['SqlQueryParameterEnumArgs']] = None, + number: Optional[pulumi.Input['SqlQueryParameterNumberArgs']] = None, + query: Optional[pulumi.Input['SqlQueryParameterQueryArgs']] = None, + text: Optional[pulumi.Input['SqlQueryParameterTextArgs']] = None, + title: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) if date is not None: - pulumi.set(__self__, "date", date) + _setter("date", date) if date_range is not None: - pulumi.set(__self__, "date_range", date_range) + _setter("date_range", date_range) if datetime is not None: - pulumi.set(__self__, "datetime", datetime) + _setter("datetime", datetime) if datetime_range is not None: - pulumi.set(__self__, "datetime_range", datetime_range) + _setter("datetime_range", datetime_range) if datetimesec is not None: - pulumi.set(__self__, "datetimesec", datetimesec) + _setter("datetimesec", datetimesec) if datetimesec_range is not None: - pulumi.set(__self__, "datetimesec_range", datetimesec_range) + _setter("datetimesec_range", datetimesec_range) if enum is not None: - pulumi.set(__self__, "enum", enum) + _setter("enum", enum) if number is not None: - pulumi.set(__self__, "number", number) + _setter("number", number) if query is not None: - pulumi.set(__self__, "query", query) + _setter("query", query) if text is not None: - pulumi.set(__self__, "text", text) + _setter("text", text) if title is not None: - pulumi.set(__self__, "title", title) + _setter("title", title) @property @pulumi.getter @@ -12673,7 +15914,16 @@ def title(self, value: Optional[pulumi.Input[str]]): class SqlQueryParameterDateArgs: def __init__(__self__, *, value: pulumi.Input[str]): - pulumi.set(__self__, "value", value) + SqlQueryParameterDateArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12690,10 +15940,21 @@ class SqlQueryParameterDateRangeArgs: def __init__(__self__, *, range: Optional[pulumi.Input['SqlQueryParameterDateRangeRangeArgs']] = None, value: Optional[pulumi.Input[str]] = None): + SqlQueryParameterDateRangeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + range=range, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + range: Optional[pulumi.Input['SqlQueryParameterDateRangeRangeArgs']] = None, + value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if range is not None: - pulumi.set(__self__, "range", range) + _setter("range", range) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12719,8 +15980,19 @@ class SqlQueryParameterDateRangeRangeArgs: def __init__(__self__, *, end: pulumi.Input[str], start: pulumi.Input[str]): - pulumi.set(__self__, "end", end) - pulumi.set(__self__, "start", start) + SqlQueryParameterDateRangeRangeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + end=end, + start=start, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + end: pulumi.Input[str], + start: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("end", end) + _setter("start", start) @property @pulumi.getter @@ -12745,7 +16017,16 @@ def start(self, value: pulumi.Input[str]): class SqlQueryParameterDatetimeArgs: def __init__(__self__, *, value: pulumi.Input[str]): - pulumi.set(__self__, "value", value) + SqlQueryParameterDatetimeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12762,10 +16043,21 @@ class SqlQueryParameterDatetimeRangeArgs: def __init__(__self__, *, range: Optional[pulumi.Input['SqlQueryParameterDatetimeRangeRangeArgs']] = None, value: Optional[pulumi.Input[str]] = None): + SqlQueryParameterDatetimeRangeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + range=range, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + range: Optional[pulumi.Input['SqlQueryParameterDatetimeRangeRangeArgs']] = None, + value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if range is not None: - pulumi.set(__self__, "range", range) + _setter("range", range) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12791,8 +16083,19 @@ class SqlQueryParameterDatetimeRangeRangeArgs: def __init__(__self__, *, end: pulumi.Input[str], start: pulumi.Input[str]): - pulumi.set(__self__, "end", end) - pulumi.set(__self__, "start", start) + SqlQueryParameterDatetimeRangeRangeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + end=end, + start=start, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + end: pulumi.Input[str], + start: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("end", end) + _setter("start", start) @property @pulumi.getter @@ -12817,7 +16120,16 @@ def start(self, value: pulumi.Input[str]): class SqlQueryParameterDatetimesecArgs: def __init__(__self__, *, value: pulumi.Input[str]): - pulumi.set(__self__, "value", value) + SqlQueryParameterDatetimesecArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12834,10 +16146,21 @@ class SqlQueryParameterDatetimesecRangeArgs: def __init__(__self__, *, range: Optional[pulumi.Input['SqlQueryParameterDatetimesecRangeRangeArgs']] = None, value: Optional[pulumi.Input[str]] = None): + SqlQueryParameterDatetimesecRangeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + range=range, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + range: Optional[pulumi.Input['SqlQueryParameterDatetimesecRangeRangeArgs']] = None, + value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if range is not None: - pulumi.set(__self__, "range", range) + _setter("range", range) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12863,8 +16186,19 @@ class SqlQueryParameterDatetimesecRangeRangeArgs: def __init__(__self__, *, end: pulumi.Input[str], start: pulumi.Input[str]): - pulumi.set(__self__, "end", end) - pulumi.set(__self__, "start", start) + SqlQueryParameterDatetimesecRangeRangeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + end=end, + start=start, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + end: pulumi.Input[str], + start: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("end", end) + _setter("start", start) @property @pulumi.getter @@ -12892,13 +16226,28 @@ def __init__(__self__, *, multiple: Optional[pulumi.Input['SqlQueryParameterEnumMultipleArgs']] = None, value: Optional[pulumi.Input[str]] = None, values: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): - pulumi.set(__self__, "options", options) + SqlQueryParameterEnumArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + options=options, + multiple=multiple, + value=value, + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + options: pulumi.Input[Sequence[pulumi.Input[str]]], + multiple: Optional[pulumi.Input['SqlQueryParameterEnumMultipleArgs']] = None, + value: Optional[pulumi.Input[str]] = None, + values: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("options", options) if multiple is not None: - pulumi.set(__self__, "multiple", multiple) + _setter("multiple", multiple) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) if values is not None: - pulumi.set(__self__, "values", values) + _setter("values", values) @property @pulumi.getter @@ -12943,9 +16292,22 @@ def __init__(__self__, *, prefix: pulumi.Input[str], separator: pulumi.Input[str], suffix: pulumi.Input[str]): - pulumi.set(__self__, "prefix", prefix) - pulumi.set(__self__, "separator", separator) - pulumi.set(__self__, "suffix", suffix) + SqlQueryParameterEnumMultipleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + prefix=prefix, + separator=separator, + suffix=suffix, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + prefix: pulumi.Input[str], + separator: pulumi.Input[str], + suffix: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("prefix", prefix) + _setter("separator", separator) + _setter("suffix", suffix) @property @pulumi.getter @@ -12979,7 +16341,16 @@ def suffix(self, value: pulumi.Input[str]): class SqlQueryParameterNumberArgs: def __init__(__self__, *, value: pulumi.Input[float]): - pulumi.set(__self__, "value", value) + SqlQueryParameterNumberArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: pulumi.Input[float], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12998,13 +16369,28 @@ def __init__(__self__, *, multiple: Optional[pulumi.Input['SqlQueryParameterQueryMultipleArgs']] = None, value: Optional[pulumi.Input[str]] = None, values: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): - pulumi.set(__self__, "query_id", query_id) + SqlQueryParameterQueryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + query_id=query_id, + multiple=multiple, + value=value, + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + query_id: pulumi.Input[str], + multiple: Optional[pulumi.Input['SqlQueryParameterQueryMultipleArgs']] = None, + value: Optional[pulumi.Input[str]] = None, + values: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("query_id", query_id) if multiple is not None: - pulumi.set(__self__, "multiple", multiple) + _setter("multiple", multiple) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) if values is not None: - pulumi.set(__self__, "values", values) + _setter("values", values) @property @pulumi.getter(name="queryId") @@ -13049,9 +16435,22 @@ def __init__(__self__, *, prefix: pulumi.Input[str], separator: pulumi.Input[str], suffix: pulumi.Input[str]): - pulumi.set(__self__, "prefix", prefix) - pulumi.set(__self__, "separator", separator) - pulumi.set(__self__, "suffix", suffix) + SqlQueryParameterQueryMultipleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + prefix=prefix, + separator=separator, + suffix=suffix, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + prefix: pulumi.Input[str], + separator: pulumi.Input[str], + suffix: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("prefix", prefix) + _setter("separator", separator) + _setter("suffix", suffix) @property @pulumi.getter @@ -13085,7 +16484,16 @@ def suffix(self, value: pulumi.Input[str]): class SqlQueryParameterTextArgs: def __init__(__self__, *, value: pulumi.Input[str]): - pulumi.set(__self__, "value", value) + SqlQueryParameterTextArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -13103,12 +16511,25 @@ def __init__(__self__, *, continuous: Optional[pulumi.Input['SqlQueryScheduleContinuousArgs']] = None, daily: Optional[pulumi.Input['SqlQueryScheduleDailyArgs']] = None, weekly: Optional[pulumi.Input['SqlQueryScheduleWeeklyArgs']] = None): + SqlQueryScheduleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + continuous=continuous, + daily=daily, + weekly=weekly, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + continuous: Optional[pulumi.Input['SqlQueryScheduleContinuousArgs']] = None, + daily: Optional[pulumi.Input['SqlQueryScheduleDailyArgs']] = None, + weekly: Optional[pulumi.Input['SqlQueryScheduleWeeklyArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if daily is not None: - pulumi.set(__self__, "daily", daily) + _setter("daily", daily) if weekly is not None: - pulumi.set(__self__, "weekly", weekly) + _setter("weekly", weekly) @property @pulumi.getter @@ -13143,9 +16564,20 @@ class SqlQueryScheduleContinuousArgs: def __init__(__self__, *, interval_seconds: pulumi.Input[int], until_date: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "interval_seconds", interval_seconds) + SqlQueryScheduleContinuousArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + interval_seconds=interval_seconds, + until_date=until_date, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + interval_seconds: pulumi.Input[int], + until_date: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("interval_seconds", interval_seconds) if until_date is not None: - pulumi.set(__self__, "until_date", until_date) + _setter("until_date", until_date) @property @pulumi.getter(name="intervalSeconds") @@ -13172,10 +16604,23 @@ def __init__(__self__, *, interval_days: pulumi.Input[int], time_of_day: pulumi.Input[str], until_date: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "interval_days", interval_days) - pulumi.set(__self__, "time_of_day", time_of_day) + SqlQueryScheduleDailyArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + interval_days=interval_days, + time_of_day=time_of_day, + until_date=until_date, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + interval_days: pulumi.Input[int], + time_of_day: pulumi.Input[str], + until_date: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("interval_days", interval_days) + _setter("time_of_day", time_of_day) if until_date is not None: - pulumi.set(__self__, "until_date", until_date) + _setter("until_date", until_date) @property @pulumi.getter(name="intervalDays") @@ -13212,11 +16657,26 @@ def __init__(__self__, *, interval_weeks: pulumi.Input[int], time_of_day: pulumi.Input[str], until_date: Optional[pulumi.Input[str]] = None): - pulumi.set(__self__, "day_of_week", day_of_week) - pulumi.set(__self__, "interval_weeks", interval_weeks) - pulumi.set(__self__, "time_of_day", time_of_day) + SqlQueryScheduleWeeklyArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + day_of_week=day_of_week, + interval_weeks=interval_weeks, + time_of_day=time_of_day, + until_date=until_date, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + day_of_week: pulumi.Input[str], + interval_weeks: pulumi.Input[int], + time_of_day: pulumi.Input[str], + until_date: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("day_of_week", day_of_week) + _setter("interval_weeks", interval_weeks) + _setter("time_of_day", time_of_day) if until_date is not None: - pulumi.set(__self__, "until_date", until_date) + _setter("until_date", until_date) @property @pulumi.getter(name="dayOfWeek") @@ -13268,13 +16728,28 @@ def __init__(__self__, *, :param pulumi.Input[bool] nullable: Whether field is nullable (Default: `true`) :param pulumi.Input[str] type: Column type spec (with metadata) as SQL text. Not supported for `VIEW` table_type. """ - pulumi.set(__self__, "name", name) + SqlTableColumnArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + comment=comment, + nullable=nullable, + type=type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + nullable: Optional[pulumi.Input[bool]] = None, + type: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if nullable is not None: - pulumi.set(__self__, "nullable", nullable) + _setter("nullable", nullable) if type is not None: - pulumi.set(__self__, "type", type) + _setter("type", type) @property @pulumi.getter @@ -13334,16 +16809,35 @@ def __init__(__self__, *, title: Optional[pulumi.Input[str]] = None, value: Optional[pulumi.Input[str]] = None, values: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None): - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "type", type) + SqlWidgetParameterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + type=type, + map_to=map_to, + title=title, + value=value, + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: pulumi.Input[str], + type: pulumi.Input[str], + map_to: Optional[pulumi.Input[str]] = None, + title: Optional[pulumi.Input[str]] = None, + value: Optional[pulumi.Input[str]] = None, + values: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("type", type) if map_to is not None: - pulumi.set(__self__, "map_to", map_to) + _setter("map_to", map_to) if title is not None: - pulumi.set(__self__, "title", title) + _setter("title", title) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) if values is not None: - pulumi.set(__self__, "values", values) + _setter("values", values) @property @pulumi.getter @@ -13408,14 +16902,31 @@ def __init__(__self__, *, auto_height: Optional[pulumi.Input[bool]] = None, pos_x: Optional[pulumi.Input[int]] = None, pos_y: Optional[pulumi.Input[int]] = None): - pulumi.set(__self__, "size_x", size_x) - pulumi.set(__self__, "size_y", size_y) + SqlWidgetPositionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + size_x=size_x, + size_y=size_y, + auto_height=auto_height, + pos_x=pos_x, + pos_y=pos_y, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + size_x: pulumi.Input[int], + size_y: pulumi.Input[int], + auto_height: Optional[pulumi.Input[bool]] = None, + pos_x: Optional[pulumi.Input[int]] = None, + pos_y: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("size_x", size_x) + _setter("size_y", size_y) if auto_height is not None: - pulumi.set(__self__, "auto_height", auto_height) + _setter("auto_height", auto_height) if pos_x is not None: - pulumi.set(__self__, "pos_x", pos_x) + _setter("pos_x", pos_x) if pos_y is not None: - pulumi.set(__self__, "pos_y", pos_y) + _setter("pos_y", pos_y) @property @pulumi.getter(name="sizeX") @@ -13472,7 +16983,16 @@ def __init__(__self__, *, `azure_managed_identity` optional configuration block for using managed identity as credential details for Azure (recommended over service principal): """ - pulumi.set(__self__, "role_arn", role_arn) + StorageCredentialAwsIamRoleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + role_arn=role_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role_arn: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role_arn", role_arn) @property @pulumi.getter(name="roleArn") @@ -13501,11 +17021,24 @@ def __init__(__self__, *, `databricks_gcp_service_account` optional configuration block for creating a Databricks-managed GCP Service Account: """ - pulumi.set(__self__, "access_connector_id", access_connector_id) + StorageCredentialAzureManagedIdentityArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_connector_id=access_connector_id, + credential_id=credential_id, + managed_identity_id=managed_identity_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_connector_id: pulumi.Input[str], + credential_id: Optional[pulumi.Input[str]] = None, + managed_identity_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("access_connector_id", access_connector_id) if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if managed_identity_id is not None: - pulumi.set(__self__, "managed_identity_id", managed_identity_id) + _setter("managed_identity_id", managed_identity_id) @property @pulumi.getter(name="accessConnectorId") @@ -13554,9 +17087,22 @@ def __init__(__self__, *, :param pulumi.Input[str] client_secret: The client secret generated for the above app ID in AAD. **This field is redacted on output** :param pulumi.Input[str] directory_id: The directory ID corresponding to the Azure Active Directory (AAD) tenant of the application """ - pulumi.set(__self__, "application_id", application_id) - pulumi.set(__self__, "client_secret", client_secret) - pulumi.set(__self__, "directory_id", directory_id) + StorageCredentialAzureServicePrincipalArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + application_id=application_id, + client_secret=client_secret, + directory_id=directory_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + application_id: pulumi.Input[str], + client_secret: pulumi.Input[str], + directory_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("application_id", application_id) + _setter("client_secret", client_secret) + _setter("directory_id", directory_id) @property @pulumi.getter(name="applicationId") @@ -13603,10 +17149,21 @@ def __init__(__self__, *, """ :param pulumi.Input[str] email: The email of the GCP service account created, to be granted access to relevant buckets. """ + StorageCredentialDatabricksGcpServiceAccountArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + credential_id=credential_id, + email=email, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + credential_id: Optional[pulumi.Input[str]] = None, + email: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if email is not None: - pulumi.set(__self__, "email", email) + _setter("email", email) @property @pulumi.getter(name="credentialId") @@ -13639,9 +17196,22 @@ def __init__(__self__, *, """ :param pulumi.Input[str] email: The email of the GCP service account created, to be granted access to relevant buckets. """ - pulumi.set(__self__, "email", email) - pulumi.set(__self__, "private_key", private_key) - pulumi.set(__self__, "private_key_id", private_key_id) + StorageCredentialGcpServiceAccountKeyArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + email=email, + private_key=private_key, + private_key_id=private_key_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + email: pulumi.Input[str], + private_key: pulumi.Input[str], + private_key_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("email", email) + _setter("private_key", private_key) + _setter("private_key_id", private_key_id) @property @pulumi.getter @@ -13688,24 +17258,53 @@ def __init__(__self__, *, type_json: Optional[pulumi.Input[str]] = None, type_precision: Optional[pulumi.Input[int]] = None, type_scale: Optional[pulumi.Input[int]] = None): - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "position", position) - pulumi.set(__self__, "type_name", type_name) - pulumi.set(__self__, "type_text", type_text) + TableColumnArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + position=position, + type_name=type_name, + type_text=type_text, + comment=comment, + nullable=nullable, + partition_index=partition_index, + type_interval_type=type_interval_type, + type_json=type_json, + type_precision=type_precision, + type_scale=type_scale, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: pulumi.Input[str], + position: pulumi.Input[int], + type_name: pulumi.Input[str], + type_text: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + nullable: Optional[pulumi.Input[bool]] = None, + partition_index: Optional[pulumi.Input[int]] = None, + type_interval_type: Optional[pulumi.Input[str]] = None, + type_json: Optional[pulumi.Input[str]] = None, + type_precision: Optional[pulumi.Input[int]] = None, + type_scale: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("position", position) + _setter("type_name", type_name) + _setter("type_text", type_text) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if nullable is not None: - pulumi.set(__self__, "nullable", nullable) + _setter("nullable", nullable) if partition_index is not None: - pulumi.set(__self__, "partition_index", partition_index) + _setter("partition_index", partition_index) if type_interval_type is not None: - pulumi.set(__self__, "type_interval_type", type_interval_type) + _setter("type_interval_type", type_interval_type) if type_json is not None: - pulumi.set(__self__, "type_json", type_json) + _setter("type_json", type_json) if type_precision is not None: - pulumi.set(__self__, "type_precision", type_precision) + _setter("type_precision", type_precision) if type_scale is not None: - pulumi.set(__self__, "type_scale", type_scale) + _setter("type_scale", type_scale) @property @pulumi.getter @@ -13873,88 +17472,181 @@ def __init__(__self__, *, :param Mapping[str, Any] spark_env_vars: Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers. :param Sequence[str] ssh_public_keys: SSH public key contents that will be added to each Spark node in this cluster. """ - pulumi.set(__self__, "default_tags", default_tags) - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "spark_version", spark_version) - pulumi.set(__self__, "state", state) + GetClusterClusterInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + default_tags=default_tags, + driver_instance_pool_id=driver_instance_pool_id, + spark_version=spark_version, + state=state, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_cores=cluster_cores, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_log_status=cluster_log_status, + cluster_memory_mb=cluster_memory_mb, + cluster_name=cluster_name, + cluster_source=cluster_source, + creator_user_name=creator_user_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver=driver, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + executors=executors, + gcp_attributes=gcp_attributes, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + jdbc_port=jdbc_port, + last_activity_time=last_activity_time, + last_state_loss_time=last_state_loss_time, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_context_id=spark_context_id, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + start_time=start_time, + state_message=state_message, + terminate_time=terminate_time, + termination_reason=termination_reason, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default_tags: Mapping[str, Any], + driver_instance_pool_id: str, + spark_version: str, + state: str, + autoscale: Optional['GetClusterClusterInfoAutoscaleArgs'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['GetClusterClusterInfoAwsAttributesArgs'] = None, + azure_attributes: Optional['GetClusterClusterInfoAzureAttributesArgs'] = None, + cluster_cores: Optional[float] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['GetClusterClusterInfoClusterLogConfArgs'] = None, + cluster_log_status: Optional['GetClusterClusterInfoClusterLogStatusArgs'] = None, + cluster_memory_mb: Optional[int] = None, + cluster_name: Optional[str] = None, + cluster_source: Optional[str] = None, + creator_user_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['GetClusterClusterInfoDockerImageArgs'] = None, + driver: Optional['GetClusterClusterInfoDriverArgs'] = None, + driver_node_type_id: Optional[str] = None, + enable_elastic_disk: Optional[bool] = None, + enable_local_disk_encryption: Optional[bool] = None, + executors: Optional[Sequence['GetClusterClusterInfoExecutorArgs']] = None, + gcp_attributes: Optional['GetClusterClusterInfoGcpAttributesArgs'] = None, + init_scripts: Optional[Sequence['GetClusterClusterInfoInitScriptArgs']] = None, + instance_pool_id: Optional[str] = None, + jdbc_port: Optional[int] = None, + last_activity_time: Optional[int] = None, + last_state_loss_time: Optional[int] = None, + node_type_id: Optional[str] = None, + num_workers: Optional[int] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_context_id: Optional[int] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + start_time: Optional[int] = None, + state_message: Optional[str] = None, + terminate_time: Optional[int] = None, + termination_reason: Optional['GetClusterClusterInfoTerminationReasonArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("default_tags", default_tags) + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("spark_version", spark_version) + _setter("state", state) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_cores is not None: - pulumi.set(__self__, "cluster_cores", cluster_cores) + _setter("cluster_cores", cluster_cores) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_log_status is not None: - pulumi.set(__self__, "cluster_log_status", cluster_log_status) + _setter("cluster_log_status", cluster_log_status) if cluster_memory_mb is not None: - pulumi.set(__self__, "cluster_memory_mb", cluster_memory_mb) + _setter("cluster_memory_mb", cluster_memory_mb) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if cluster_source is not None: - pulumi.set(__self__, "cluster_source", cluster_source) + _setter("cluster_source", cluster_source) if creator_user_name is not None: - pulumi.set(__self__, "creator_user_name", creator_user_name) + _setter("creator_user_name", creator_user_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver is not None: - pulumi.set(__self__, "driver", driver) + _setter("driver", driver) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if executors is not None: - pulumi.set(__self__, "executors", executors) + _setter("executors", executors) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if jdbc_port is not None: - pulumi.set(__self__, "jdbc_port", jdbc_port) + _setter("jdbc_port", jdbc_port) if last_activity_time is not None: - pulumi.set(__self__, "last_activity_time", last_activity_time) + _setter("last_activity_time", last_activity_time) if last_state_loss_time is not None: - pulumi.set(__self__, "last_state_loss_time", last_state_loss_time) + _setter("last_state_loss_time", last_state_loss_time) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_context_id is not None: - pulumi.set(__self__, "spark_context_id", spark_context_id) + _setter("spark_context_id", spark_context_id) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if start_time is not None: - pulumi.set(__self__, "start_time", start_time) + _setter("start_time", start_time) if state_message is not None: - pulumi.set(__self__, "state_message", state_message) + _setter("state_message", state_message) if terminate_time is not None: - pulumi.set(__self__, "terminate_time", terminate_time) + _setter("terminate_time", terminate_time) if termination_reason is not None: - pulumi.set(__self__, "termination_reason", termination_reason) + _setter("termination_reason", termination_reason) @property @pulumi.getter(name="defaultTags") @@ -14403,10 +18095,21 @@ class GetClusterClusterInfoAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetClusterClusterInfoAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -14438,22 +18141,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetClusterClusterInfoAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -14534,12 +18260,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetClusterClusterInfoAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -14574,10 +18313,21 @@ class GetClusterClusterInfoClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional['GetClusterClusterInfoClusterLogConfDbfsArgs'] = None, s3: Optional['GetClusterClusterInfoClusterLogConfS3Args'] = None): + GetClusterClusterInfoClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['GetClusterClusterInfoClusterLogConfDbfsArgs'] = None, + s3: Optional['GetClusterClusterInfoClusterLogConfS3Args'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -14602,7 +18352,16 @@ def s3(self, value: Optional['GetClusterClusterInfoClusterLogConfS3Args']): class GetClusterClusterInfoClusterLogConfDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -14624,19 +18383,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -14707,10 +18487,21 @@ class GetClusterClusterInfoClusterLogStatusArgs: def __init__(__self__, *, last_attempted: Optional[int] = None, last_exception: Optional[str] = None): + GetClusterClusterInfoClusterLogStatusArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + last_attempted=last_attempted, + last_exception=last_exception, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + last_attempted: Optional[int] = None, + last_exception: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if last_attempted is not None: - pulumi.set(__self__, "last_attempted", last_attempted) + _setter("last_attempted", last_attempted) if last_exception is not None: - pulumi.set(__self__, "last_exception", last_exception) + _setter("last_exception", last_exception) @property @pulumi.getter(name="lastAttempted") @@ -14736,9 +18527,20 @@ class GetClusterClusterInfoDockerImageArgs: def __init__(__self__, *, url: str, basic_auth: Optional['GetClusterClusterInfoDockerImageBasicAuthArgs'] = None): - pulumi.set(__self__, "url", url) + GetClusterClusterInfoDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['GetClusterClusterInfoDockerImageBasicAuthArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -14764,8 +18566,19 @@ class GetClusterClusterInfoDockerImageBasicAuthArgs: def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetClusterClusterInfoDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -14796,20 +18609,41 @@ def __init__(__self__, *, private_ip: Optional[str] = None, public_dns: Optional[str] = None, start_timestamp: Optional[int] = None): + GetClusterClusterInfoDriverArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + host_private_ip=host_private_ip, + instance_id=instance_id, + node_aws_attributes=node_aws_attributes, + node_id=node_id, + private_ip=private_ip, + public_dns=public_dns, + start_timestamp=start_timestamp, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + host_private_ip: Optional[str] = None, + instance_id: Optional[str] = None, + node_aws_attributes: Optional['GetClusterClusterInfoDriverNodeAwsAttributesArgs'] = None, + node_id: Optional[str] = None, + private_ip: Optional[str] = None, + public_dns: Optional[str] = None, + start_timestamp: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if host_private_ip is not None: - pulumi.set(__self__, "host_private_ip", host_private_ip) + _setter("host_private_ip", host_private_ip) if instance_id is not None: - pulumi.set(__self__, "instance_id", instance_id) + _setter("instance_id", instance_id) if node_aws_attributes is not None: - pulumi.set(__self__, "node_aws_attributes", node_aws_attributes) + _setter("node_aws_attributes", node_aws_attributes) if node_id is not None: - pulumi.set(__self__, "node_id", node_id) + _setter("node_id", node_id) if private_ip is not None: - pulumi.set(__self__, "private_ip", private_ip) + _setter("private_ip", private_ip) if public_dns is not None: - pulumi.set(__self__, "public_dns", public_dns) + _setter("public_dns", public_dns) if start_timestamp is not None: - pulumi.set(__self__, "start_timestamp", start_timestamp) + _setter("start_timestamp", start_timestamp) @property @pulumi.getter(name="hostPrivateIp") @@ -14879,8 +18713,17 @@ def start_timestamp(self, value: Optional[int]): class GetClusterClusterInfoDriverNodeAwsAttributesArgs: def __init__(__self__, *, is_spot: Optional[bool] = None): + GetClusterClusterInfoDriverNodeAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + is_spot=is_spot, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + is_spot: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if is_spot is not None: - pulumi.set(__self__, "is_spot", is_spot) + _setter("is_spot", is_spot) @property @pulumi.getter(name="isSpot") @@ -14902,20 +18745,41 @@ def __init__(__self__, *, private_ip: Optional[str] = None, public_dns: Optional[str] = None, start_timestamp: Optional[int] = None): + GetClusterClusterInfoExecutorArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + host_private_ip=host_private_ip, + instance_id=instance_id, + node_aws_attributes=node_aws_attributes, + node_id=node_id, + private_ip=private_ip, + public_dns=public_dns, + start_timestamp=start_timestamp, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + host_private_ip: Optional[str] = None, + instance_id: Optional[str] = None, + node_aws_attributes: Optional['GetClusterClusterInfoExecutorNodeAwsAttributesArgs'] = None, + node_id: Optional[str] = None, + private_ip: Optional[str] = None, + public_dns: Optional[str] = None, + start_timestamp: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if host_private_ip is not None: - pulumi.set(__self__, "host_private_ip", host_private_ip) + _setter("host_private_ip", host_private_ip) if instance_id is not None: - pulumi.set(__self__, "instance_id", instance_id) + _setter("instance_id", instance_id) if node_aws_attributes is not None: - pulumi.set(__self__, "node_aws_attributes", node_aws_attributes) + _setter("node_aws_attributes", node_aws_attributes) if node_id is not None: - pulumi.set(__self__, "node_id", node_id) + _setter("node_id", node_id) if private_ip is not None: - pulumi.set(__self__, "private_ip", private_ip) + _setter("private_ip", private_ip) if public_dns is not None: - pulumi.set(__self__, "public_dns", public_dns) + _setter("public_dns", public_dns) if start_timestamp is not None: - pulumi.set(__self__, "start_timestamp", start_timestamp) + _setter("start_timestamp", start_timestamp) @property @pulumi.getter(name="hostPrivateIp") @@ -14985,8 +18849,17 @@ def start_timestamp(self, value: Optional[int]): class GetClusterClusterInfoExecutorNodeAwsAttributesArgs: def __init__(__self__, *, is_spot: Optional[bool] = None): + GetClusterClusterInfoExecutorNodeAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + is_spot=is_spot, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + is_spot: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if is_spot is not None: - pulumi.set(__self__, "is_spot", is_spot) + _setter("is_spot", is_spot) @property @pulumi.getter(name="isSpot") @@ -15007,18 +18880,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetClusterClusterInfoGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -15085,20 +18977,41 @@ def __init__(__self__, *, s3: Optional['GetClusterClusterInfoInitScriptS3Args'] = None, volumes: Optional['GetClusterClusterInfoInitScriptVolumesArgs'] = None, workspace: Optional['GetClusterClusterInfoInitScriptWorkspaceArgs'] = None): + GetClusterClusterInfoInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['GetClusterClusterInfoInitScriptAbfssArgs'] = None, + dbfs: Optional['GetClusterClusterInfoInitScriptDbfsArgs'] = None, + file: Optional['GetClusterClusterInfoInitScriptFileArgs'] = None, + gcs: Optional['GetClusterClusterInfoInitScriptGcsArgs'] = None, + s3: Optional['GetClusterClusterInfoInitScriptS3Args'] = None, + volumes: Optional['GetClusterClusterInfoInitScriptVolumesArgs'] = None, + workspace: Optional['GetClusterClusterInfoInitScriptWorkspaceArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -15168,8 +19081,17 @@ def workspace(self, value: Optional['GetClusterClusterInfoInitScriptWorkspaceArg class GetClusterClusterInfoInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -15185,7 +19107,16 @@ def destination(self, value: Optional[str]): class GetClusterClusterInfoInitScriptDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -15201,8 +19132,17 @@ def destination(self, value: str): class GetClusterClusterInfoInitScriptFileArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -15218,8 +19158,17 @@ def destination(self, value: Optional[str]): class GetClusterClusterInfoInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -15241,19 +19190,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -15323,8 +19293,17 @@ def region(self, value: Optional[str]): class GetClusterClusterInfoInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -15340,8 +19319,17 @@ def destination(self, value: Optional[str]): class GetClusterClusterInfoInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -15359,12 +19347,25 @@ def __init__(__self__, *, code: Optional[str] = None, parameters: Optional[Mapping[str, Any]] = None, type: Optional[str] = None): + GetClusterClusterInfoTerminationReasonArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + code=code, + parameters=parameters, + type=type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + code: Optional[str] = None, + parameters: Optional[Mapping[str, Any]] = None, + type: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if code is not None: - pulumi.set(__self__, "code", code) + _setter("code", code) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if type is not None: - pulumi.set(__self__, "type", type) + _setter("type", type) @property @pulumi.getter @@ -15415,38 +19416,81 @@ def __init__(__self__, *, preloaded_spark_versions: Optional[Sequence[str]] = None, state: Optional[str] = None, stats: Optional['GetInstancePoolPoolInfoStatsArgs'] = None): - pulumi.set(__self__, "default_tags", default_tags) - pulumi.set(__self__, "idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) - pulumi.set(__self__, "instance_pool_id", instance_pool_id) - pulumi.set(__self__, "instance_pool_name", instance_pool_name) + GetInstancePoolPoolInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + default_tags=default_tags, + idle_instance_autotermination_minutes=idle_instance_autotermination_minutes, + instance_pool_id=instance_pool_id, + instance_pool_name=instance_pool_name, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + custom_tags=custom_tags, + disk_spec=disk_spec, + enable_elastic_disk=enable_elastic_disk, + gcp_attributes=gcp_attributes, + instance_pool_fleet_attributes=instance_pool_fleet_attributes, + max_capacity=max_capacity, + min_idle_instances=min_idle_instances, + node_type_id=node_type_id, + preloaded_docker_images=preloaded_docker_images, + preloaded_spark_versions=preloaded_spark_versions, + state=state, + stats=stats, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default_tags: Mapping[str, Any], + idle_instance_autotermination_minutes: int, + instance_pool_id: str, + instance_pool_name: str, + aws_attributes: Optional['GetInstancePoolPoolInfoAwsAttributesArgs'] = None, + azure_attributes: Optional['GetInstancePoolPoolInfoAzureAttributesArgs'] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + disk_spec: Optional['GetInstancePoolPoolInfoDiskSpecArgs'] = None, + enable_elastic_disk: Optional[bool] = None, + gcp_attributes: Optional['GetInstancePoolPoolInfoGcpAttributesArgs'] = None, + instance_pool_fleet_attributes: Optional[Sequence['GetInstancePoolPoolInfoInstancePoolFleetAttributeArgs']] = None, + max_capacity: Optional[int] = None, + min_idle_instances: Optional[int] = None, + node_type_id: Optional[str] = None, + preloaded_docker_images: Optional[Sequence['GetInstancePoolPoolInfoPreloadedDockerImageArgs']] = None, + preloaded_spark_versions: Optional[Sequence[str]] = None, + state: Optional[str] = None, + stats: Optional['GetInstancePoolPoolInfoStatsArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("default_tags", default_tags) + _setter("idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) + _setter("instance_pool_id", instance_pool_id) + _setter("instance_pool_name", instance_pool_name) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if disk_spec is not None: - pulumi.set(__self__, "disk_spec", disk_spec) + _setter("disk_spec", disk_spec) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if instance_pool_fleet_attributes is not None: - pulumi.set(__self__, "instance_pool_fleet_attributes", instance_pool_fleet_attributes) + _setter("instance_pool_fleet_attributes", instance_pool_fleet_attributes) if max_capacity is not None: - pulumi.set(__self__, "max_capacity", max_capacity) + _setter("max_capacity", max_capacity) if min_idle_instances is not None: - pulumi.set(__self__, "min_idle_instances", min_idle_instances) + _setter("min_idle_instances", min_idle_instances) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if preloaded_docker_images is not None: - pulumi.set(__self__, "preloaded_docker_images", preloaded_docker_images) + _setter("preloaded_docker_images", preloaded_docker_images) if preloaded_spark_versions is not None: - pulumi.set(__self__, "preloaded_spark_versions", preloaded_spark_versions) + _setter("preloaded_spark_versions", preloaded_spark_versions) if state is not None: - pulumi.set(__self__, "state", state) + _setter("state", state) if stats is not None: - pulumi.set(__self__, "stats", stats) + _setter("stats", stats) @property @pulumi.getter(name="defaultTags") @@ -15617,11 +19661,24 @@ def __init__(__self__, *, zone_id: str, availability: Optional[str] = None, spot_bid_price_percent: Optional[int] = None): - pulumi.set(__self__, "zone_id", zone_id) + GetInstancePoolPoolInfoAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + zone_id=zone_id, + availability=availability, + spot_bid_price_percent=spot_bid_price_percent, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + zone_id: str, + availability: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("zone_id", zone_id) if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) @property @pulumi.getter(name="zoneId") @@ -15656,10 +19713,21 @@ class GetInstancePoolPoolInfoAzureAttributesArgs: def __init__(__self__, *, availability: Optional[str] = None, spot_bid_max_price: Optional[float] = None): + GetInstancePoolPoolInfoAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -15686,12 +19754,25 @@ def __init__(__self__, *, disk_count: Optional[int] = None, disk_size: Optional[int] = None, disk_type: Optional['GetInstancePoolPoolInfoDiskSpecDiskTypeArgs'] = None): + GetInstancePoolPoolInfoDiskSpecArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + disk_count=disk_count, + disk_size=disk_size, + disk_type=disk_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + disk_count: Optional[int] = None, + disk_size: Optional[int] = None, + disk_type: Optional['GetInstancePoolPoolInfoDiskSpecDiskTypeArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if disk_count is not None: - pulumi.set(__self__, "disk_count", disk_count) + _setter("disk_count", disk_count) if disk_size is not None: - pulumi.set(__self__, "disk_size", disk_size) + _setter("disk_size", disk_size) if disk_type is not None: - pulumi.set(__self__, "disk_type", disk_type) + _setter("disk_type", disk_type) @property @pulumi.getter(name="diskCount") @@ -15726,10 +19807,21 @@ class GetInstancePoolPoolInfoDiskSpecDiskTypeArgs: def __init__(__self__, *, azure_disk_volume_type: Optional[str] = None, ebs_volume_type: Optional[str] = None): + GetInstancePoolPoolInfoDiskSpecDiskTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + azure_disk_volume_type=azure_disk_volume_type, + ebs_volume_type=ebs_volume_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + azure_disk_volume_type: Optional[str] = None, + ebs_volume_type: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if azure_disk_volume_type is not None: - pulumi.set(__self__, "azure_disk_volume_type", azure_disk_volume_type) + _setter("azure_disk_volume_type", azure_disk_volume_type) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) @property @pulumi.getter(name="azureDiskVolumeType") @@ -15755,10 +19847,21 @@ class GetInstancePoolPoolInfoGcpAttributesArgs: def __init__(__self__, *, gcp_availability: Optional[str] = None, local_ssd_count: Optional[int] = None): + GetInstancePoolPoolInfoGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + gcp_availability=gcp_availability, + local_ssd_count=local_ssd_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gcp_availability: Optional[str] = None, + local_ssd_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if gcp_availability is not None: - pulumi.set(__self__, "gcp_availability", gcp_availability) + _setter("gcp_availability", gcp_availability) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) @property @pulumi.getter(name="gcpAvailability") @@ -15785,11 +19888,24 @@ def __init__(__self__, *, launch_template_overrides: Sequence['GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideArgs'], fleet_on_demand_option: Optional['GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionArgs'] = None, fleet_spot_option: Optional['GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionArgs'] = None): - pulumi.set(__self__, "launch_template_overrides", launch_template_overrides) + GetInstancePoolPoolInfoInstancePoolFleetAttributeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + launch_template_overrides=launch_template_overrides, + fleet_on_demand_option=fleet_on_demand_option, + fleet_spot_option=fleet_spot_option, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + launch_template_overrides: Sequence['GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideArgs'], + fleet_on_demand_option: Optional['GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionArgs'] = None, + fleet_spot_option: Optional['GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("launch_template_overrides", launch_template_overrides) if fleet_on_demand_option is not None: - pulumi.set(__self__, "fleet_on_demand_option", fleet_on_demand_option) + _setter("fleet_on_demand_option", fleet_on_demand_option) if fleet_spot_option is not None: - pulumi.set(__self__, "fleet_spot_option", fleet_spot_option) + _setter("fleet_spot_option", fleet_spot_option) @property @pulumi.getter(name="launchTemplateOverrides") @@ -15824,9 +19940,20 @@ class GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionArgs: def __init__(__self__, *, allocation_strategy: str, instance_pools_to_use_count: Optional[int] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: str, + instance_pools_to_use_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -15852,9 +19979,20 @@ class GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionArgs: def __init__(__self__, *, allocation_strategy: str, instance_pools_to_use_count: Optional[int] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: str, + instance_pools_to_use_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -15880,8 +20018,19 @@ class GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideArg def __init__(__self__, *, availability_zone: str, instance_type: str): - pulumi.set(__self__, "availability_zone", availability_zone) - pulumi.set(__self__, "instance_type", instance_type) + GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability_zone=availability_zone, + instance_type=instance_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability_zone: str, + instance_type: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("availability_zone", availability_zone) + _setter("instance_type", instance_type) @property @pulumi.getter(name="availabilityZone") @@ -15907,9 +20056,20 @@ class GetInstancePoolPoolInfoPreloadedDockerImageArgs: def __init__(__self__, *, url: str, basic_auth: Optional['GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthArgs'] = None): - pulumi.set(__self__, "url", url) + GetInstancePoolPoolInfoPreloadedDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -15935,8 +20095,19 @@ class GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthArgs: def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -15964,14 +20135,29 @@ def __init__(__self__, *, pending_idle_count: Optional[int] = None, pending_used_count: Optional[int] = None, used_count: Optional[int] = None): + GetInstancePoolPoolInfoStatsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + idle_count=idle_count, + pending_idle_count=pending_idle_count, + pending_used_count=pending_used_count, + used_count=used_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + idle_count: Optional[int] = None, + pending_idle_count: Optional[int] = None, + pending_used_count: Optional[int] = None, + used_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if idle_count is not None: - pulumi.set(__self__, "idle_count", idle_count) + _setter("idle_count", idle_count) if pending_idle_count is not None: - pulumi.set(__self__, "pending_idle_count", pending_idle_count) + _setter("pending_idle_count", pending_idle_count) if pending_used_count is not None: - pulumi.set(__self__, "pending_used_count", pending_used_count) + _setter("pending_used_count", pending_used_count) if used_count is not None: - pulumi.set(__self__, "used_count", used_count) + _setter("used_count", used_count) @property @pulumi.getter(name="idleCount") @@ -16018,15 +20204,32 @@ def __init__(__self__, *, creator_user_name: Optional[str] = None, job_id: Optional[int] = None, settings: Optional['GetJobJobSettingsSettingsArgs'] = None): - pulumi.set(__self__, "run_as_user_name", run_as_user_name) + GetJobJobSettingsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + run_as_user_name=run_as_user_name, + created_time=created_time, + creator_user_name=creator_user_name, + job_id=job_id, + settings=settings, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + run_as_user_name: str, + created_time: Optional[int] = None, + creator_user_name: Optional[str] = None, + job_id: Optional[int] = None, + settings: Optional['GetJobJobSettingsSettingsArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("run_as_user_name", run_as_user_name) if created_time is not None: - pulumi.set(__self__, "created_time", created_time) + _setter("created_time", created_time) if creator_user_name is not None: - pulumi.set(__self__, "creator_user_name", creator_user_name) + _setter("creator_user_name", creator_user_name) if job_id is not None: - pulumi.set(__self__, "job_id", job_id) + _setter("job_id", job_id) if settings is not None: - pulumi.set(__self__, "settings", settings) + _setter("settings", settings) @property @pulumi.getter(name="runAsUserName") @@ -16113,71 +20316,144 @@ def __init__(__self__, *, """ :param str name: the job name of Job if the resource was matched by id. """ - pulumi.set(__self__, "format", format) + GetJobJobSettingsSettingsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + format=format, + computes=computes, + continuous=continuous, + dbt_task=dbt_task, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + git_source=git_source, + health=health, + job_clusters=job_clusters, + libraries=libraries, + max_concurrent_runs=max_concurrent_runs, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + name=name, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + parameters=parameters, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + queue=queue, + retry_on_timeout=retry_on_timeout, + run_as=run_as, + run_job_task=run_job_task, + schedule=schedule, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + tags=tags, + tasks=tasks, + timeout_seconds=timeout_seconds, + trigger=trigger, + webhook_notifications=webhook_notifications, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + format: str, + computes: Optional[Sequence['GetJobJobSettingsSettingsComputeArgs']] = None, + continuous: Optional['GetJobJobSettingsSettingsContinuousArgs'] = None, + dbt_task: Optional['GetJobJobSettingsSettingsDbtTaskArgs'] = None, + email_notifications: Optional['GetJobJobSettingsSettingsEmailNotificationsArgs'] = None, + existing_cluster_id: Optional[str] = None, + git_source: Optional['GetJobJobSettingsSettingsGitSourceArgs'] = None, + health: Optional['GetJobJobSettingsSettingsHealthArgs'] = None, + job_clusters: Optional[Sequence['GetJobJobSettingsSettingsJobClusterArgs']] = None, + libraries: Optional[Sequence['GetJobJobSettingsSettingsLibraryArgs']] = None, + max_concurrent_runs: Optional[int] = None, + max_retries: Optional[int] = None, + min_retry_interval_millis: Optional[int] = None, + name: Optional[str] = None, + new_cluster: Optional['GetJobJobSettingsSettingsNewClusterArgs'] = None, + notebook_task: Optional['GetJobJobSettingsSettingsNotebookTaskArgs'] = None, + notification_settings: Optional['GetJobJobSettingsSettingsNotificationSettingsArgs'] = None, + parameters: Optional[Sequence['GetJobJobSettingsSettingsParameterArgs']] = None, + pipeline_task: Optional['GetJobJobSettingsSettingsPipelineTaskArgs'] = None, + python_wheel_task: Optional['GetJobJobSettingsSettingsPythonWheelTaskArgs'] = None, + queue: Optional['GetJobJobSettingsSettingsQueueArgs'] = None, + retry_on_timeout: Optional[bool] = None, + run_as: Optional['GetJobJobSettingsSettingsRunAsArgs'] = None, + run_job_task: Optional['GetJobJobSettingsSettingsRunJobTaskArgs'] = None, + schedule: Optional['GetJobJobSettingsSettingsScheduleArgs'] = None, + spark_jar_task: Optional['GetJobJobSettingsSettingsSparkJarTaskArgs'] = None, + spark_python_task: Optional['GetJobJobSettingsSettingsSparkPythonTaskArgs'] = None, + spark_submit_task: Optional['GetJobJobSettingsSettingsSparkSubmitTaskArgs'] = None, + tags: Optional[Mapping[str, Any]] = None, + tasks: Optional[Sequence['GetJobJobSettingsSettingsTaskArgs']] = None, + timeout_seconds: Optional[int] = None, + trigger: Optional['GetJobJobSettingsSettingsTriggerArgs'] = None, + webhook_notifications: Optional['GetJobJobSettingsSettingsWebhookNotificationsArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("format", format) if computes is not None: - pulumi.set(__self__, "computes", computes) + _setter("computes", computes) if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if git_source is not None: - pulumi.set(__self__, "git_source", git_source) + _setter("git_source", git_source) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_clusters is not None: - pulumi.set(__self__, "job_clusters", job_clusters) + _setter("job_clusters", job_clusters) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_concurrent_runs is not None: - pulumi.set(__self__, "max_concurrent_runs", max_concurrent_runs) + _setter("max_concurrent_runs", max_concurrent_runs) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if queue is not None: - pulumi.set(__self__, "queue", queue) + _setter("queue", queue) if retry_on_timeout is not None: - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + _setter("retry_on_timeout", retry_on_timeout) if run_as is not None: - pulumi.set(__self__, "run_as", run_as) + _setter("run_as", run_as) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if schedule is not None: - pulumi.set(__self__, "schedule", schedule) + _setter("schedule", schedule) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if tasks is not None: - pulumi.set(__self__, "tasks", tasks) + _setter("tasks", tasks) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) if trigger is not None: - pulumi.set(__self__, "trigger", trigger) + _setter("trigger", trigger) if webhook_notifications is not None: - pulumi.set(__self__, "webhook_notifications", webhook_notifications) + _setter("webhook_notifications", webhook_notifications) @property @pulumi.getter @@ -16485,10 +20761,21 @@ class GetJobJobSettingsSettingsComputeArgs: def __init__(__self__, *, compute_key: Optional[str] = None, spec: Optional['GetJobJobSettingsSettingsComputeSpecArgs'] = None): + GetJobJobSettingsSettingsComputeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + compute_key=compute_key, + spec=spec, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + compute_key: Optional[str] = None, + spec: Optional['GetJobJobSettingsSettingsComputeSpecArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if spec is not None: - pulumi.set(__self__, "spec", spec) + _setter("spec", spec) @property @pulumi.getter(name="computeKey") @@ -16513,8 +20800,17 @@ def spec(self, value: Optional['GetJobJobSettingsSettingsComputeSpecArgs']): class GetJobJobSettingsSettingsComputeSpecArgs: def __init__(__self__, *, kind: Optional[str] = None): + GetJobJobSettingsSettingsComputeSpecArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + kind=kind, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + kind: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if kind is not None: - pulumi.set(__self__, "kind", kind) + _setter("kind", kind) @property @pulumi.getter @@ -16530,8 +20826,17 @@ def kind(self, value: Optional[str]): class GetJobJobSettingsSettingsContinuousArgs: def __init__(__self__, *, pause_status: Optional[str] = None): + GetJobJobSettingsSettingsContinuousArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="pauseStatus") @@ -16552,17 +20857,36 @@ def __init__(__self__, *, project_directory: Optional[str] = None, schema: Optional[str] = None, warehouse_id: Optional[str] = None): - pulumi.set(__self__, "commands", commands) + GetJobJobSettingsSettingsDbtTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: Sequence[str], + catalog: Optional[str] = None, + profiles_directory: Optional[str] = None, + project_directory: Optional[str] = None, + schema: Optional[str] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -16628,18 +20952,37 @@ def __init__(__self__, *, on_failures: Optional[Sequence[str]] = None, on_starts: Optional[Sequence[str]] = None, on_successes: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsEmailNotificationsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + on_duration_warning_threshold_exceededs: Optional[Sequence[str]] = None, + on_failures: Optional[Sequence[str]] = None, + on_starts: Optional[Sequence[str]] = None, + on_successes: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -16705,17 +21048,36 @@ def __init__(__self__, *, job_source: Optional['GetJobJobSettingsSettingsGitSourceJobSourceArgs'] = None, provider: Optional[str] = None, tag: Optional[str] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsGitSourceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + branch=branch, + commit=commit, + job_source=job_source, + provider=provider, + tag=tag, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + branch: Optional[str] = None, + commit: Optional[str] = None, + job_source: Optional['GetJobJobSettingsSettingsGitSourceJobSourceArgs'] = None, + provider: Optional[str] = None, + tag: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if branch is not None: - pulumi.set(__self__, "branch", branch) + _setter("branch", branch) if commit is not None: - pulumi.set(__self__, "commit", commit) + _setter("commit", commit) if job_source is not None: - pulumi.set(__self__, "job_source", job_source) + _setter("job_source", job_source) if provider is not None: - pulumi.set(__self__, "provider", provider) + _setter("provider", provider) if tag is not None: - pulumi.set(__self__, "tag", tag) + _setter("tag", tag) @property @pulumi.getter @@ -16778,10 +21140,23 @@ def __init__(__self__, *, import_from_git_branch: str, job_config_path: str, dirty_state: Optional[str] = None): - pulumi.set(__self__, "import_from_git_branch", import_from_git_branch) - pulumi.set(__self__, "job_config_path", job_config_path) + GetJobJobSettingsSettingsGitSourceJobSourceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + import_from_git_branch=import_from_git_branch, + job_config_path=job_config_path, + dirty_state=dirty_state, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + import_from_git_branch: str, + job_config_path: str, + dirty_state: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("import_from_git_branch", import_from_git_branch) + _setter("job_config_path", job_config_path) if dirty_state is not None: - pulumi.set(__self__, "dirty_state", dirty_state) + _setter("dirty_state", dirty_state) @property @pulumi.getter(name="importFromGitBranch") @@ -16815,7 +21190,16 @@ def dirty_state(self, value: Optional[str]): class GetJobJobSettingsSettingsHealthArgs: def __init__(__self__, *, rules: Sequence['GetJobJobSettingsSettingsHealthRuleArgs']): - pulumi.set(__self__, "rules", rules) + GetJobJobSettingsSettingsHealthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: Sequence['GetJobJobSettingsSettingsHealthRuleArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -16833,12 +21217,25 @@ def __init__(__self__, *, metric: Optional[str] = None, op: Optional[str] = None, value: Optional[int] = None): + GetJobJobSettingsSettingsHealthRuleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[str] = None, + op: Optional[str] = None, + value: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -16873,10 +21270,21 @@ class GetJobJobSettingsSettingsJobClusterArgs: def __init__(__self__, *, job_cluster_key: Optional[str] = None, new_cluster: Optional['GetJobJobSettingsSettingsJobClusterNewClusterArgs'] = None): + GetJobJobSettingsSettingsJobClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_cluster_key=job_cluster_key, + new_cluster=new_cluster, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_cluster_key: Optional[str] = None, + new_cluster: Optional['GetJobJobSettingsSettingsJobClusterNewClusterArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") @@ -16930,59 +21338,126 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeArgs'] = None): - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) - pulumi.set(__self__, "node_type_id", node_type_id) - pulumi.set(__self__, "num_workers", num_workers) - pulumi.set(__self__, "spark_version", spark_version) + GetJobJobSettingsSettingsJobClusterNewClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + node_type_id=node_type_id, + num_workers=num_workers, + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + driver_instance_pool_id: str, + driver_node_type_id: str, + enable_elastic_disk: bool, + enable_local_disk_encryption: bool, + node_type_id: str, + num_workers: int, + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleArgs'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesArgs'] = None, + azure_attributes: Optional['GetJobJobSettingsSettingsJobClusterNewClusterAzureAttributesArgs'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfArgs'] = None, + cluster_mount_infos: Optional[Sequence['GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoArgs']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['GetJobJobSettingsSettingsJobClusterNewClusterDockerImageArgs'] = None, + gcp_attributes: Optional['GetJobJobSettingsSettingsJobClusterNewClusterGcpAttributesArgs'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptArgs']] = None, + instance_pool_id: Optional[str] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_node_type_id", driver_node_type_id) + _setter("enable_elastic_disk", enable_elastic_disk) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) + _setter("node_type_id", node_type_id) + _setter("num_workers", num_workers) + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="driverInstancePoolId") @@ -17260,10 +21735,21 @@ class GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -17295,22 +21781,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -17391,12 +21900,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetJobJobSettingsSettingsJobClusterNewClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -17431,10 +21953,21 @@ class GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional['GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsArgs'] = None, s3: Optional['GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfS3Args'] = None): + GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsArgs'] = None, + s3: Optional['GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfS3Args'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -17459,7 +21992,16 @@ def s3(self, value: Optional['GetJobJobSettingsSettingsJobClusterNewClusterClust class GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -17481,19 +22023,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -17565,10 +22128,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoArgs', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoArgs', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -17603,9 +22179,20 @@ class GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesy def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -17631,9 +22218,20 @@ class GetJobJobSettingsSettingsJobClusterNewClusterDockerImageArgs: def __init__(__self__, *, url: str, basic_auth: Optional['GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthArgs'] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsJobClusterNewClusterDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -17659,8 +22257,19 @@ class GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthArgs: def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -17690,18 +22299,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -17768,20 +22396,41 @@ def __init__(__self__, *, s3: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptS3Args'] = None, volumes: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesArgs'] = None, workspace: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceArgs'] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptAbfssArgs'] = None, + dbfs: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptDbfsArgs'] = None, + file: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptFileArgs'] = None, + gcs: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptGcsArgs'] = None, + s3: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptS3Args'] = None, + volumes: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesArgs'] = None, + workspace: Optional['GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -17851,8 +22500,17 @@ def workspace(self, value: Optional['GetJobJobSettingsSettingsJobClusterNewClust class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17868,7 +22526,16 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -17884,8 +22551,17 @@ def destination(self, value: str): class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17901,8 +22577,17 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17924,19 +22609,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -18006,8 +22712,17 @@ def region(self, value: Optional[str]): class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -18023,8 +22738,17 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -18040,7 +22764,16 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeArgs: def __init__(__self__, *, clients: 'GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsArgs'): - pulumi.set(__self__, "clients", clients) + GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsArgs', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -18057,10 +22790,21 @@ class GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsArgs: def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -18090,18 +22834,37 @@ def __init__(__self__, *, maven: Optional['GetJobJobSettingsSettingsLibraryMavenArgs'] = None, pypi: Optional['GetJobJobSettingsSettingsLibraryPypiArgs'] = None, whl: Optional[str] = None): + GetJobJobSettingsSettingsLibraryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional['GetJobJobSettingsSettingsLibraryCranArgs'] = None, + egg: Optional[str] = None, + jar: Optional[str] = None, + maven: Optional['GetJobJobSettingsSettingsLibraryMavenArgs'] = None, + pypi: Optional['GetJobJobSettingsSettingsLibraryPypiArgs'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -18163,9 +22926,20 @@ class GetJobJobSettingsSettingsLibraryCranArgs: def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsLibraryCranArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -18192,11 +22966,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + GetJobJobSettingsSettingsLibraryMavenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -18231,9 +23018,20 @@ class GetJobJobSettingsSettingsLibraryPypiArgs: def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsLibraryPypiArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -18287,59 +23085,126 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['GetJobJobSettingsSettingsNewClusterWorkloadTypeArgs'] = None): - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) - pulumi.set(__self__, "node_type_id", node_type_id) - pulumi.set(__self__, "num_workers", num_workers) - pulumi.set(__self__, "spark_version", spark_version) + GetJobJobSettingsSettingsNewClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + node_type_id=node_type_id, + num_workers=num_workers, + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + driver_instance_pool_id: str, + driver_node_type_id: str, + enable_elastic_disk: bool, + enable_local_disk_encryption: bool, + node_type_id: str, + num_workers: int, + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['GetJobJobSettingsSettingsNewClusterAutoscaleArgs'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['GetJobJobSettingsSettingsNewClusterAwsAttributesArgs'] = None, + azure_attributes: Optional['GetJobJobSettingsSettingsNewClusterAzureAttributesArgs'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['GetJobJobSettingsSettingsNewClusterClusterLogConfArgs'] = None, + cluster_mount_infos: Optional[Sequence['GetJobJobSettingsSettingsNewClusterClusterMountInfoArgs']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['GetJobJobSettingsSettingsNewClusterDockerImageArgs'] = None, + gcp_attributes: Optional['GetJobJobSettingsSettingsNewClusterGcpAttributesArgs'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['GetJobJobSettingsSettingsNewClusterInitScriptArgs']] = None, + instance_pool_id: Optional[str] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['GetJobJobSettingsSettingsNewClusterWorkloadTypeArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_node_type_id", driver_node_type_id) + _setter("enable_elastic_disk", enable_elastic_disk) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) + _setter("node_type_id", node_type_id) + _setter("num_workers", num_workers) + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="driverInstancePoolId") @@ -18617,10 +23482,21 @@ class GetJobJobSettingsSettingsNewClusterAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetJobJobSettingsSettingsNewClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -18652,22 +23528,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -18748,12 +23647,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetJobJobSettingsSettingsNewClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -18788,10 +23700,21 @@ class GetJobJobSettingsSettingsNewClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional['GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsArgs'] = None, s3: Optional['GetJobJobSettingsSettingsNewClusterClusterLogConfS3Args'] = None): + GetJobJobSettingsSettingsNewClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsArgs'] = None, + s3: Optional['GetJobJobSettingsSettingsNewClusterClusterLogConfS3Args'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -18816,7 +23739,16 @@ def s3(self, value: Optional['GetJobJobSettingsSettingsNewClusterClusterLogConfS class GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -18838,19 +23770,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -18922,10 +23875,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoArgs', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + GetJobJobSettingsSettingsNewClusterClusterMountInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoArgs', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -18960,9 +23926,20 @@ class GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoAr def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -18988,9 +23965,20 @@ class GetJobJobSettingsSettingsNewClusterDockerImageArgs: def __init__(__self__, *, url: str, basic_auth: Optional['GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthArgs'] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsNewClusterDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -19016,8 +24004,19 @@ class GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthArgs: def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -19047,18 +24046,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -19125,20 +24143,41 @@ def __init__(__self__, *, s3: Optional['GetJobJobSettingsSettingsNewClusterInitScriptS3Args'] = None, volumes: Optional['GetJobJobSettingsSettingsNewClusterInitScriptVolumesArgs'] = None, workspace: Optional['GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceArgs'] = None): + GetJobJobSettingsSettingsNewClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['GetJobJobSettingsSettingsNewClusterInitScriptAbfssArgs'] = None, + dbfs: Optional['GetJobJobSettingsSettingsNewClusterInitScriptDbfsArgs'] = None, + file: Optional['GetJobJobSettingsSettingsNewClusterInitScriptFileArgs'] = None, + gcs: Optional['GetJobJobSettingsSettingsNewClusterInitScriptGcsArgs'] = None, + s3: Optional['GetJobJobSettingsSettingsNewClusterInitScriptS3Args'] = None, + volumes: Optional['GetJobJobSettingsSettingsNewClusterInitScriptVolumesArgs'] = None, + workspace: Optional['GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -19208,8 +24247,17 @@ def workspace(self, value: Optional['GetJobJobSettingsSettingsNewClusterInitScri class GetJobJobSettingsSettingsNewClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -19225,7 +24273,16 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsNewClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -19241,8 +24298,17 @@ def destination(self, value: str): class GetJobJobSettingsSettingsNewClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -19258,8 +24324,17 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsNewClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -19281,19 +24356,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -19363,8 +24459,17 @@ def region(self, value: Optional[str]): class GetJobJobSettingsSettingsNewClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -19380,8 +24485,17 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -19397,7 +24511,16 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsNewClusterWorkloadTypeArgs: def __init__(__self__, *, clients: 'GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsArgs'): - pulumi.set(__self__, "clients", clients) + GetJobJobSettingsSettingsNewClusterWorkloadTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsArgs', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -19414,10 +24537,21 @@ class GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsArgs: def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -19444,11 +24578,24 @@ def __init__(__self__, *, notebook_path: str, base_parameters: Optional[Mapping[str, Any]] = None, source: Optional[str] = None): - pulumi.set(__self__, "notebook_path", notebook_path) + GetJobJobSettingsSettingsNotebookTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: str, + base_parameters: Optional[Mapping[str, Any]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -19483,10 +24630,21 @@ class GetJobJobSettingsSettingsNotificationSettingsArgs: def __init__(__self__, *, no_alert_for_canceled_runs: Optional[bool] = None, no_alert_for_skipped_runs: Optional[bool] = None): + GetJobJobSettingsSettingsNotificationSettingsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + no_alert_for_canceled_runs: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="noAlertForCanceledRuns") @@ -19515,10 +24673,21 @@ def __init__(__self__, *, """ :param str name: the job name of Job if the resource was matched by id. """ + GetJobJobSettingsSettingsParameterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + default=default, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default: Optional[str] = None, + name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if default is not None: - pulumi.set(__self__, "default", default) + _setter("default", default) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -19547,9 +24716,20 @@ class GetJobJobSettingsSettingsPipelineTaskArgs: def __init__(__self__, *, pipeline_id: str, full_refresh: Optional[bool] = None): - pulumi.set(__self__, "pipeline_id", pipeline_id) + GetJobJobSettingsSettingsPipelineTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: str, + full_refresh: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -19577,14 +24757,29 @@ def __init__(__self__, *, named_parameters: Optional[Mapping[str, Any]] = None, package_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsPythonWheelTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[str] = None, + named_parameters: Optional[Mapping[str, Any]] = None, + package_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -19627,6 +24822,11 @@ def parameters(self, value: Optional[Sequence[str]]): class GetJobJobSettingsSettingsQueueArgs: def __init__(__self__): pass + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + opts: Optional[pulumi.ResourceOptions]=None): + pass @pulumi.input_type @@ -19634,10 +24834,21 @@ class GetJobJobSettingsSettingsRunAsArgs: def __init__(__self__, *, service_principal_name: Optional[str] = None, user_name: Optional[str] = None): + GetJobJobSettingsSettingsRunAsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + service_principal_name=service_principal_name, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + service_principal_name: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if service_principal_name is not None: - pulumi.set(__self__, "service_principal_name", service_principal_name) + _setter("service_principal_name", service_principal_name) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="servicePrincipalName") @@ -19663,9 +24874,20 @@ class GetJobJobSettingsSettingsRunJobTaskArgs: def __init__(__self__, *, job_id: int, job_parameters: Optional[Mapping[str, Any]] = None): - pulumi.set(__self__, "job_id", job_id) + GetJobJobSettingsSettingsRunJobTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: int, + job_parameters: Optional[Mapping[str, Any]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -19692,10 +24914,23 @@ def __init__(__self__, *, quartz_cron_expression: str, timezone_id: str, pause_status: Optional[str] = None): - pulumi.set(__self__, "quartz_cron_expression", quartz_cron_expression) - pulumi.set(__self__, "timezone_id", timezone_id) + GetJobJobSettingsSettingsScheduleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + quartz_cron_expression=quartz_cron_expression, + timezone_id=timezone_id, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + quartz_cron_expression: str, + timezone_id: str, + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("quartz_cron_expression", quartz_cron_expression) + _setter("timezone_id", timezone_id) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="quartzCronExpression") @@ -19731,12 +24966,25 @@ def __init__(__self__, *, jar_uri: Optional[str] = None, main_class_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsSparkJarTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[str] = None, + main_class_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -19772,11 +25020,24 @@ def __init__(__self__, *, python_file: str, parameters: Optional[Sequence[str]] = None, source: Optional[str] = None): - pulumi.set(__self__, "python_file", python_file) + GetJobJobSettingsSettingsSparkPythonTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: str, + parameters: Optional[Sequence[str]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -19810,8 +25071,17 @@ def source(self, value: Optional[str]): class GetJobJobSettingsSettingsSparkSubmitTaskArgs: def __init__(__self__, *, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsSparkSubmitTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -19852,57 +25122,116 @@ def __init__(__self__, *, sql_task: Optional['GetJobJobSettingsSettingsTaskSqlTaskArgs'] = None, task_key: Optional[str] = None, timeout_seconds: Optional[int] = None): - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + GetJobJobSettingsSettingsTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + retry_on_timeout=retry_on_timeout, + compute_key=compute_key, + condition_task=condition_task, + dbt_task=dbt_task, + depends_ons=depends_ons, + description=description, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + health=health, + job_cluster_key=job_cluster_key, + libraries=libraries, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + run_if=run_if, + run_job_task=run_job_task, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + sql_task=sql_task, + task_key=task_key, + timeout_seconds=timeout_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + retry_on_timeout: bool, + compute_key: Optional[str] = None, + condition_task: Optional['GetJobJobSettingsSettingsTaskConditionTaskArgs'] = None, + dbt_task: Optional['GetJobJobSettingsSettingsTaskDbtTaskArgs'] = None, + depends_ons: Optional[Sequence['GetJobJobSettingsSettingsTaskDependsOnArgs']] = None, + description: Optional[str] = None, + email_notifications: Optional['GetJobJobSettingsSettingsTaskEmailNotificationsArgs'] = None, + existing_cluster_id: Optional[str] = None, + health: Optional['GetJobJobSettingsSettingsTaskHealthArgs'] = None, + job_cluster_key: Optional[str] = None, + libraries: Optional[Sequence['GetJobJobSettingsSettingsTaskLibraryArgs']] = None, + max_retries: Optional[int] = None, + min_retry_interval_millis: Optional[int] = None, + new_cluster: Optional['GetJobJobSettingsSettingsTaskNewClusterArgs'] = None, + notebook_task: Optional['GetJobJobSettingsSettingsTaskNotebookTaskArgs'] = None, + notification_settings: Optional['GetJobJobSettingsSettingsTaskNotificationSettingsArgs'] = None, + pipeline_task: Optional['GetJobJobSettingsSettingsTaskPipelineTaskArgs'] = None, + python_wheel_task: Optional['GetJobJobSettingsSettingsTaskPythonWheelTaskArgs'] = None, + run_if: Optional[str] = None, + run_job_task: Optional['GetJobJobSettingsSettingsTaskRunJobTaskArgs'] = None, + spark_jar_task: Optional['GetJobJobSettingsSettingsTaskSparkJarTaskArgs'] = None, + spark_python_task: Optional['GetJobJobSettingsSettingsTaskSparkPythonTaskArgs'] = None, + spark_submit_task: Optional['GetJobJobSettingsSettingsTaskSparkSubmitTaskArgs'] = None, + sql_task: Optional['GetJobJobSettingsSettingsTaskSqlTaskArgs'] = None, + task_key: Optional[str] = None, + timeout_seconds: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("retry_on_timeout", retry_on_timeout) if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if condition_task is not None: - pulumi.set(__self__, "condition_task", condition_task) + _setter("condition_task", condition_task) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if depends_ons is not None: - pulumi.set(__self__, "depends_ons", depends_ons) + _setter("depends_ons", depends_ons) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if run_if is not None: - pulumi.set(__self__, "run_if", run_if) + _setter("run_if", run_if) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if sql_task is not None: - pulumi.set(__self__, "sql_task", sql_task) + _setter("sql_task", sql_task) if task_key is not None: - pulumi.set(__self__, "task_key", task_key) + _setter("task_key", task_key) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) @property @pulumi.getter(name="retryOnTimeout") @@ -20145,12 +25474,25 @@ def __init__(__self__, *, left: Optional[str] = None, op: Optional[str] = None, right: Optional[str] = None): + GetJobJobSettingsSettingsTaskConditionTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + left=left, + op=op, + right=right, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + left: Optional[str] = None, + op: Optional[str] = None, + right: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if left is not None: - pulumi.set(__self__, "left", left) + _setter("left", left) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if right is not None: - pulumi.set(__self__, "right", right) + _setter("right", right) @property @pulumi.getter @@ -20189,17 +25531,36 @@ def __init__(__self__, *, project_directory: Optional[str] = None, schema: Optional[str] = None, warehouse_id: Optional[str] = None): - pulumi.set(__self__, "commands", commands) + GetJobJobSettingsSettingsTaskDbtTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: Sequence[str], + catalog: Optional[str] = None, + profiles_directory: Optional[str] = None, + project_directory: Optional[str] = None, + schema: Optional[str] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -20261,9 +25622,20 @@ class GetJobJobSettingsSettingsTaskDependsOnArgs: def __init__(__self__, *, task_key: str, outcome: Optional[str] = None): - pulumi.set(__self__, "task_key", task_key) + GetJobJobSettingsSettingsTaskDependsOnArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + task_key=task_key, + outcome=outcome, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + task_key: str, + outcome: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("task_key", task_key) if outcome is not None: - pulumi.set(__self__, "outcome", outcome) + _setter("outcome", outcome) @property @pulumi.getter(name="taskKey") @@ -20293,18 +25665,37 @@ def __init__(__self__, *, on_failures: Optional[Sequence[str]] = None, on_starts: Optional[Sequence[str]] = None, on_successes: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskEmailNotificationsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + on_duration_warning_threshold_exceededs: Optional[Sequence[str]] = None, + on_failures: Optional[Sequence[str]] = None, + on_starts: Optional[Sequence[str]] = None, + on_successes: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -20365,7 +25756,16 @@ def on_successes(self, value: Optional[Sequence[str]]): class GetJobJobSettingsSettingsTaskHealthArgs: def __init__(__self__, *, rules: Sequence['GetJobJobSettingsSettingsTaskHealthRuleArgs']): - pulumi.set(__self__, "rules", rules) + GetJobJobSettingsSettingsTaskHealthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: Sequence['GetJobJobSettingsSettingsTaskHealthRuleArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -20383,12 +25783,25 @@ def __init__(__self__, *, metric: Optional[str] = None, op: Optional[str] = None, value: Optional[int] = None): + GetJobJobSettingsSettingsTaskHealthRuleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[str] = None, + op: Optional[str] = None, + value: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -20427,18 +25840,37 @@ def __init__(__self__, *, maven: Optional['GetJobJobSettingsSettingsTaskLibraryMavenArgs'] = None, pypi: Optional['GetJobJobSettingsSettingsTaskLibraryPypiArgs'] = None, whl: Optional[str] = None): + GetJobJobSettingsSettingsTaskLibraryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional['GetJobJobSettingsSettingsTaskLibraryCranArgs'] = None, + egg: Optional[str] = None, + jar: Optional[str] = None, + maven: Optional['GetJobJobSettingsSettingsTaskLibraryMavenArgs'] = None, + pypi: Optional['GetJobJobSettingsSettingsTaskLibraryPypiArgs'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -20500,9 +25932,20 @@ class GetJobJobSettingsSettingsTaskLibraryCranArgs: def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsTaskLibraryCranArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -20529,11 +25972,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + GetJobJobSettingsSettingsTaskLibraryMavenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -20568,9 +26024,20 @@ class GetJobJobSettingsSettingsTaskLibraryPypiArgs: def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsTaskLibraryPypiArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -20624,59 +26091,126 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeArgs'] = None): - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) - pulumi.set(__self__, "node_type_id", node_type_id) - pulumi.set(__self__, "num_workers", num_workers) - pulumi.set(__self__, "spark_version", spark_version) + GetJobJobSettingsSettingsTaskNewClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + node_type_id=node_type_id, + num_workers=num_workers, + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + driver_instance_pool_id: str, + driver_node_type_id: str, + enable_elastic_disk: bool, + enable_local_disk_encryption: bool, + node_type_id: str, + num_workers: int, + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['GetJobJobSettingsSettingsTaskNewClusterAutoscaleArgs'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['GetJobJobSettingsSettingsTaskNewClusterAwsAttributesArgs'] = None, + azure_attributes: Optional['GetJobJobSettingsSettingsTaskNewClusterAzureAttributesArgs'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['GetJobJobSettingsSettingsTaskNewClusterClusterLogConfArgs'] = None, + cluster_mount_infos: Optional[Sequence['GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoArgs']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['GetJobJobSettingsSettingsTaskNewClusterDockerImageArgs'] = None, + gcp_attributes: Optional['GetJobJobSettingsSettingsTaskNewClusterGcpAttributesArgs'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['GetJobJobSettingsSettingsTaskNewClusterInitScriptArgs']] = None, + instance_pool_id: Optional[str] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_node_type_id", driver_node_type_id) + _setter("enable_elastic_disk", enable_elastic_disk) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) + _setter("node_type_id", node_type_id) + _setter("num_workers", num_workers) + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="driverInstancePoolId") @@ -20954,10 +26488,21 @@ class GetJobJobSettingsSettingsTaskNewClusterAutoscaleArgs: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetJobJobSettingsSettingsTaskNewClusterAutoscaleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -20989,22 +26534,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterAwsAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -21085,12 +26653,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetJobJobSettingsSettingsTaskNewClusterAzureAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -21125,10 +26706,21 @@ class GetJobJobSettingsSettingsTaskNewClusterClusterLogConfArgs: def __init__(__self__, *, dbfs: Optional['GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsArgs'] = None, s3: Optional['GetJobJobSettingsSettingsTaskNewClusterClusterLogConfS3Args'] = None): + GetJobJobSettingsSettingsTaskNewClusterClusterLogConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsArgs'] = None, + s3: Optional['GetJobJobSettingsSettingsTaskNewClusterClusterLogConfS3Args'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -21153,7 +26745,16 @@ def s3(self, value: Optional['GetJobJobSettingsSettingsTaskNewClusterClusterLogC class GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -21175,19 +26776,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterClusterLogConfS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -21259,10 +26881,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemInfoArgs', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemInfoArgs', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -21297,9 +26932,20 @@ class GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemIn def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -21325,9 +26971,20 @@ class GetJobJobSettingsSettingsTaskNewClusterDockerImageArgs: def __init__(__self__, *, url: str, basic_auth: Optional['GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthArgs'] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsTaskNewClusterDockerImageArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -21353,8 +27010,19 @@ class GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthArgs: def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -21384,18 +27052,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterGcpAttributesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -21462,20 +27149,41 @@ def __init__(__self__, *, s3: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptS3Args'] = None, volumes: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesArgs'] = None, workspace: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceArgs'] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptAbfssArgs'] = None, + dbfs: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptDbfsArgs'] = None, + file: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptFileArgs'] = None, + gcs: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptGcsArgs'] = None, + s3: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptS3Args'] = None, + volumes: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesArgs'] = None, + workspace: Optional['GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceArgs'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -21545,8 +27253,17 @@ def workspace(self, value: Optional['GetJobJobSettingsSettingsTaskNewClusterInit class GetJobJobSettingsSettingsTaskNewClusterInitScriptAbfssArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptAbfssArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -21562,7 +27279,16 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsTaskNewClusterInitScriptDbfsArgs: def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterInitScriptDbfsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -21578,8 +27304,17 @@ def destination(self, value: str): class GetJobJobSettingsSettingsTaskNewClusterInitScriptFileArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -21595,8 +27330,17 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsTaskNewClusterInitScriptGcsArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptGcsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -21618,19 +27362,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterInitScriptS3Args._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -21700,8 +27465,17 @@ def region(self, value: Optional[str]): class GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -21717,8 +27491,17 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceArgs: def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -21734,7 +27517,16 @@ def destination(self, value: Optional[str]): class GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeArgs: def __init__(__self__, *, clients: 'GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsArgs'): - pulumi.set(__self__, "clients", clients) + GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsArgs', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -21751,10 +27543,21 @@ class GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsArgs: def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -21781,11 +27584,24 @@ def __init__(__self__, *, notebook_path: str, base_parameters: Optional[Mapping[str, Any]] = None, source: Optional[str] = None): - pulumi.set(__self__, "notebook_path", notebook_path) + GetJobJobSettingsSettingsTaskNotebookTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: str, + base_parameters: Optional[Mapping[str, Any]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -21821,12 +27637,25 @@ def __init__(__self__, *, alert_on_last_attempt: Optional[bool] = None, no_alert_for_canceled_runs: Optional[bool] = None, no_alert_for_skipped_runs: Optional[bool] = None): + GetJobJobSettingsSettingsTaskNotificationSettingsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_canceled_runs: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="alertOnLastAttempt") @@ -21861,9 +27690,20 @@ class GetJobJobSettingsSettingsTaskPipelineTaskArgs: def __init__(__self__, *, pipeline_id: str, full_refresh: Optional[bool] = None): - pulumi.set(__self__, "pipeline_id", pipeline_id) + GetJobJobSettingsSettingsTaskPipelineTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: str, + full_refresh: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -21891,14 +27731,29 @@ def __init__(__self__, *, named_parameters: Optional[Mapping[str, Any]] = None, package_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskPythonWheelTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[str] = None, + named_parameters: Optional[Mapping[str, Any]] = None, + package_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -21942,9 +27797,20 @@ class GetJobJobSettingsSettingsTaskRunJobTaskArgs: def __init__(__self__, *, job_id: int, job_parameters: Optional[Mapping[str, Any]] = None): - pulumi.set(__self__, "job_id", job_id) + GetJobJobSettingsSettingsTaskRunJobTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: int, + job_parameters: Optional[Mapping[str, Any]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -21971,12 +27837,25 @@ def __init__(__self__, *, jar_uri: Optional[str] = None, main_class_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskSparkJarTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[str] = None, + main_class_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -22012,11 +27891,24 @@ def __init__(__self__, *, python_file: str, parameters: Optional[Sequence[str]] = None, source: Optional[str] = None): - pulumi.set(__self__, "python_file", python_file) + GetJobJobSettingsSettingsTaskSparkPythonTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: str, + parameters: Optional[Sequence[str]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -22050,8 +27942,17 @@ def source(self, value: Optional[str]): class GetJobJobSettingsSettingsTaskSparkSubmitTaskArgs: def __init__(__self__, *, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskSparkSubmitTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -22072,18 +27973,37 @@ def __init__(__self__, *, parameters: Optional[Mapping[str, Any]] = None, query: Optional['GetJobJobSettingsSettingsTaskSqlTaskQueryArgs'] = None, warehouse_id: Optional[str] = None): + GetJobJobSettingsSettingsTaskSqlTaskArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert=alert, + dashboard=dashboard, + file=file, + parameters=parameters, + query=query, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert: Optional['GetJobJobSettingsSettingsTaskSqlTaskAlertArgs'] = None, + dashboard: Optional['GetJobJobSettingsSettingsTaskSqlTaskDashboardArgs'] = None, + file: Optional['GetJobJobSettingsSettingsTaskSqlTaskFileArgs'] = None, + parameters: Optional[Mapping[str, Any]] = None, + query: Optional['GetJobJobSettingsSettingsTaskSqlTaskQueryArgs'] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert is not None: - pulumi.set(__self__, "alert", alert) + _setter("alert", alert) if dashboard is not None: - pulumi.set(__self__, "dashboard", dashboard) + _setter("dashboard", dashboard) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if query is not None: - pulumi.set(__self__, "query", query) + _setter("query", query) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -22146,10 +28066,23 @@ def __init__(__self__, *, alert_id: str, subscriptions: Sequence['GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionArgs'], pause_subscriptions: Optional[bool] = None): - pulumi.set(__self__, "alert_id", alert_id) - pulumi.set(__self__, "subscriptions", subscriptions) + GetJobJobSettingsSettingsTaskSqlTaskAlertArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_id=alert_id, + subscriptions=subscriptions, + pause_subscriptions=pause_subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_id: str, + subscriptions: Sequence['GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionArgs'], + pause_subscriptions: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("alert_id", alert_id) + _setter("subscriptions", subscriptions) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) @property @pulumi.getter(name="alertId") @@ -22184,10 +28117,21 @@ class GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionArgs: def __init__(__self__, *, destination_id: Optional[str] = None, user_name: Optional[str] = None): + GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -22215,13 +28159,28 @@ def __init__(__self__, *, custom_subject: Optional[str] = None, pause_subscriptions: Optional[bool] = None, subscriptions: Optional[Sequence['GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionArgs']] = None): - pulumi.set(__self__, "dashboard_id", dashboard_id) + GetJobJobSettingsSettingsTaskSqlTaskDashboardArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dashboard_id=dashboard_id, + custom_subject=custom_subject, + pause_subscriptions=pause_subscriptions, + subscriptions=subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dashboard_id: str, + custom_subject: Optional[str] = None, + pause_subscriptions: Optional[bool] = None, + subscriptions: Optional[Sequence['GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dashboard_id", dashboard_id) if custom_subject is not None: - pulumi.set(__self__, "custom_subject", custom_subject) + _setter("custom_subject", custom_subject) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) if subscriptions is not None: - pulumi.set(__self__, "subscriptions", subscriptions) + _setter("subscriptions", subscriptions) @property @pulumi.getter(name="dashboardId") @@ -22265,10 +28224,21 @@ class GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionArgs: def __init__(__self__, *, destination_id: Optional[str] = None, user_name: Optional[str] = None): + GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -22293,7 +28263,16 @@ def user_name(self, value: Optional[str]): class GetJobJobSettingsSettingsTaskSqlTaskFileArgs: def __init__(__self__, *, path: str): - pulumi.set(__self__, "path", path) + GetJobJobSettingsSettingsTaskSqlTaskFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -22309,7 +28288,16 @@ def path(self, value: str): class GetJobJobSettingsSettingsTaskSqlTaskQueryArgs: def __init__(__self__, *, query_id: str): - pulumi.set(__self__, "query_id", query_id) + GetJobJobSettingsSettingsTaskSqlTaskQueryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + query_id=query_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + query_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("query_id", query_id) @property @pulumi.getter(name="queryId") @@ -22326,9 +28314,20 @@ class GetJobJobSettingsSettingsTriggerArgs: def __init__(__self__, *, file_arrival: 'GetJobJobSettingsSettingsTriggerFileArrivalArgs', pause_status: Optional[str] = None): - pulumi.set(__self__, "file_arrival", file_arrival) + GetJobJobSettingsSettingsTriggerArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + file_arrival=file_arrival, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + file_arrival: 'GetJobJobSettingsSettingsTriggerFileArrivalArgs', + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("file_arrival", file_arrival) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="fileArrival") @@ -22355,11 +28354,24 @@ def __init__(__self__, *, url: str, min_time_between_triggers_seconds: Optional[int] = None, wait_after_last_change_seconds: Optional[int] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsTriggerFileArrivalArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + min_time_between_triggers_seconds=min_time_between_triggers_seconds, + wait_after_last_change_seconds=wait_after_last_change_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + min_time_between_triggers_seconds: Optional[int] = None, + wait_after_last_change_seconds: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if min_time_between_triggers_seconds is not None: - pulumi.set(__self__, "min_time_between_triggers_seconds", min_time_between_triggers_seconds) + _setter("min_time_between_triggers_seconds", min_time_between_triggers_seconds) if wait_after_last_change_seconds is not None: - pulumi.set(__self__, "wait_after_last_change_seconds", wait_after_last_change_seconds) + _setter("wait_after_last_change_seconds", wait_after_last_change_seconds) @property @pulumi.getter @@ -22396,14 +28408,29 @@ def __init__(__self__, *, on_failures: Optional[Sequence['GetJobJobSettingsSettingsWebhookNotificationsOnFailureArgs']] = None, on_starts: Optional[Sequence['GetJobJobSettingsSettingsWebhookNotificationsOnStartArgs']] = None, on_successes: Optional[Sequence['GetJobJobSettingsSettingsWebhookNotificationsOnSuccessArgs']] = None): + GetJobJobSettingsSettingsWebhookNotificationsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + on_duration_warning_threshold_exceededs: Optional[Sequence['GetJobJobSettingsSettingsWebhookNotificationsOnDurationWarningThresholdExceededArgs']] = None, + on_failures: Optional[Sequence['GetJobJobSettingsSettingsWebhookNotificationsOnFailureArgs']] = None, + on_starts: Optional[Sequence['GetJobJobSettingsSettingsWebhookNotificationsOnStartArgs']] = None, + on_successes: Optional[Sequence['GetJobJobSettingsSettingsWebhookNotificationsOnSuccessArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="onDurationWarningThresholdExceededs") @@ -22449,7 +28476,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnDurationWarningThresholdExceededArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -22471,7 +28507,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnFailureArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -22493,7 +28538,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnStartArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -22515,7 +28569,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnSuccessArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -22560,42 +28623,85 @@ def __init__(__self__, *, :param str owner: Username/groupname/sp application_id of the metastore owner. :param str storage_root: Path on cloud storage account, where managed `Table` are stored. Change forces creation of a new resource. """ + GetMetastoreMetastoreInfoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cloud=cloud, + created_at=created_at, + created_by=created_by, + default_data_access_config_id=default_data_access_config_id, + delta_sharing_organization_name=delta_sharing_organization_name, + delta_sharing_recipient_token_lifetime_in_seconds=delta_sharing_recipient_token_lifetime_in_seconds, + delta_sharing_scope=delta_sharing_scope, + global_metastore_id=global_metastore_id, + metastore_id=metastore_id, + name=name, + owner=owner, + privilege_model_version=privilege_model_version, + region=region, + storage_root=storage_root, + storage_root_credential_id=storage_root_credential_id, + storage_root_credential_name=storage_root_credential_name, + updated_at=updated_at, + updated_by=updated_by, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cloud: Optional[str] = None, + created_at: Optional[int] = None, + created_by: Optional[str] = None, + default_data_access_config_id: Optional[str] = None, + delta_sharing_organization_name: Optional[str] = None, + delta_sharing_recipient_token_lifetime_in_seconds: Optional[int] = None, + delta_sharing_scope: Optional[str] = None, + global_metastore_id: Optional[str] = None, + metastore_id: Optional[str] = None, + name: Optional[str] = None, + owner: Optional[str] = None, + privilege_model_version: Optional[str] = None, + region: Optional[str] = None, + storage_root: Optional[str] = None, + storage_root_credential_id: Optional[str] = None, + storage_root_credential_name: Optional[str] = None, + updated_at: Optional[int] = None, + updated_by: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cloud is not None: - pulumi.set(__self__, "cloud", cloud) + _setter("cloud", cloud) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if default_data_access_config_id is not None: - pulumi.set(__self__, "default_data_access_config_id", default_data_access_config_id) + _setter("default_data_access_config_id", default_data_access_config_id) if delta_sharing_organization_name is not None: - pulumi.set(__self__, "delta_sharing_organization_name", delta_sharing_organization_name) + _setter("delta_sharing_organization_name", delta_sharing_organization_name) if delta_sharing_recipient_token_lifetime_in_seconds is not None: - pulumi.set(__self__, "delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) + _setter("delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) if delta_sharing_scope is not None: - pulumi.set(__self__, "delta_sharing_scope", delta_sharing_scope) + _setter("delta_sharing_scope", delta_sharing_scope) if global_metastore_id is not None: - pulumi.set(__self__, "global_metastore_id", global_metastore_id) + _setter("global_metastore_id", global_metastore_id) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if privilege_model_version is not None: - pulumi.set(__self__, "privilege_model_version", privilege_model_version) + _setter("privilege_model_version", privilege_model_version) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) if storage_root is not None: - pulumi.set(__self__, "storage_root", storage_root) + _setter("storage_root", storage_root) if storage_root_credential_id is not None: - pulumi.set(__self__, "storage_root_credential_id", storage_root_credential_id) + _setter("storage_root_credential_id", storage_root_credential_id) if storage_root_credential_name is not None: - pulumi.set(__self__, "storage_root_credential_name", storage_root_credential_name) + _setter("storage_root_credential_name", storage_root_credential_name) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) if updated_by is not None: - pulumi.set(__self__, "updated_by", updated_by) + _setter("updated_by", updated_by) @property @pulumi.getter @@ -22800,23 +28906,52 @@ def __init__(__self__, *, :param str name: The name of the share :param str comment: Description about the object. """ - pulumi.set(__self__, "added_at", added_at) - pulumi.set(__self__, "added_by", added_by) - pulumi.set(__self__, "data_object_type", data_object_type) - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "status", status) + GetShareObjectArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + added_at=added_at, + added_by=added_by, + data_object_type=data_object_type, + name=name, + status=status, + cdf_enabled=cdf_enabled, + comment=comment, + history_data_sharing_status=history_data_sharing_status, + partitions=partitions, + shared_as=shared_as, + start_version=start_version, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + added_at: int, + added_by: str, + data_object_type: str, + name: str, + status: str, + cdf_enabled: Optional[bool] = None, + comment: Optional[str] = None, + history_data_sharing_status: Optional[str] = None, + partitions: Optional[Sequence['GetShareObjectPartitionArgs']] = None, + shared_as: Optional[str] = None, + start_version: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("added_at", added_at) + _setter("added_by", added_by) + _setter("data_object_type", data_object_type) + _setter("name", name) + _setter("status", status) if cdf_enabled is not None: - pulumi.set(__self__, "cdf_enabled", cdf_enabled) + _setter("cdf_enabled", cdf_enabled) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if history_data_sharing_status is not None: - pulumi.set(__self__, "history_data_sharing_status", history_data_sharing_status) + _setter("history_data_sharing_status", history_data_sharing_status) if partitions is not None: - pulumi.set(__self__, "partitions", partitions) + _setter("partitions", partitions) if shared_as is not None: - pulumi.set(__self__, "shared_as", shared_as) + _setter("shared_as", shared_as) if start_version is not None: - pulumi.set(__self__, "start_version", start_version) + _setter("start_version", start_version) @property @pulumi.getter(name="addedAt") @@ -22931,7 +29066,16 @@ def start_version(self, value: Optional[int]): class GetShareObjectPartitionArgs: def __init__(__self__, *, values: Sequence['GetShareObjectPartitionValueArgs']): - pulumi.set(__self__, "values", values) + GetShareObjectPartitionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + values: Sequence['GetShareObjectPartitionValueArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("values", values) @property @pulumi.getter @@ -22953,12 +29097,27 @@ def __init__(__self__, *, """ :param str name: The name of the share """ - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "op", op) + GetShareObjectPartitionValueArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + op=op, + recipient_property_key=recipient_property_key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: str, + op: str, + recipient_property_key: Optional[str] = None, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("op", op) if recipient_property_key is not None: - pulumi.set(__self__, "recipient_property_key", recipient_property_key) + _setter("recipient_property_key", recipient_property_key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -23007,8 +29166,17 @@ def __init__(__self__, *, """ :param str name: Name of the SQL warehouse to search (case-sensitive). """ + GetSqlWarehouseChannelArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -23031,13 +29199,30 @@ def __init__(__self__, *, protocol: str, host: Optional[str] = None, hostname: Optional[str] = None): - pulumi.set(__self__, "path", path) - pulumi.set(__self__, "port", port) - pulumi.set(__self__, "protocol", protocol) + GetSqlWarehouseOdbcParamsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + port=port, + protocol=protocol, + host=host, + hostname=hostname, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + port: int, + protocol: str, + host: Optional[str] = None, + hostname: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) + _setter("port", port) + _setter("protocol", protocol) if host is not None: - pulumi.set(__self__, "host", host) + _setter("host", host) if hostname is not None: - pulumi.set(__self__, "hostname", hostname) + _setter("hostname", hostname) @property @pulumi.getter @@ -23089,7 +29274,16 @@ def hostname(self, value: Optional[str]): class GetSqlWarehouseTagsArgs: def __init__(__self__, *, custom_tags: Sequence['GetSqlWarehouseTagsCustomTagArgs']): - pulumi.set(__self__, "custom_tags", custom_tags) + GetSqlWarehouseTagsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + custom_tags=custom_tags, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + custom_tags: Sequence['GetSqlWarehouseTagsCustomTagArgs'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("custom_tags", custom_tags) @property @pulumi.getter(name="customTags") @@ -23106,8 +29300,19 @@ class GetSqlWarehouseTagsCustomTagArgs: def __init__(__self__, *, key: str, value: str): - pulumi.set(__self__, "key", key) - pulumi.set(__self__, "value", value) + GetSqlWarehouseTagsCustomTagArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: str, + value: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key", key) + _setter("value", value) @property @pulumi.getter diff --git a/sdk/python/pulumi_databricks/access_control_rule_set.py b/sdk/python/pulumi_databricks/access_control_rule_set.py index 9d91571f..a0ff39b7 100644 --- a/sdk/python/pulumi_databricks/access_control_rule_set.py +++ b/sdk/python/pulumi_databricks/access_control_rule_set.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -28,10 +28,21 @@ def __init__(__self__, *, * `accounts/{account_id}/groups/{group_id}/ruleSets/default` * `accounts/{account_id}/ruleSets/default` """ + AccessControlRuleSetArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + grant_rules=grant_rules, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + grant_rules: Optional[pulumi.Input[Sequence[pulumi.Input['AccessControlRuleSetGrantRuleArgs']]]] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if grant_rules is not None: - pulumi.set(__self__, "grant_rules", grant_rules) + _setter("grant_rules", grant_rules) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter(name="grantRules") @@ -79,12 +90,25 @@ def __init__(__self__, *, * `accounts/{account_id}/groups/{group_id}/ruleSets/default` * `accounts/{account_id}/ruleSets/default` """ + _AccessControlRuleSetState._configure( + lambda key, value: pulumi.set(__self__, key, value), + etag=etag, + grant_rules=grant_rules, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + etag: Optional[pulumi.Input[str]] = None, + grant_rules: Optional[pulumi.Input[Sequence[pulumi.Input['AccessControlRuleSetGrantRuleArgs']]]] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if etag is not None: - pulumi.set(__self__, "etag", etag) + _setter("etag", etag) if grant_rules is not None: - pulumi.set(__self__, "grant_rules", grant_rules) + _setter("grant_rules", grant_rules) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -413,6 +437,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + AccessControlRuleSetArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/catalog.py b/sdk/python/pulumi_databricks/catalog.py index 27977b2e..93d60c3d 100644 --- a/sdk/python/pulumi_databricks/catalog.py +++ b/sdk/python/pulumi_databricks/catalog.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['CatalogArgs', 'Catalog'] @@ -40,30 +40,61 @@ def __init__(__self__, *, :param pulumi.Input[str] share_name: For Delta Sharing Catalogs: the name of the share under the share provider. Change forces creation of a new resource. :param pulumi.Input[str] storage_root: Managed location of the catalog. Location in cloud storage where data for managed tables will be stored. If not specified, the location will default to the metastore root location. Change forces creation of a new resource. """ + CatalogArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + comment=comment, + connection_name=connection_name, + force_destroy=force_destroy, + isolation_mode=isolation_mode, + metastore_id=metastore_id, + name=name, + options=options, + owner=owner, + properties=properties, + provider_name=provider_name, + share_name=share_name, + storage_root=storage_root, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + comment: Optional[pulumi.Input[str]] = None, + connection_name: Optional[pulumi.Input[str]] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + isolation_mode: Optional[pulumi.Input[str]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + options: Optional[pulumi.Input[Mapping[str, Any]]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + provider_name: Optional[pulumi.Input[str]] = None, + share_name: Optional[pulumi.Input[str]] = None, + storage_root: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if connection_name is not None: - pulumi.set(__self__, "connection_name", connection_name) + _setter("connection_name", connection_name) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if isolation_mode is not None: - pulumi.set(__self__, "isolation_mode", isolation_mode) + _setter("isolation_mode", isolation_mode) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if options is not None: - pulumi.set(__self__, "options", options) + _setter("options", options) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if provider_name is not None: - pulumi.set(__self__, "provider_name", provider_name) + _setter("provider_name", provider_name) if share_name is not None: - pulumi.set(__self__, "share_name", share_name) + _setter("share_name", share_name) if storage_root is not None: - pulumi.set(__self__, "storage_root", storage_root) + _setter("storage_root", storage_root) @property @pulumi.getter @@ -236,30 +267,61 @@ def __init__(__self__, *, :param pulumi.Input[str] share_name: For Delta Sharing Catalogs: the name of the share under the share provider. Change forces creation of a new resource. :param pulumi.Input[str] storage_root: Managed location of the catalog. Location in cloud storage where data for managed tables will be stored. If not specified, the location will default to the metastore root location. Change forces creation of a new resource. """ + _CatalogState._configure( + lambda key, value: pulumi.set(__self__, key, value), + comment=comment, + connection_name=connection_name, + force_destroy=force_destroy, + isolation_mode=isolation_mode, + metastore_id=metastore_id, + name=name, + options=options, + owner=owner, + properties=properties, + provider_name=provider_name, + share_name=share_name, + storage_root=storage_root, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + comment: Optional[pulumi.Input[str]] = None, + connection_name: Optional[pulumi.Input[str]] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + isolation_mode: Optional[pulumi.Input[str]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + options: Optional[pulumi.Input[Mapping[str, Any]]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + provider_name: Optional[pulumi.Input[str]] = None, + share_name: Optional[pulumi.Input[str]] = None, + storage_root: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if connection_name is not None: - pulumi.set(__self__, "connection_name", connection_name) + _setter("connection_name", connection_name) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if isolation_mode is not None: - pulumi.set(__self__, "isolation_mode", isolation_mode) + _setter("isolation_mode", isolation_mode) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if options is not None: - pulumi.set(__self__, "options", options) + _setter("options", options) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if provider_name is not None: - pulumi.set(__self__, "provider_name", provider_name) + _setter("provider_name", provider_name) if share_name is not None: - pulumi.set(__self__, "share_name", share_name) + _setter("share_name", share_name) if storage_root is not None: - pulumi.set(__self__, "storage_root", storage_root) + _setter("storage_root", storage_root) @property @pulumi.getter @@ -511,6 +573,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + CatalogArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/catalog_workspace_binding.py b/sdk/python/pulumi_databricks/catalog_workspace_binding.py index cc3ce20d..565f7031 100644 --- a/sdk/python/pulumi_databricks/catalog_workspace_binding.py +++ b/sdk/python/pulumi_databricks/catalog_workspace_binding.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['CatalogWorkspaceBindingArgs', 'CatalogWorkspaceBinding'] @@ -21,8 +21,19 @@ def __init__(__self__, *, :param pulumi.Input[str] catalog_name: Name of Catalog. Change forces creation of a new resource. :param pulumi.Input[str] workspace_id: ID of the workspace. Change forces creation of a new resource. """ - pulumi.set(__self__, "catalog_name", catalog_name) - pulumi.set(__self__, "workspace_id", workspace_id) + CatalogWorkspaceBindingArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + workspace_id=workspace_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: pulumi.Input[str], + workspace_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("catalog_name", catalog_name) + _setter("workspace_id", workspace_id) @property @pulumi.getter(name="catalogName") @@ -59,10 +70,21 @@ def __init__(__self__, *, :param pulumi.Input[str] catalog_name: Name of Catalog. Change forces creation of a new resource. :param pulumi.Input[str] workspace_id: ID of the workspace. Change forces creation of a new resource. """ + _CatalogWorkspaceBindingState._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + workspace_id=workspace_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: Optional[pulumi.Input[str]] = None, + workspace_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if catalog_name is not None: - pulumi.set(__self__, "catalog_name", catalog_name) + _setter("catalog_name", catalog_name) if workspace_id is not None: - pulumi.set(__self__, "workspace_id", workspace_id) + _setter("workspace_id", workspace_id) @property @pulumi.getter(name="catalogName") @@ -152,6 +174,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + CatalogWorkspaceBindingArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/cluster.py b/sdk/python/pulumi_databricks/cluster.py index 56598ab1..c515e62e 100644 --- a/sdk/python/pulumi_databricks/cluster.py +++ b/sdk/python/pulumi_databricks/cluster.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -95,72 +95,143 @@ def __init__(__self__, *, :param pulumi.Input[Mapping[str, Any]] spark_env_vars: Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers. :param pulumi.Input[Sequence[pulumi.Input[str]]] ssh_public_keys: SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys. """ - pulumi.set(__self__, "spark_version", spark_version) + ClusterArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + is_pinned=is_pinned, + libraries=libraries, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + spark_version: pulumi.Input[str], + apply_policy_default_values: Optional[pulumi.Input[bool]] = None, + autoscale: Optional[pulumi.Input['ClusterAutoscaleArgs']] = None, + autotermination_minutes: Optional[pulumi.Input[int]] = None, + aws_attributes: Optional[pulumi.Input['ClusterAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['ClusterAzureAttributesArgs']] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + cluster_log_conf: Optional[pulumi.Input['ClusterClusterLogConfArgs']] = None, + cluster_mount_infos: Optional[pulumi.Input[Sequence[pulumi.Input['ClusterClusterMountInfoArgs']]]] = None, + cluster_name: Optional[pulumi.Input[str]] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + data_security_mode: Optional[pulumi.Input[str]] = None, + docker_image: Optional[pulumi.Input['ClusterDockerImageArgs']] = None, + driver_instance_pool_id: Optional[pulumi.Input[str]] = None, + driver_node_type_id: Optional[pulumi.Input[str]] = None, + enable_elastic_disk: Optional[pulumi.Input[bool]] = None, + enable_local_disk_encryption: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['ClusterGcpAttributesArgs']] = None, + idempotency_token: Optional[pulumi.Input[str]] = None, + init_scripts: Optional[pulumi.Input[Sequence[pulumi.Input['ClusterInitScriptArgs']]]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + is_pinned: Optional[pulumi.Input[bool]] = None, + libraries: Optional[pulumi.Input[Sequence[pulumi.Input['ClusterLibraryArgs']]]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + num_workers: Optional[pulumi.Input[int]] = None, + policy_id: Optional[pulumi.Input[str]] = None, + runtime_engine: Optional[pulumi.Input[str]] = None, + single_user_name: Optional[pulumi.Input[str]] = None, + spark_conf: Optional[pulumi.Input[Mapping[str, Any]]] = None, + spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, + ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + workload_type: Optional[pulumi.Input['ClusterWorkloadTypeArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: warnings.warn("""cluster_mount_info block is deprecated due the Clusters API changes.""", DeprecationWarning) pulumi.log.warn("""cluster_mount_infos is deprecated: cluster_mount_info block is deprecated due the Clusters API changes.""") if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if is_pinned is not None: - pulumi.set(__self__, "is_pinned", is_pinned) + _setter("is_pinned", is_pinned) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="sparkVersion") @@ -625,79 +696,156 @@ def __init__(__self__, *, :param pulumi.Input[Sequence[pulumi.Input[str]]] ssh_public_keys: SSH public key contents that will be added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu on port 2200. You can specify up to 10 keys. :param pulumi.Input[str] state: (string) State of the cluster. """ + _ClusterState._configure( + lambda key, value: pulumi.set(__self__, key, value), + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + default_tags=default_tags, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + is_pinned=is_pinned, + libraries=libraries, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + spark_version=spark_version, + ssh_public_keys=ssh_public_keys, + state=state, + url=url, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + apply_policy_default_values: Optional[pulumi.Input[bool]] = None, + autoscale: Optional[pulumi.Input['ClusterAutoscaleArgs']] = None, + autotermination_minutes: Optional[pulumi.Input[int]] = None, + aws_attributes: Optional[pulumi.Input['ClusterAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['ClusterAzureAttributesArgs']] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + cluster_log_conf: Optional[pulumi.Input['ClusterClusterLogConfArgs']] = None, + cluster_mount_infos: Optional[pulumi.Input[Sequence[pulumi.Input['ClusterClusterMountInfoArgs']]]] = None, + cluster_name: Optional[pulumi.Input[str]] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + data_security_mode: Optional[pulumi.Input[str]] = None, + default_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + docker_image: Optional[pulumi.Input['ClusterDockerImageArgs']] = None, + driver_instance_pool_id: Optional[pulumi.Input[str]] = None, + driver_node_type_id: Optional[pulumi.Input[str]] = None, + enable_elastic_disk: Optional[pulumi.Input[bool]] = None, + enable_local_disk_encryption: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['ClusterGcpAttributesArgs']] = None, + idempotency_token: Optional[pulumi.Input[str]] = None, + init_scripts: Optional[pulumi.Input[Sequence[pulumi.Input['ClusterInitScriptArgs']]]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + is_pinned: Optional[pulumi.Input[bool]] = None, + libraries: Optional[pulumi.Input[Sequence[pulumi.Input['ClusterLibraryArgs']]]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + num_workers: Optional[pulumi.Input[int]] = None, + policy_id: Optional[pulumi.Input[str]] = None, + runtime_engine: Optional[pulumi.Input[str]] = None, + single_user_name: Optional[pulumi.Input[str]] = None, + spark_conf: Optional[pulumi.Input[Mapping[str, Any]]] = None, + spark_env_vars: Optional[pulumi.Input[Mapping[str, Any]]] = None, + spark_version: Optional[pulumi.Input[str]] = None, + ssh_public_keys: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + state: Optional[pulumi.Input[str]] = None, + url: Optional[pulumi.Input[str]] = None, + workload_type: Optional[pulumi.Input['ClusterWorkloadTypeArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: warnings.warn("""cluster_mount_info block is deprecated due the Clusters API changes.""", DeprecationWarning) pulumi.log.warn("""cluster_mount_infos is deprecated: cluster_mount_info block is deprecated due the Clusters API changes.""") if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if default_tags is not None: - pulumi.set(__self__, "default_tags", default_tags) + _setter("default_tags", default_tags) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if is_pinned is not None: - pulumi.set(__self__, "is_pinned", is_pinned) + _setter("is_pinned", is_pinned) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if spark_version is not None: - pulumi.set(__self__, "spark_version", spark_version) + _setter("spark_version", spark_version) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if state is not None: - pulumi.set(__self__, "state", state) + _setter("state", state) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="applyPolicyDefaultValues") @@ -1227,6 +1375,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ClusterArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -1274,24 +1426,51 @@ def _internal_init(__self__, __props__ = ClusterArgs.__new__(ClusterArgs) __props__.__dict__["apply_policy_default_values"] = apply_policy_default_values + if autoscale is not None and not isinstance(autoscale, ClusterAutoscaleArgs): + autoscale = autoscale or {} + def _setter(key, value): + autoscale[key] = value + ClusterAutoscaleArgs._configure(_setter, **autoscale) __props__.__dict__["autoscale"] = autoscale __props__.__dict__["autotermination_minutes"] = autotermination_minutes + if aws_attributes is not None and not isinstance(aws_attributes, ClusterAwsAttributesArgs): + aws_attributes = aws_attributes or {} + def _setter(key, value): + aws_attributes[key] = value + ClusterAwsAttributesArgs._configure(_setter, **aws_attributes) __props__.__dict__["aws_attributes"] = aws_attributes + if azure_attributes is not None and not isinstance(azure_attributes, ClusterAzureAttributesArgs): + azure_attributes = azure_attributes or {} + def _setter(key, value): + azure_attributes[key] = value + ClusterAzureAttributesArgs._configure(_setter, **azure_attributes) __props__.__dict__["azure_attributes"] = azure_attributes __props__.__dict__["cluster_id"] = cluster_id + if cluster_log_conf is not None and not isinstance(cluster_log_conf, ClusterClusterLogConfArgs): + cluster_log_conf = cluster_log_conf or {} + def _setter(key, value): + cluster_log_conf[key] = value + ClusterClusterLogConfArgs._configure(_setter, **cluster_log_conf) __props__.__dict__["cluster_log_conf"] = cluster_log_conf - if cluster_mount_infos is not None and not opts.urn: - warnings.warn("""cluster_mount_info block is deprecated due the Clusters API changes.""", DeprecationWarning) - pulumi.log.warn("""cluster_mount_infos is deprecated: cluster_mount_info block is deprecated due the Clusters API changes.""") __props__.__dict__["cluster_mount_infos"] = cluster_mount_infos __props__.__dict__["cluster_name"] = cluster_name __props__.__dict__["custom_tags"] = custom_tags __props__.__dict__["data_security_mode"] = data_security_mode + if docker_image is not None and not isinstance(docker_image, ClusterDockerImageArgs): + docker_image = docker_image or {} + def _setter(key, value): + docker_image[key] = value + ClusterDockerImageArgs._configure(_setter, **docker_image) __props__.__dict__["docker_image"] = docker_image __props__.__dict__["driver_instance_pool_id"] = driver_instance_pool_id __props__.__dict__["driver_node_type_id"] = driver_node_type_id __props__.__dict__["enable_elastic_disk"] = enable_elastic_disk __props__.__dict__["enable_local_disk_encryption"] = enable_local_disk_encryption + if gcp_attributes is not None and not isinstance(gcp_attributes, ClusterGcpAttributesArgs): + gcp_attributes = gcp_attributes or {} + def _setter(key, value): + gcp_attributes[key] = value + ClusterGcpAttributesArgs._configure(_setter, **gcp_attributes) __props__.__dict__["gcp_attributes"] = gcp_attributes __props__.__dict__["idempotency_token"] = idempotency_token __props__.__dict__["init_scripts"] = init_scripts @@ -1309,6 +1488,11 @@ def _internal_init(__self__, raise TypeError("Missing required property 'spark_version'") __props__.__dict__["spark_version"] = spark_version __props__.__dict__["ssh_public_keys"] = ssh_public_keys + if workload_type is not None and not isinstance(workload_type, ClusterWorkloadTypeArgs): + workload_type = workload_type or {} + def _setter(key, value): + workload_type[key] = value + ClusterWorkloadTypeArgs._configure(_setter, **workload_type) __props__.__dict__["workload_type"] = workload_type __props__.__dict__["default_tags"] = None __props__.__dict__["state"] = None diff --git a/sdk/python/pulumi_databricks/cluster_policy.py b/sdk/python/pulumi_databricks/cluster_policy.py index 6d4f2008..c6f2b97b 100644 --- a/sdk/python/pulumi_databricks/cluster_policy.py +++ b/sdk/python/pulumi_databricks/cluster_policy.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['ClusterPolicyArgs', 'ClusterPolicy'] @@ -29,18 +29,37 @@ def __init__(__self__, *, :param pulumi.Input[str] policy_family_definition_overrides: Policy definition JSON document expressed in Databricks Policy Definition Language. The JSON document must be passed as a string and cannot be embedded in the requests. You can use this to customize the policy definition inherited from the policy family. Policy rules specified here are merged into the inherited policy definition. :param pulumi.Input[str] policy_family_id: ID of the policy family. The cluster policy's policy definition inherits the policy family's policy definition. Cannot be used with `definition`. Use `policy_family_definition_overrides` instead to customize the policy definition. """ + ClusterPolicyArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + definition=definition, + description=description, + max_clusters_per_user=max_clusters_per_user, + name=name, + policy_family_definition_overrides=policy_family_definition_overrides, + policy_family_id=policy_family_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + definition: Optional[pulumi.Input[str]] = None, + description: Optional[pulumi.Input[str]] = None, + max_clusters_per_user: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + policy_family_definition_overrides: Optional[pulumi.Input[str]] = None, + policy_family_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if definition is not None: - pulumi.set(__self__, "definition", definition) + _setter("definition", definition) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if max_clusters_per_user is not None: - pulumi.set(__self__, "max_clusters_per_user", max_clusters_per_user) + _setter("max_clusters_per_user", max_clusters_per_user) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if policy_family_definition_overrides is not None: - pulumi.set(__self__, "policy_family_definition_overrides", policy_family_definition_overrides) + _setter("policy_family_definition_overrides", policy_family_definition_overrides) if policy_family_id is not None: - pulumi.set(__self__, "policy_family_id", policy_family_id) + _setter("policy_family_id", policy_family_id) @property @pulumi.getter @@ -135,20 +154,41 @@ def __init__(__self__, *, :param pulumi.Input[str] policy_family_id: ID of the policy family. The cluster policy's policy definition inherits the policy family's policy definition. Cannot be used with `definition`. Use `policy_family_definition_overrides` instead to customize the policy definition. :param pulumi.Input[str] policy_id: Canonical unique identifier for the cluster policy. """ + _ClusterPolicyState._configure( + lambda key, value: pulumi.set(__self__, key, value), + definition=definition, + description=description, + max_clusters_per_user=max_clusters_per_user, + name=name, + policy_family_definition_overrides=policy_family_definition_overrides, + policy_family_id=policy_family_id, + policy_id=policy_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + definition: Optional[pulumi.Input[str]] = None, + description: Optional[pulumi.Input[str]] = None, + max_clusters_per_user: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + policy_family_definition_overrides: Optional[pulumi.Input[str]] = None, + policy_family_id: Optional[pulumi.Input[str]] = None, + policy_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if definition is not None: - pulumi.set(__self__, "definition", definition) + _setter("definition", definition) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if max_clusters_per_user is not None: - pulumi.set(__self__, "max_clusters_per_user", max_clusters_per_user) + _setter("max_clusters_per_user", max_clusters_per_user) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if policy_family_definition_overrides is not None: - pulumi.set(__self__, "policy_family_definition_overrides", policy_family_definition_overrides) + _setter("policy_family_definition_overrides", policy_family_definition_overrides) if policy_family_id is not None: - pulumi.set(__self__, "policy_family_id", policy_family_id) + _setter("policy_family_id", policy_family_id) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) @property @pulumi.getter @@ -360,6 +400,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ClusterPolicyArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/config/__init__.pyi b/sdk/python/pulumi_databricks/config/__init__.pyi index 89961853..e8ef74bd 100644 --- a/sdk/python/pulumi_databricks/config/__init__.pyi +++ b/sdk/python/pulumi_databricks/config/__init__.pyi @@ -6,7 +6,7 @@ import copy import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from .. import _utilities accountId: Optional[str] diff --git a/sdk/python/pulumi_databricks/config/vars.py b/sdk/python/pulumi_databricks/config/vars.py index 39f54a4a..518b8772 100644 --- a/sdk/python/pulumi_databricks/config/vars.py +++ b/sdk/python/pulumi_databricks/config/vars.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from .. import _utilities import types diff --git a/sdk/python/pulumi_databricks/connection.py b/sdk/python/pulumi_databricks/connection.py index c9bbe80a..409800fe 100644 --- a/sdk/python/pulumi_databricks/connection.py +++ b/sdk/python/pulumi_databricks/connection.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['ConnectionArgs', 'Connection'] @@ -31,20 +31,43 @@ def __init__(__self__, *, :param pulumi.Input[str] owner: Name of the connection owner. :param pulumi.Input[Mapping[str, Any]] properties: Free-form connection properties. """ - pulumi.set(__self__, "connection_type", connection_type) - pulumi.set(__self__, "options", options) + ConnectionArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + connection_type=connection_type, + options=options, + comment=comment, + metastore_id=metastore_id, + name=name, + owner=owner, + properties=properties, + read_only=read_only, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + connection_type: pulumi.Input[str], + options: pulumi.Input[Mapping[str, Any]], + comment: Optional[pulumi.Input[str]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("connection_type", connection_type) + _setter("options", options) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) @property @pulumi.getter(name="connectionType") @@ -157,22 +180,45 @@ def __init__(__self__, *, :param pulumi.Input[str] owner: Name of the connection owner. :param pulumi.Input[Mapping[str, Any]] properties: Free-form connection properties. """ + _ConnectionState._configure( + lambda key, value: pulumi.set(__self__, key, value), + comment=comment, + connection_type=connection_type, + metastore_id=metastore_id, + name=name, + options=options, + owner=owner, + properties=properties, + read_only=read_only, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + comment: Optional[pulumi.Input[str]] = None, + connection_type: Optional[pulumi.Input[str]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + options: Optional[pulumi.Input[Mapping[str, Any]]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if connection_type is not None: - pulumi.set(__self__, "connection_type", connection_type) + _setter("connection_type", connection_type) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if options is not None: - pulumi.set(__self__, "options", options) + _setter("options", options) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) @property @pulumi.getter @@ -376,6 +422,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ConnectionArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/dbfs_file.py b/sdk/python/pulumi_databricks/dbfs_file.py index 6e0f4b4b..0f8d7cd8 100644 --- a/sdk/python/pulumi_databricks/dbfs_file.py +++ b/sdk/python/pulumi_databricks/dbfs_file.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['DbfsFileArgs', 'DbfsFile'] @@ -23,13 +23,28 @@ def __init__(__self__, *, :param pulumi.Input[str] path: The path of the file in which you wish to save. :param pulumi.Input[str] source: The full absolute path to the file. Conflicts with `content_base64`. """ - pulumi.set(__self__, "path", path) + DbfsFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + content_base64=content_base64, + md5=md5, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + content_base64: Optional[pulumi.Input[str]] = None, + md5: Optional[pulumi.Input[str]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter @@ -90,18 +105,37 @@ def __init__(__self__, *, :param pulumi.Input[str] path: The path of the file in which you wish to save. :param pulumi.Input[str] source: The full absolute path to the file. Conflicts with `content_base64`. """ + _DbfsFileState._configure( + lambda key, value: pulumi.set(__self__, key, value), + content_base64=content_base64, + dbfs_path=dbfs_path, + file_size=file_size, + md5=md5, + path=path, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + content_base64: Optional[pulumi.Input[str]] = None, + dbfs_path: Optional[pulumi.Input[str]] = None, + file_size: Optional[pulumi.Input[int]] = None, + md5: Optional[pulumi.Input[str]] = None, + path: Optional[pulumi.Input[str]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if dbfs_path is not None: - pulumi.set(__self__, "dbfs_path", dbfs_path) + _setter("dbfs_path", dbfs_path) if file_size is not None: - pulumi.set(__self__, "file_size", file_size) + _setter("file_size", file_size) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="contentBase64") @@ -219,6 +253,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + DbfsFileArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/directory.py b/sdk/python/pulumi_databricks/directory.py index 315fe39a..98c83f7e 100644 --- a/sdk/python/pulumi_databricks/directory.py +++ b/sdk/python/pulumi_databricks/directory.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['DirectoryArgs', 'Directory'] @@ -22,11 +22,24 @@ def __init__(__self__, *, :param pulumi.Input[str] path: The absolute path of the directory, beginning with "/", e.g. "/Demo". :param pulumi.Input[int] object_id: Unique identifier for a DIRECTORY """ - pulumi.set(__self__, "path", path) + DirectoryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + delete_recursive=delete_recursive, + object_id=object_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + delete_recursive: Optional[pulumi.Input[bool]] = None, + object_id: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) if delete_recursive is not None: - pulumi.set(__self__, "delete_recursive", delete_recursive) + _setter("delete_recursive", delete_recursive) if object_id is not None: - pulumi.set(__self__, "object_id", object_id) + _setter("object_id", object_id) @property @pulumi.getter @@ -73,12 +86,25 @@ def __init__(__self__, *, :param pulumi.Input[int] object_id: Unique identifier for a DIRECTORY :param pulumi.Input[str] path: The absolute path of the directory, beginning with "/", e.g. "/Demo". """ + _DirectoryState._configure( + lambda key, value: pulumi.set(__self__, key, value), + delete_recursive=delete_recursive, + object_id=object_id, + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + delete_recursive: Optional[pulumi.Input[bool]] = None, + object_id: Optional[pulumi.Input[int]] = None, + path: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if delete_recursive is not None: - pulumi.set(__self__, "delete_recursive", delete_recursive) + _setter("delete_recursive", delete_recursive) if object_id is not None: - pulumi.set(__self__, "object_id", object_id) + _setter("object_id", object_id) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) @property @pulumi.getter(name="deleteRecursive") @@ -162,6 +188,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + DirectoryArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/entitlements.py b/sdk/python/pulumi_databricks/entitlements.py index 8cedbace..67d5a299 100644 --- a/sdk/python/pulumi_databricks/entitlements.py +++ b/sdk/python/pulumi_databricks/entitlements.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['EntitlementsArgs', 'Entitlements'] @@ -33,20 +33,41 @@ def __init__(__self__, *, :param pulumi.Input[str] user_id: Canonical unique identifier for the user. :param pulumi.Input[bool] workspace_access: This is a field to allow the principal to have access to Databricks Workspace. """ + EntitlementsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + databricks_sql_access=databricks_sql_access, + group_id=group_id, + service_principal_id=service_principal_id, + user_id=user_id, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + group_id: Optional[pulumi.Input[str]] = None, + service_principal_id: Optional[pulumi.Input[str]] = None, + user_id: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if group_id is not None: - pulumi.set(__self__, "group_id", group_id) + _setter("group_id", group_id) if service_principal_id is not None: - pulumi.set(__self__, "service_principal_id", service_principal_id) + _setter("service_principal_id", service_principal_id) if user_id is not None: - pulumi.set(__self__, "user_id", user_id) + _setter("user_id", user_id) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="allowClusterCreate") @@ -157,20 +178,41 @@ def __init__(__self__, *, :param pulumi.Input[str] user_id: Canonical unique identifier for the user. :param pulumi.Input[bool] workspace_access: This is a field to allow the principal to have access to Databricks Workspace. """ + _EntitlementsState._configure( + lambda key, value: pulumi.set(__self__, key, value), + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + databricks_sql_access=databricks_sql_access, + group_id=group_id, + service_principal_id=service_principal_id, + user_id=user_id, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + group_id: Optional[pulumi.Input[str]] = None, + service_principal_id: Optional[pulumi.Input[str]] = None, + user_id: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if group_id is not None: - pulumi.set(__self__, "group_id", group_id) + _setter("group_id", group_id) if service_principal_id is not None: - pulumi.set(__self__, "service_principal_id", service_principal_id) + _setter("service_principal_id", service_principal_id) if user_id is not None: - pulumi.set(__self__, "user_id", user_id) + _setter("user_id", user_id) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="allowClusterCreate") @@ -430,6 +472,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + EntitlementsArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/external_location.py b/sdk/python/pulumi_databricks/external_location.py index 3fb998e8..ad36ff25 100644 --- a/sdk/python/pulumi_databricks/external_location.py +++ b/sdk/python/pulumi_databricks/external_location.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -42,28 +42,59 @@ def __init__(__self__, *, :param pulumi.Input[bool] read_only: Indicates whether the external location is read-only. :param pulumi.Input[bool] skip_validation: Suppress validation errors if any & force save the external location """ - pulumi.set(__self__, "credential_name", credential_name) - pulumi.set(__self__, "url", url) + ExternalLocationArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + credential_name=credential_name, + url=url, + access_point=access_point, + comment=comment, + encryption_details=encryption_details, + force_destroy=force_destroy, + force_update=force_update, + metastore_id=metastore_id, + name=name, + owner=owner, + read_only=read_only, + skip_validation=skip_validation, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + credential_name: pulumi.Input[str], + url: pulumi.Input[str], + access_point: Optional[pulumi.Input[str]] = None, + comment: Optional[pulumi.Input[str]] = None, + encryption_details: Optional[pulumi.Input['ExternalLocationEncryptionDetailsArgs']] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + force_update: Optional[pulumi.Input[bool]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + skip_validation: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("credential_name", credential_name) + _setter("url", url) if access_point is not None: - pulumi.set(__self__, "access_point", access_point) + _setter("access_point", access_point) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if encryption_details is not None: - pulumi.set(__self__, "encryption_details", encryption_details) + _setter("encryption_details", encryption_details) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if force_update is not None: - pulumi.set(__self__, "force_update", force_update) + _setter("force_update", force_update) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) if skip_validation is not None: - pulumi.set(__self__, "skip_validation", skip_validation) + _setter("skip_validation", skip_validation) @property @pulumi.getter(name="credentialName") @@ -236,30 +267,61 @@ def __init__(__self__, *, :param pulumi.Input[bool] skip_validation: Suppress validation errors if any & force save the external location :param pulumi.Input[str] url: Path URL in cloud storage, of the form: `s3://[bucket-host]/[bucket-dir]` (AWS), `abfss://[user]@[host]/[path]` (Azure), `gs://[bucket-host]/[bucket-dir]` (GCP). """ + _ExternalLocationState._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_point=access_point, + comment=comment, + credential_name=credential_name, + encryption_details=encryption_details, + force_destroy=force_destroy, + force_update=force_update, + metastore_id=metastore_id, + name=name, + owner=owner, + read_only=read_only, + skip_validation=skip_validation, + url=url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_point: Optional[pulumi.Input[str]] = None, + comment: Optional[pulumi.Input[str]] = None, + credential_name: Optional[pulumi.Input[str]] = None, + encryption_details: Optional[pulumi.Input['ExternalLocationEncryptionDetailsArgs']] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + force_update: Optional[pulumi.Input[bool]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + skip_validation: Optional[pulumi.Input[bool]] = None, + url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if access_point is not None: - pulumi.set(__self__, "access_point", access_point) + _setter("access_point", access_point) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if credential_name is not None: - pulumi.set(__self__, "credential_name", credential_name) + _setter("credential_name", credential_name) if encryption_details is not None: - pulumi.set(__self__, "encryption_details", encryption_details) + _setter("encryption_details", encryption_details) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if force_update is not None: - pulumi.set(__self__, "force_update", force_update) + _setter("force_update", force_update) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) if skip_validation is not None: - pulumi.set(__self__, "skip_validation", skip_validation) + _setter("skip_validation", skip_validation) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) @property @pulumi.getter(name="accessPoint") @@ -479,6 +541,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ExternalLocationArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -510,6 +576,11 @@ def _internal_init(__self__, if credential_name is None and not opts.urn: raise TypeError("Missing required property 'credential_name'") __props__.__dict__["credential_name"] = credential_name + if encryption_details is not None and not isinstance(encryption_details, ExternalLocationEncryptionDetailsArgs): + encryption_details = encryption_details or {} + def _setter(key, value): + encryption_details[key] = value + ExternalLocationEncryptionDetailsArgs._configure(_setter, **encryption_details) __props__.__dict__["encryption_details"] = encryption_details __props__.__dict__["force_destroy"] = force_destroy __props__.__dict__["force_update"] = force_update diff --git a/sdk/python/pulumi_databricks/get_aws_assume_role_policy.py b/sdk/python/pulumi_databricks/get_aws_assume_role_policy.py index 22b828f8..1e822b5e 100644 --- a/sdk/python/pulumi_databricks/get_aws_assume_role_policy.py +++ b/sdk/python/pulumi_databricks/get_aws_assume_role_policy.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_aws_bucket_policy.py b/sdk/python/pulumi_databricks/get_aws_bucket_policy.py index c3b43291..3cfc02ce 100644 --- a/sdk/python/pulumi_databricks/get_aws_bucket_policy.py +++ b/sdk/python/pulumi_databricks/get_aws_bucket_policy.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_aws_cross_account_policy.py b/sdk/python/pulumi_databricks/get_aws_cross_account_policy.py index b66ff5e5..75e53dbe 100644 --- a/sdk/python/pulumi_databricks/get_aws_cross_account_policy.py +++ b/sdk/python/pulumi_databricks/get_aws_cross_account_policy.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_catalogs.py b/sdk/python/pulumi_databricks/get_catalogs.py index 9ae77634..cba84ca5 100644 --- a/sdk/python/pulumi_databricks/get_catalogs.py +++ b/sdk/python/pulumi_databricks/get_catalogs.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_cluster.py b/sdk/python/pulumi_databricks/get_cluster.py index 49588594..694c67b9 100644 --- a/sdk/python/pulumi_databricks/get_cluster.py +++ b/sdk/python/pulumi_databricks/get_cluster.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * diff --git a/sdk/python/pulumi_databricks/get_cluster_policy.py b/sdk/python/pulumi_databricks/get_cluster_policy.py index e73211e8..92afd2cc 100644 --- a/sdk/python/pulumi_databricks/get_cluster_policy.py +++ b/sdk/python/pulumi_databricks/get_cluster_policy.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_clusters.py b/sdk/python/pulumi_databricks/get_clusters.py index 120372db..05cc7dd4 100644 --- a/sdk/python/pulumi_databricks/get_clusters.py +++ b/sdk/python/pulumi_databricks/get_clusters.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_current_user.py b/sdk/python/pulumi_databricks/get_current_user.py index 9d7396f3..717c6670 100644 --- a/sdk/python/pulumi_databricks/get_current_user.py +++ b/sdk/python/pulumi_databricks/get_current_user.py @@ -6,13 +6,14 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ 'GetCurrentUserResult', 'AwaitableGetCurrentUserResult', 'get_current_user', + 'get_current_user_output', ] @pulumi.output_type @@ -143,3 +144,31 @@ def get_current_user(opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGe repos=pulumi.get(__ret__, 'repos'), user_name=pulumi.get(__ret__, 'user_name'), workspace_url=pulumi.get(__ret__, 'workspace_url')) + + +@_utilities.lift_output_func(get_current_user) +def get_current_user_output(opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetCurrentUserResult]: + """ + ## Exported attributes + + Data source exposes the following attributes: + + * `id` - The id of the calling user. + * `external_id` - ID of the user in an external identity provider. + * `user_name` - Name of the user, e.g. `mr.foo@example.com`. If the currently logged-in identity is a service principal, returns the application ID, e.g. `11111111-2222-3333-4444-555666777888` + * `home` - Home folder of the user, e.g. `/Users/mr.foo@example.com`. + * `repos` - Personal Repos location of the user, e.g. `/Repos/mr.foo@example.com`. + * `alphanumeric` - Alphanumeric representation of user local name. e.g. `mr_foo`. + * `workspace_url` - URL of the current Databricks workspace. + * `acl_principal_id` - identifier for use in databricks_access_control_rule_set, e.g. `users/mr.foo@example.com` if current user is user, or `servicePrincipals/00000000-0000-0000-0000-000000000000` if current user is service principal. + + ## Related Resources + + The following resources are used in the same context: + + * End to end workspace management guide + * Directory to manage directories in [Databricks Workpace](https://docs.databricks.com/workspace/workspace-objects.html). + * Notebook to manage [Databricks Notebooks](https://docs.databricks.com/notebooks/index.html). + * Repo to manage [Databricks Repos](https://docs.databricks.com/repos.html). + """ + ... diff --git a/sdk/python/pulumi_databricks/get_dbfs_file.py b/sdk/python/pulumi_databricks/get_dbfs_file.py index de5ea301..e74cb5ec 100644 --- a/sdk/python/pulumi_databricks/get_dbfs_file.py +++ b/sdk/python/pulumi_databricks/get_dbfs_file.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_dbfs_file_paths.py b/sdk/python/pulumi_databricks/get_dbfs_file_paths.py index 2a6744b7..dd0a968f 100644 --- a/sdk/python/pulumi_databricks/get_dbfs_file_paths.py +++ b/sdk/python/pulumi_databricks/get_dbfs_file_paths.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs diff --git a/sdk/python/pulumi_databricks/get_directory.py b/sdk/python/pulumi_databricks/get_directory.py index 55b0551f..87a69558 100644 --- a/sdk/python/pulumi_databricks/get_directory.py +++ b/sdk/python/pulumi_databricks/get_directory.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_group.py b/sdk/python/pulumi_databricks/get_group.py index 4c02b4ee..dadde1ea 100644 --- a/sdk/python/pulumi_databricks/get_group.py +++ b/sdk/python/pulumi_databricks/get_group.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_instance_pool.py b/sdk/python/pulumi_databricks/get_instance_pool.py index d2cad66b..223f9b66 100644 --- a/sdk/python/pulumi_databricks/get_instance_pool.py +++ b/sdk/python/pulumi_databricks/get_instance_pool.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * diff --git a/sdk/python/pulumi_databricks/get_job.py b/sdk/python/pulumi_databricks/get_job.py index e64a2a8d..a00c3d99 100644 --- a/sdk/python/pulumi_databricks/get_job.py +++ b/sdk/python/pulumi_databricks/get_job.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * diff --git a/sdk/python/pulumi_databricks/get_jobs.py b/sdk/python/pulumi_databricks/get_jobs.py index 8b3c1869..92c02469 100644 --- a/sdk/python/pulumi_databricks/get_jobs.py +++ b/sdk/python/pulumi_databricks/get_jobs.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_metastore.py b/sdk/python/pulumi_databricks/get_metastore.py index b8969e5f..ca522ff7 100644 --- a/sdk/python/pulumi_databricks/get_metastore.py +++ b/sdk/python/pulumi_databricks/get_metastore.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * diff --git a/sdk/python/pulumi_databricks/get_metastores.py b/sdk/python/pulumi_databricks/get_metastores.py index 65b8f03f..a178c973 100644 --- a/sdk/python/pulumi_databricks/get_metastores.py +++ b/sdk/python/pulumi_databricks/get_metastores.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_mws_credentials.py b/sdk/python/pulumi_databricks/get_mws_credentials.py index 85ade29f..577eee63 100644 --- a/sdk/python/pulumi_databricks/get_mws_credentials.py +++ b/sdk/python/pulumi_databricks/get_mws_credentials.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_mws_workspaces.py b/sdk/python/pulumi_databricks/get_mws_workspaces.py index f143cb85..9b4552b2 100644 --- a/sdk/python/pulumi_databricks/get_mws_workspaces.py +++ b/sdk/python/pulumi_databricks/get_mws_workspaces.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_node_type.py b/sdk/python/pulumi_databricks/get_node_type.py index aa78704c..ab681a9e 100644 --- a/sdk/python/pulumi_databricks/get_node_type.py +++ b/sdk/python/pulumi_databricks/get_node_type.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_notebook.py b/sdk/python/pulumi_databricks/get_notebook.py index 0fe92953..0c93ceeb 100644 --- a/sdk/python/pulumi_databricks/get_notebook.py +++ b/sdk/python/pulumi_databricks/get_notebook.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_notebook_paths.py b/sdk/python/pulumi_databricks/get_notebook_paths.py index 16afeefe..26c399e2 100644 --- a/sdk/python/pulumi_databricks/get_notebook_paths.py +++ b/sdk/python/pulumi_databricks/get_notebook_paths.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs diff --git a/sdk/python/pulumi_databricks/get_pipelines.py b/sdk/python/pulumi_databricks/get_pipelines.py index c282c981..9746e3d5 100644 --- a/sdk/python/pulumi_databricks/get_pipelines.py +++ b/sdk/python/pulumi_databricks/get_pipelines.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_schemas.py b/sdk/python/pulumi_databricks/get_schemas.py index 5a358f0c..c85af254 100644 --- a/sdk/python/pulumi_databricks/get_schemas.py +++ b/sdk/python/pulumi_databricks/get_schemas.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_service_principal.py b/sdk/python/pulumi_databricks/get_service_principal.py index 69cef6eb..10a52054 100644 --- a/sdk/python/pulumi_databricks/get_service_principal.py +++ b/sdk/python/pulumi_databricks/get_service_principal.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_service_principals.py b/sdk/python/pulumi_databricks/get_service_principals.py index 3fe42115..0617eeec 100644 --- a/sdk/python/pulumi_databricks/get_service_principals.py +++ b/sdk/python/pulumi_databricks/get_service_principals.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_share.py b/sdk/python/pulumi_databricks/get_share.py index 68044a0e..d1699247 100644 --- a/sdk/python/pulumi_databricks/get_share.py +++ b/sdk/python/pulumi_databricks/get_share.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * diff --git a/sdk/python/pulumi_databricks/get_shares.py b/sdk/python/pulumi_databricks/get_shares.py index 5b63a01c..5d6d8712 100644 --- a/sdk/python/pulumi_databricks/get_shares.py +++ b/sdk/python/pulumi_databricks/get_shares.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_spark_version.py b/sdk/python/pulumi_databricks/get_spark_version.py index ce1c852b..259cd8ad 100644 --- a/sdk/python/pulumi_databricks/get_spark_version.py +++ b/sdk/python/pulumi_databricks/get_spark_version.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_sql_warehouse.py b/sdk/python/pulumi_databricks/get_sql_warehouse.py index 63f7e1ce..31a676ae 100644 --- a/sdk/python/pulumi_databricks/get_sql_warehouse.py +++ b/sdk/python/pulumi_databricks/get_sql_warehouse.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * diff --git a/sdk/python/pulumi_databricks/get_sql_warehouses.py b/sdk/python/pulumi_databricks/get_sql_warehouses.py index 3d7cb02d..b2876acb 100644 --- a/sdk/python/pulumi_databricks/get_sql_warehouses.py +++ b/sdk/python/pulumi_databricks/get_sql_warehouses.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_tables.py b/sdk/python/pulumi_databricks/get_tables.py index 2d91cc73..b9799bf9 100644 --- a/sdk/python/pulumi_databricks/get_tables.py +++ b/sdk/python/pulumi_databricks/get_tables.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_user.py b/sdk/python/pulumi_databricks/get_user.py index 9ee6dc50..992c2068 100644 --- a/sdk/python/pulumi_databricks/get_user.py +++ b/sdk/python/pulumi_databricks/get_user.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_views.py b/sdk/python/pulumi_databricks/get_views.py index f7aade14..19327eef 100644 --- a/sdk/python/pulumi_databricks/get_views.py +++ b/sdk/python/pulumi_databricks/get_views.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ diff --git a/sdk/python/pulumi_databricks/get_zones.py b/sdk/python/pulumi_databricks/get_zones.py index 1141c1a4..89b4e63a 100644 --- a/sdk/python/pulumi_databricks/get_zones.py +++ b/sdk/python/pulumi_databricks/get_zones.py @@ -6,13 +6,14 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = [ 'GetZonesResult', 'AwaitableGetZonesResult', 'get_zones', + 'get_zones_output', ] @pulumi.output_type @@ -90,3 +91,22 @@ def get_zones(opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetZonesR default_zone=pulumi.get(__ret__, 'default_zone'), id=pulumi.get(__ret__, 'id'), zones=pulumi.get(__ret__, 'zones')) + + +@_utilities.lift_output_func(get_zones) +def get_zones_output(opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetZonesResult]: + """ + > **Note** If you have a fully automated setup with workspaces created by databricks_mws_workspaces, please make sure to add depends_on attribute in order to prevent _default auth: cannot configure default credentials_ errors. + + This data source allows you to fetch all available AWS availability zones on your workspace on AWS. + + ## Example Usage + + ```python + import pulumi + import pulumi_databricks as databricks + + zones = databricks.get_zones() + ``` + """ + ... diff --git a/sdk/python/pulumi_databricks/git_credential.py b/sdk/python/pulumi_databricks/git_credential.py index d4c5791d..cbd3c885 100644 --- a/sdk/python/pulumi_databricks/git_credential.py +++ b/sdk/python/pulumi_databricks/git_credential.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['GitCredentialArgs', 'GitCredential'] @@ -25,13 +25,28 @@ def __init__(__self__, *, :param pulumi.Input[str] git_username: user name at Git provider. :param pulumi.Input[str] personal_access_token: The personal access token used to authenticate to the corresponding Git provider. If value is not provided, it's sourced from the first environment variable of `GITHUB_TOKEN`, that has a non-empty value. """ - pulumi.set(__self__, "git_provider", git_provider) + GitCredentialArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + git_provider=git_provider, + force=force, + git_username=git_username, + personal_access_token=personal_access_token, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + git_provider: pulumi.Input[str], + force: Optional[pulumi.Input[bool]] = None, + git_username: Optional[pulumi.Input[str]] = None, + personal_access_token: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("git_provider", git_provider) if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if git_username is not None: - pulumi.set(__self__, "git_username", git_username) + _setter("git_username", git_username) if personal_access_token is not None: - pulumi.set(__self__, "personal_access_token", personal_access_token) + _setter("personal_access_token", personal_access_token) @property @pulumi.getter(name="gitProvider") @@ -96,14 +111,29 @@ def __init__(__self__, *, :param pulumi.Input[str] git_username: user name at Git provider. :param pulumi.Input[str] personal_access_token: The personal access token used to authenticate to the corresponding Git provider. If value is not provided, it's sourced from the first environment variable of `GITHUB_TOKEN`, that has a non-empty value. """ + _GitCredentialState._configure( + lambda key, value: pulumi.set(__self__, key, value), + force=force, + git_provider=git_provider, + git_username=git_username, + personal_access_token=personal_access_token, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + force: Optional[pulumi.Input[bool]] = None, + git_provider: Optional[pulumi.Input[str]] = None, + git_username: Optional[pulumi.Input[str]] = None, + personal_access_token: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if git_provider is not None: - pulumi.set(__self__, "git_provider", git_provider) + _setter("git_provider", git_provider) if git_username is not None: - pulumi.set(__self__, "git_username", git_username) + _setter("git_username", git_username) if personal_access_token is not None: - pulumi.set(__self__, "personal_access_token", personal_access_token) + _setter("personal_access_token", personal_access_token) @property @pulumi.getter @@ -205,6 +235,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + GitCredentialArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/global_init_script.py b/sdk/python/pulumi_databricks/global_init_script.py index 1d4398ea..8f68151c 100644 --- a/sdk/python/pulumi_databricks/global_init_script.py +++ b/sdk/python/pulumi_databricks/global_init_script.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['GlobalInitScriptArgs', 'GlobalInitScript'] @@ -27,18 +27,37 @@ def __init__(__self__, *, :param pulumi.Input[int] position: the position of a global init script, where `0` represents the first global init script to run, `1` is the second global init script to run, and so on. When omitted, the script gets the last position. :param pulumi.Input[str] source: Path to script's source code on local filesystem. Conflicts with `content_base64` """ + GlobalInitScriptArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + content_base64=content_base64, + enabled=enabled, + md5=md5, + name=name, + position=position, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + content_base64: Optional[pulumi.Input[str]] = None, + enabled: Optional[pulumi.Input[bool]] = None, + md5: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + position: Optional[pulumi.Input[int]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if enabled is not None: - pulumi.set(__self__, "enabled", enabled) + _setter("enabled", enabled) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if position is not None: - pulumi.set(__self__, "position", position) + _setter("position", position) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="contentBase64") @@ -123,18 +142,37 @@ def __init__(__self__, *, :param pulumi.Input[int] position: the position of a global init script, where `0` represents the first global init script to run, `1` is the second global init script to run, and so on. When omitted, the script gets the last position. :param pulumi.Input[str] source: Path to script's source code on local filesystem. Conflicts with `content_base64` """ + _GlobalInitScriptState._configure( + lambda key, value: pulumi.set(__self__, key, value), + content_base64=content_base64, + enabled=enabled, + md5=md5, + name=name, + position=position, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + content_base64: Optional[pulumi.Input[str]] = None, + enabled: Optional[pulumi.Input[bool]] = None, + md5: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + position: Optional[pulumi.Input[int]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if enabled is not None: - pulumi.set(__self__, "enabled", enabled) + _setter("enabled", enabled) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if position is not None: - pulumi.set(__self__, "position", position) + _setter("position", position) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="contentBase64") @@ -256,6 +294,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + GlobalInitScriptArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/grants.py b/sdk/python/pulumi_databricks/grants.py index 4e626284..83f8ccfd 100644 --- a/sdk/python/pulumi_databricks/grants.py +++ b/sdk/python/pulumi_databricks/grants.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -32,31 +32,64 @@ def __init__(__self__, *, """ The set of arguments for constructing a Grants resource. """ - pulumi.set(__self__, "grants", grants) + GrantsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + grants=grants, + catalog=catalog, + external_location=external_location, + foreign_connection=foreign_connection, + function=function, + materialized_view=materialized_view, + metastore=metastore, + schema=schema, + share=share, + storage_credential=storage_credential, + table=table, + view=view, + volume=volume, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + grants: pulumi.Input[Sequence[pulumi.Input['GrantsGrantArgs']]], + catalog: Optional[pulumi.Input[str]] = None, + external_location: Optional[pulumi.Input[str]] = None, + foreign_connection: Optional[pulumi.Input[str]] = None, + function: Optional[pulumi.Input[str]] = None, + materialized_view: Optional[pulumi.Input[str]] = None, + metastore: Optional[pulumi.Input[str]] = None, + schema: Optional[pulumi.Input[str]] = None, + share: Optional[pulumi.Input[str]] = None, + storage_credential: Optional[pulumi.Input[str]] = None, + table: Optional[pulumi.Input[str]] = None, + view: Optional[pulumi.Input[str]] = None, + volume: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("grants", grants) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if external_location is not None: - pulumi.set(__self__, "external_location", external_location) + _setter("external_location", external_location) if foreign_connection is not None: - pulumi.set(__self__, "foreign_connection", foreign_connection) + _setter("foreign_connection", foreign_connection) if function is not None: - pulumi.set(__self__, "function", function) + _setter("function", function) if materialized_view is not None: - pulumi.set(__self__, "materialized_view", materialized_view) + _setter("materialized_view", materialized_view) if metastore is not None: - pulumi.set(__self__, "metastore", metastore) + _setter("metastore", metastore) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if share is not None: - pulumi.set(__self__, "share", share) + _setter("share", share) if storage_credential is not None: - pulumi.set(__self__, "storage_credential", storage_credential) + _setter("storage_credential", storage_credential) if table is not None: - pulumi.set(__self__, "table", table) + _setter("table", table) if view is not None: - pulumi.set(__self__, "view", view) + _setter("view", view) if volume is not None: - pulumi.set(__self__, "volume", volume) + _setter("volume", volume) @property @pulumi.getter @@ -195,32 +228,65 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering Grants resources. """ + _GrantsState._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog=catalog, + external_location=external_location, + foreign_connection=foreign_connection, + function=function, + grants=grants, + materialized_view=materialized_view, + metastore=metastore, + schema=schema, + share=share, + storage_credential=storage_credential, + table=table, + view=view, + volume=volume, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog: Optional[pulumi.Input[str]] = None, + external_location: Optional[pulumi.Input[str]] = None, + foreign_connection: Optional[pulumi.Input[str]] = None, + function: Optional[pulumi.Input[str]] = None, + grants: Optional[pulumi.Input[Sequence[pulumi.Input['GrantsGrantArgs']]]] = None, + materialized_view: Optional[pulumi.Input[str]] = None, + metastore: Optional[pulumi.Input[str]] = None, + schema: Optional[pulumi.Input[str]] = None, + share: Optional[pulumi.Input[str]] = None, + storage_credential: Optional[pulumi.Input[str]] = None, + table: Optional[pulumi.Input[str]] = None, + view: Optional[pulumi.Input[str]] = None, + volume: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if external_location is not None: - pulumi.set(__self__, "external_location", external_location) + _setter("external_location", external_location) if foreign_connection is not None: - pulumi.set(__self__, "foreign_connection", foreign_connection) + _setter("foreign_connection", foreign_connection) if function is not None: - pulumi.set(__self__, "function", function) + _setter("function", function) if grants is not None: - pulumi.set(__self__, "grants", grants) + _setter("grants", grants) if materialized_view is not None: - pulumi.set(__self__, "materialized_view", materialized_view) + _setter("materialized_view", materialized_view) if metastore is not None: - pulumi.set(__self__, "metastore", metastore) + _setter("metastore", metastore) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if share is not None: - pulumi.set(__self__, "share", share) + _setter("share", share) if storage_credential is not None: - pulumi.set(__self__, "storage_credential", storage_credential) + _setter("storage_credential", storage_credential) if table is not None: - pulumi.set(__self__, "table", table) + _setter("table", table) if view is not None: - pulumi.set(__self__, "view", view) + _setter("view", view) if volume is not None: - pulumi.set(__self__, "volume", volume) + _setter("volume", volume) @property @pulumi.getter @@ -382,6 +448,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + GrantsArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/group.py b/sdk/python/pulumi_databricks/group.py index 69168fcd..c809e08d 100644 --- a/sdk/python/pulumi_databricks/group.py +++ b/sdk/python/pulumi_databricks/group.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['GroupArgs', 'Group'] @@ -33,24 +33,49 @@ def __init__(__self__, *, :param pulumi.Input[str] external_id: ID of the group in an external identity provider. :param pulumi.Input[bool] workspace_access: This is a field to allow the group to have access to Databricks Workspace. """ + GroupArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + acl_principal_id=acl_principal_id, + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + databricks_sql_access=databricks_sql_access, + display_name=display_name, + external_id=external_id, + force=force, + url=url, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + acl_principal_id: Optional[pulumi.Input[str]] = None, + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + display_name: Optional[pulumi.Input[str]] = None, + external_id: Optional[pulumi.Input[str]] = None, + force: Optional[pulumi.Input[bool]] = None, + url: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if acl_principal_id is not None: - pulumi.set(__self__, "acl_principal_id", acl_principal_id) + _setter("acl_principal_id", acl_principal_id) if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if display_name is not None: - pulumi.set(__self__, "display_name", display_name) + _setter("display_name", display_name) if external_id is not None: - pulumi.set(__self__, "external_id", external_id) + _setter("external_id", external_id) if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="aclPrincipalId") @@ -177,24 +202,49 @@ def __init__(__self__, *, :param pulumi.Input[str] external_id: ID of the group in an external identity provider. :param pulumi.Input[bool] workspace_access: This is a field to allow the group to have access to Databricks Workspace. """ + _GroupState._configure( + lambda key, value: pulumi.set(__self__, key, value), + acl_principal_id=acl_principal_id, + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + databricks_sql_access=databricks_sql_access, + display_name=display_name, + external_id=external_id, + force=force, + url=url, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + acl_principal_id: Optional[pulumi.Input[str]] = None, + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + display_name: Optional[pulumi.Input[str]] = None, + external_id: Optional[pulumi.Input[str]] = None, + force: Optional[pulumi.Input[bool]] = None, + url: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if acl_principal_id is not None: - pulumi.set(__self__, "acl_principal_id", acl_principal_id) + _setter("acl_principal_id", acl_principal_id) if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if display_name is not None: - pulumi.set(__self__, "display_name", display_name) + _setter("display_name", display_name) if external_id is not None: - pulumi.set(__self__, "external_id", external_id) + _setter("external_id", external_id) if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="aclPrincipalId") @@ -496,6 +546,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + GroupArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/group_instance_profile.py b/sdk/python/pulumi_databricks/group_instance_profile.py index 537dd838..0ee1c3de 100644 --- a/sdk/python/pulumi_databricks/group_instance_profile.py +++ b/sdk/python/pulumi_databricks/group_instance_profile.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['GroupInstanceProfileArgs', 'GroupInstanceProfile'] @@ -21,8 +21,19 @@ def __init__(__self__, *, :param pulumi.Input[str] group_id: This is the id of the group resource. :param pulumi.Input[str] instance_profile_id: This is the id of the instance profile resource. """ - pulumi.set(__self__, "group_id", group_id) - pulumi.set(__self__, "instance_profile_id", instance_profile_id) + GroupInstanceProfileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + group_id=group_id, + instance_profile_id=instance_profile_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + group_id: pulumi.Input[str], + instance_profile_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("group_id", group_id) + _setter("instance_profile_id", instance_profile_id) @property @pulumi.getter(name="groupId") @@ -59,10 +70,21 @@ def __init__(__self__, *, :param pulumi.Input[str] group_id: This is the id of the group resource. :param pulumi.Input[str] instance_profile_id: This is the id of the instance profile resource. """ + _GroupInstanceProfileState._configure( + lambda key, value: pulumi.set(__self__, key, value), + group_id=group_id, + instance_profile_id=instance_profile_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + group_id: Optional[pulumi.Input[str]] = None, + instance_profile_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if group_id is not None: - pulumi.set(__self__, "group_id", group_id) + _setter("group_id", group_id) if instance_profile_id is not None: - pulumi.set(__self__, "instance_profile_id", instance_profile_id) + _setter("instance_profile_id", instance_profile_id) @property @pulumi.getter(name="groupId") @@ -188,6 +210,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + GroupInstanceProfileArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/group_member.py b/sdk/python/pulumi_databricks/group_member.py index 3ffc8135..83eb4fec 100644 --- a/sdk/python/pulumi_databricks/group_member.py +++ b/sdk/python/pulumi_databricks/group_member.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['GroupMemberArgs', 'GroupMember'] @@ -21,8 +21,19 @@ def __init__(__self__, *, :param pulumi.Input[str] group_id: This is the id of the `group` resource. :param pulumi.Input[str] member_id: This is the id of the `group`, `service principal`, or `user`. """ - pulumi.set(__self__, "group_id", group_id) - pulumi.set(__self__, "member_id", member_id) + GroupMemberArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + group_id=group_id, + member_id=member_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + group_id: pulumi.Input[str], + member_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("group_id", group_id) + _setter("member_id", member_id) @property @pulumi.getter(name="groupId") @@ -59,10 +70,21 @@ def __init__(__self__, *, :param pulumi.Input[str] group_id: This is the id of the `group` resource. :param pulumi.Input[str] member_id: This is the id of the `group`, `service principal`, or `user`. """ + _GroupMemberState._configure( + lambda key, value: pulumi.set(__self__, key, value), + group_id=group_id, + member_id=member_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + group_id: Optional[pulumi.Input[str]] = None, + member_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if group_id is not None: - pulumi.set(__self__, "group_id", group_id) + _setter("group_id", group_id) if member_id is not None: - pulumi.set(__self__, "member_id", member_id) + _setter("member_id", member_id) @property @pulumi.getter(name="groupId") @@ -164,6 +186,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + GroupMemberArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/group_role.py b/sdk/python/pulumi_databricks/group_role.py index f4741861..f3b38df3 100644 --- a/sdk/python/pulumi_databricks/group_role.py +++ b/sdk/python/pulumi_databricks/group_role.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['GroupRoleArgs', 'GroupRole'] @@ -21,8 +21,19 @@ def __init__(__self__, *, :param pulumi.Input[str] group_id: This is the id of the group resource. :param pulumi.Input[str] role: Either a role name or the ARN/ID of the instance profile resource. """ - pulumi.set(__self__, "group_id", group_id) - pulumi.set(__self__, "role", role) + GroupRoleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + group_id=group_id, + role=role, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + group_id: pulumi.Input[str], + role: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("group_id", group_id) + _setter("role", role) @property @pulumi.getter(name="groupId") @@ -59,10 +70,21 @@ def __init__(__self__, *, :param pulumi.Input[str] group_id: This is the id of the group resource. :param pulumi.Input[str] role: Either a role name or the ARN/ID of the instance profile resource. """ + _GroupRoleState._configure( + lambda key, value: pulumi.set(__self__, key, value), + group_id=group_id, + role=role, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + group_id: Optional[pulumi.Input[str]] = None, + role: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if group_id is not None: - pulumi.set(__self__, "group_id", group_id) + _setter("group_id", group_id) if role is not None: - pulumi.set(__self__, "role", role) + _setter("role", role) @property @pulumi.getter(name="groupId") @@ -212,6 +234,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + GroupRoleArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/instance_pool.py b/sdk/python/pulumi_databricks/instance_pool.py index f4f5b32f..3d9e0092 100644 --- a/sdk/python/pulumi_databricks/instance_pool.py +++ b/sdk/python/pulumi_databricks/instance_pool.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -42,34 +42,71 @@ def __init__(__self__, *, :param pulumi.Input[str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool’s idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call. :param pulumi.Input[Sequence[pulumi.Input[str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call. """ - pulumi.set(__self__, "idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) - pulumi.set(__self__, "instance_pool_name", instance_pool_name) + InstancePoolArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + idle_instance_autotermination_minutes=idle_instance_autotermination_minutes, + instance_pool_name=instance_pool_name, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + custom_tags=custom_tags, + disk_spec=disk_spec, + enable_elastic_disk=enable_elastic_disk, + gcp_attributes=gcp_attributes, + instance_pool_fleet_attributes=instance_pool_fleet_attributes, + instance_pool_id=instance_pool_id, + max_capacity=max_capacity, + min_idle_instances=min_idle_instances, + node_type_id=node_type_id, + preloaded_docker_images=preloaded_docker_images, + preloaded_spark_versions=preloaded_spark_versions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + idle_instance_autotermination_minutes: pulumi.Input[int], + instance_pool_name: pulumi.Input[str], + aws_attributes: Optional[pulumi.Input['InstancePoolAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['InstancePoolAzureAttributesArgs']] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + disk_spec: Optional[pulumi.Input['InstancePoolDiskSpecArgs']] = None, + enable_elastic_disk: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['InstancePoolGcpAttributesArgs']] = None, + instance_pool_fleet_attributes: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + max_capacity: Optional[pulumi.Input[int]] = None, + min_idle_instances: Optional[pulumi.Input[int]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + preloaded_docker_images: Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]] = None, + preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) + _setter("instance_pool_name", instance_pool_name) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if disk_spec is not None: - pulumi.set(__self__, "disk_spec", disk_spec) + _setter("disk_spec", disk_spec) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if instance_pool_fleet_attributes is not None: - pulumi.set(__self__, "instance_pool_fleet_attributes", instance_pool_fleet_attributes) + _setter("instance_pool_fleet_attributes", instance_pool_fleet_attributes) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if max_capacity is not None: - pulumi.set(__self__, "max_capacity", max_capacity) + _setter("max_capacity", max_capacity) if min_idle_instances is not None: - pulumi.set(__self__, "min_idle_instances", min_idle_instances) + _setter("min_idle_instances", min_idle_instances) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if preloaded_docker_images is not None: - pulumi.set(__self__, "preloaded_docker_images", preloaded_docker_images) + _setter("preloaded_docker_images", preloaded_docker_images) if preloaded_spark_versions is not None: - pulumi.set(__self__, "preloaded_spark_versions", preloaded_spark_versions) + _setter("preloaded_spark_versions", preloaded_spark_versions) @property @pulumi.getter(name="idleInstanceAutoterminationMinutes") @@ -260,36 +297,73 @@ def __init__(__self__, *, :param pulumi.Input[str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool’s idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call. :param pulumi.Input[Sequence[pulumi.Input[str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call. """ + _InstancePoolState._configure( + lambda key, value: pulumi.set(__self__, key, value), + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + custom_tags=custom_tags, + disk_spec=disk_spec, + enable_elastic_disk=enable_elastic_disk, + gcp_attributes=gcp_attributes, + idle_instance_autotermination_minutes=idle_instance_autotermination_minutes, + instance_pool_fleet_attributes=instance_pool_fleet_attributes, + instance_pool_id=instance_pool_id, + instance_pool_name=instance_pool_name, + max_capacity=max_capacity, + min_idle_instances=min_idle_instances, + node_type_id=node_type_id, + preloaded_docker_images=preloaded_docker_images, + preloaded_spark_versions=preloaded_spark_versions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + aws_attributes: Optional[pulumi.Input['InstancePoolAwsAttributesArgs']] = None, + azure_attributes: Optional[pulumi.Input['InstancePoolAzureAttributesArgs']] = None, + custom_tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + disk_spec: Optional[pulumi.Input['InstancePoolDiskSpecArgs']] = None, + enable_elastic_disk: Optional[pulumi.Input[bool]] = None, + gcp_attributes: Optional[pulumi.Input['InstancePoolGcpAttributesArgs']] = None, + idle_instance_autotermination_minutes: Optional[pulumi.Input[int]] = None, + instance_pool_fleet_attributes: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + instance_pool_name: Optional[pulumi.Input[str]] = None, + max_capacity: Optional[pulumi.Input[int]] = None, + min_idle_instances: Optional[pulumi.Input[int]] = None, + node_type_id: Optional[pulumi.Input[str]] = None, + preloaded_docker_images: Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]] = None, + preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if disk_spec is not None: - pulumi.set(__self__, "disk_spec", disk_spec) + _setter("disk_spec", disk_spec) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idle_instance_autotermination_minutes is not None: - pulumi.set(__self__, "idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) + _setter("idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) if instance_pool_fleet_attributes is not None: - pulumi.set(__self__, "instance_pool_fleet_attributes", instance_pool_fleet_attributes) + _setter("instance_pool_fleet_attributes", instance_pool_fleet_attributes) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if instance_pool_name is not None: - pulumi.set(__self__, "instance_pool_name", instance_pool_name) + _setter("instance_pool_name", instance_pool_name) if max_capacity is not None: - pulumi.set(__self__, "max_capacity", max_capacity) + _setter("max_capacity", max_capacity) if min_idle_instances is not None: - pulumi.set(__self__, "min_idle_instances", min_idle_instances) + _setter("min_idle_instances", min_idle_instances) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if preloaded_docker_images is not None: - pulumi.set(__self__, "preloaded_docker_images", preloaded_docker_images) + _setter("preloaded_docker_images", preloaded_docker_images) if preloaded_spark_versions is not None: - pulumi.set(__self__, "preloaded_spark_versions", preloaded_spark_versions) + _setter("preloaded_spark_versions", preloaded_spark_versions) @property @pulumi.getter(name="awsAttributes") @@ -643,6 +717,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + InstancePoolArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -672,15 +750,40 @@ def _internal_init(__self__, raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = InstancePoolArgs.__new__(InstancePoolArgs) + if aws_attributes is not None and not isinstance(aws_attributes, InstancePoolAwsAttributesArgs): + aws_attributes = aws_attributes or {} + def _setter(key, value): + aws_attributes[key] = value + InstancePoolAwsAttributesArgs._configure(_setter, **aws_attributes) __props__.__dict__["aws_attributes"] = aws_attributes + if azure_attributes is not None and not isinstance(azure_attributes, InstancePoolAzureAttributesArgs): + azure_attributes = azure_attributes or {} + def _setter(key, value): + azure_attributes[key] = value + InstancePoolAzureAttributesArgs._configure(_setter, **azure_attributes) __props__.__dict__["azure_attributes"] = azure_attributes __props__.__dict__["custom_tags"] = custom_tags + if disk_spec is not None and not isinstance(disk_spec, InstancePoolDiskSpecArgs): + disk_spec = disk_spec or {} + def _setter(key, value): + disk_spec[key] = value + InstancePoolDiskSpecArgs._configure(_setter, **disk_spec) __props__.__dict__["disk_spec"] = disk_spec __props__.__dict__["enable_elastic_disk"] = enable_elastic_disk + if gcp_attributes is not None and not isinstance(gcp_attributes, InstancePoolGcpAttributesArgs): + gcp_attributes = gcp_attributes or {} + def _setter(key, value): + gcp_attributes[key] = value + InstancePoolGcpAttributesArgs._configure(_setter, **gcp_attributes) __props__.__dict__["gcp_attributes"] = gcp_attributes if idle_instance_autotermination_minutes is None and not opts.urn: raise TypeError("Missing required property 'idle_instance_autotermination_minutes'") __props__.__dict__["idle_instance_autotermination_minutes"] = idle_instance_autotermination_minutes + if instance_pool_fleet_attributes is not None and not isinstance(instance_pool_fleet_attributes, InstancePoolInstancePoolFleetAttributesArgs): + instance_pool_fleet_attributes = instance_pool_fleet_attributes or {} + def _setter(key, value): + instance_pool_fleet_attributes[key] = value + InstancePoolInstancePoolFleetAttributesArgs._configure(_setter, **instance_pool_fleet_attributes) __props__.__dict__["instance_pool_fleet_attributes"] = instance_pool_fleet_attributes __props__.__dict__["instance_pool_id"] = instance_pool_id if instance_pool_name is None and not opts.urn: diff --git a/sdk/python/pulumi_databricks/instance_profile.py b/sdk/python/pulumi_databricks/instance_profile.py index 201a6fed..0ac12ad2 100644 --- a/sdk/python/pulumi_databricks/instance_profile.py +++ b/sdk/python/pulumi_databricks/instance_profile.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['InstanceProfileArgs', 'InstanceProfile'] @@ -25,13 +25,28 @@ def __init__(__self__, *, :param pulumi.Input[bool] is_meta_instance_profile: Whether the instance profile is a meta instance profile. Used only in [IAM credential passthrough](https://docs.databricks.com/security/credential-passthrough/iam-passthrough.html). :param pulumi.Input[bool] skip_validation: **For advanced usage only.** If validation fails with an error message that does not indicate an IAM related permission issue, (e.g. "Your requested instance type is not supported in your requested availability zone"), you can pass this flag to skip the validation and forcibly add the instance profile. """ - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + InstanceProfileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + instance_profile_arn=instance_profile_arn, + iam_role_arn=iam_role_arn, + is_meta_instance_profile=is_meta_instance_profile, + skip_validation=skip_validation, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + instance_profile_arn: pulumi.Input[str], + iam_role_arn: Optional[pulumi.Input[str]] = None, + is_meta_instance_profile: Optional[pulumi.Input[bool]] = None, + skip_validation: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("instance_profile_arn", instance_profile_arn) if iam_role_arn is not None: - pulumi.set(__self__, "iam_role_arn", iam_role_arn) + _setter("iam_role_arn", iam_role_arn) if is_meta_instance_profile is not None: - pulumi.set(__self__, "is_meta_instance_profile", is_meta_instance_profile) + _setter("is_meta_instance_profile", is_meta_instance_profile) if skip_validation is not None: - pulumi.set(__self__, "skip_validation", skip_validation) + _setter("skip_validation", skip_validation) @property @pulumi.getter(name="instanceProfileArn") @@ -96,14 +111,29 @@ def __init__(__self__, *, :param pulumi.Input[bool] is_meta_instance_profile: Whether the instance profile is a meta instance profile. Used only in [IAM credential passthrough](https://docs.databricks.com/security/credential-passthrough/iam-passthrough.html). :param pulumi.Input[bool] skip_validation: **For advanced usage only.** If validation fails with an error message that does not indicate an IAM related permission issue, (e.g. "Your requested instance type is not supported in your requested availability zone"), you can pass this flag to skip the validation and forcibly add the instance profile. """ + _InstanceProfileState._configure( + lambda key, value: pulumi.set(__self__, key, value), + iam_role_arn=iam_role_arn, + instance_profile_arn=instance_profile_arn, + is_meta_instance_profile=is_meta_instance_profile, + skip_validation=skip_validation, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + iam_role_arn: Optional[pulumi.Input[str]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + is_meta_instance_profile: Optional[pulumi.Input[bool]] = None, + skip_validation: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if iam_role_arn is not None: - pulumi.set(__self__, "iam_role_arn", iam_role_arn) + _setter("iam_role_arn", iam_role_arn) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if is_meta_instance_profile is not None: - pulumi.set(__self__, "is_meta_instance_profile", is_meta_instance_profile) + _setter("is_meta_instance_profile", is_meta_instance_profile) if skip_validation is not None: - pulumi.set(__self__, "skip_validation", skip_validation) + _setter("skip_validation", skip_validation) @property @pulumi.getter(name="iamRoleArn") @@ -441,6 +471,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + InstanceProfileArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/ip_access_list.py b/sdk/python/pulumi_databricks/ip_access_list.py index b740e6fe..602dbb10 100644 --- a/sdk/python/pulumi_databricks/ip_access_list.py +++ b/sdk/python/pulumi_databricks/ip_access_list.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['IpAccessListArgs', 'IpAccessList'] @@ -25,11 +25,26 @@ def __init__(__self__, *, :param pulumi.Input[str] list_type: Can only be "ALLOW" or "BLOCK". :param pulumi.Input[bool] enabled: Boolean `true` or `false` indicating whether this list should be active. Defaults to `true` """ - pulumi.set(__self__, "ip_addresses", ip_addresses) - pulumi.set(__self__, "label", label) - pulumi.set(__self__, "list_type", list_type) + IpAccessListArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + ip_addresses=ip_addresses, + label=label, + list_type=list_type, + enabled=enabled, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + ip_addresses: pulumi.Input[Sequence[pulumi.Input[str]]], + label: pulumi.Input[str], + list_type: pulumi.Input[str], + enabled: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("ip_addresses", ip_addresses) + _setter("label", label) + _setter("list_type", list_type) if enabled is not None: - pulumi.set(__self__, "enabled", enabled) + _setter("enabled", enabled) @property @pulumi.getter(name="ipAddresses") @@ -94,14 +109,29 @@ def __init__(__self__, *, :param pulumi.Input[str] label: This is the display name for the given IP ACL List. :param pulumi.Input[str] list_type: Can only be "ALLOW" or "BLOCK". """ + _IpAccessListState._configure( + lambda key, value: pulumi.set(__self__, key, value), + enabled=enabled, + ip_addresses=ip_addresses, + label=label, + list_type=list_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + enabled: Optional[pulumi.Input[bool]] = None, + ip_addresses: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + label: Optional[pulumi.Input[str]] = None, + list_type: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if enabled is not None: - pulumi.set(__self__, "enabled", enabled) + _setter("enabled", enabled) if ip_addresses is not None: - pulumi.set(__self__, "ip_addresses", ip_addresses) + _setter("ip_addresses", ip_addresses) if label is not None: - pulumi.set(__self__, "label", label) + _setter("label", label) if list_type is not None: - pulumi.set(__self__, "list_type", list_type) + _setter("list_type", list_type) @property @pulumi.getter @@ -271,6 +301,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + IpAccessListArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/job.py b/sdk/python/pulumi_databricks/job.py index 93869782..a49c9ed7 100644 --- a/sdk/python/pulumi_databricks/job.py +++ b/sdk/python/pulumi_databricks/job.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -75,79 +75,156 @@ def __init__(__self__, *, :param pulumi.Input[int] timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. :param pulumi.Input['JobWebhookNotificationsArgs'] webhook_notifications: (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this job begins, completes and fails. The default behavior is to not send any notifications. This field is a block and is documented below. """ + JobArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + always_running=always_running, + computes=computes, + continuous=continuous, + control_run_state=control_run_state, + dbt_task=dbt_task, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + format=format, + git_source=git_source, + health=health, + job_clusters=job_clusters, + libraries=libraries, + max_concurrent_runs=max_concurrent_runs, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + name=name, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + parameters=parameters, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + queue=queue, + retry_on_timeout=retry_on_timeout, + run_as=run_as, + run_job_task=run_job_task, + schedule=schedule, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + tags=tags, + tasks=tasks, + timeout_seconds=timeout_seconds, + trigger=trigger, + webhook_notifications=webhook_notifications, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + always_running: Optional[pulumi.Input[bool]] = None, + computes: Optional[pulumi.Input[Sequence[pulumi.Input['JobComputeArgs']]]] = None, + continuous: Optional[pulumi.Input['JobContinuousArgs']] = None, + control_run_state: Optional[pulumi.Input[bool]] = None, + dbt_task: Optional[pulumi.Input['JobDbtTaskArgs']] = None, + email_notifications: Optional[pulumi.Input['JobEmailNotificationsArgs']] = None, + existing_cluster_id: Optional[pulumi.Input[str]] = None, + format: Optional[pulumi.Input[str]] = None, + git_source: Optional[pulumi.Input['JobGitSourceArgs']] = None, + health: Optional[pulumi.Input['JobHealthArgs']] = None, + job_clusters: Optional[pulumi.Input[Sequence[pulumi.Input['JobJobClusterArgs']]]] = None, + libraries: Optional[pulumi.Input[Sequence[pulumi.Input['JobLibraryArgs']]]] = None, + max_concurrent_runs: Optional[pulumi.Input[int]] = None, + max_retries: Optional[pulumi.Input[int]] = None, + min_retry_interval_millis: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + new_cluster: Optional[pulumi.Input['JobNewClusterArgs']] = None, + notebook_task: Optional[pulumi.Input['JobNotebookTaskArgs']] = None, + notification_settings: Optional[pulumi.Input['JobNotificationSettingsArgs']] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input['JobParameterArgs']]]] = None, + pipeline_task: Optional[pulumi.Input['JobPipelineTaskArgs']] = None, + python_wheel_task: Optional[pulumi.Input['JobPythonWheelTaskArgs']] = None, + queue: Optional[pulumi.Input['JobQueueArgs']] = None, + retry_on_timeout: Optional[pulumi.Input[bool]] = None, + run_as: Optional[pulumi.Input['JobRunAsArgs']] = None, + run_job_task: Optional[pulumi.Input['JobRunJobTaskArgs']] = None, + schedule: Optional[pulumi.Input['JobScheduleArgs']] = None, + spark_jar_task: Optional[pulumi.Input['JobSparkJarTaskArgs']] = None, + spark_python_task: Optional[pulumi.Input['JobSparkPythonTaskArgs']] = None, + spark_submit_task: Optional[pulumi.Input['JobSparkSubmitTaskArgs']] = None, + tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + tasks: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskArgs']]]] = None, + timeout_seconds: Optional[pulumi.Input[int]] = None, + trigger: Optional[pulumi.Input['JobTriggerArgs']] = None, + webhook_notifications: Optional[pulumi.Input['JobWebhookNotificationsArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if always_running is not None: warnings.warn("""always_running will be replaced by control_run_state in the next major release.""", DeprecationWarning) pulumi.log.warn("""always_running is deprecated: always_running will be replaced by control_run_state in the next major release.""") if always_running is not None: - pulumi.set(__self__, "always_running", always_running) + _setter("always_running", always_running) if computes is not None: - pulumi.set(__self__, "computes", computes) + _setter("computes", computes) if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if control_run_state is not None: - pulumi.set(__self__, "control_run_state", control_run_state) + _setter("control_run_state", control_run_state) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if format is not None: - pulumi.set(__self__, "format", format) + _setter("format", format) if git_source is not None: - pulumi.set(__self__, "git_source", git_source) + _setter("git_source", git_source) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_clusters is not None: - pulumi.set(__self__, "job_clusters", job_clusters) + _setter("job_clusters", job_clusters) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_concurrent_runs is not None: - pulumi.set(__self__, "max_concurrent_runs", max_concurrent_runs) + _setter("max_concurrent_runs", max_concurrent_runs) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if queue is not None: - pulumi.set(__self__, "queue", queue) + _setter("queue", queue) if retry_on_timeout is not None: - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + _setter("retry_on_timeout", retry_on_timeout) if run_as is not None: - pulumi.set(__self__, "run_as", run_as) + _setter("run_as", run_as) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if schedule is not None: - pulumi.set(__self__, "schedule", schedule) + _setter("schedule", schedule) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if tasks is not None: - pulumi.set(__self__, "tasks", tasks) + _setter("tasks", tasks) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) if trigger is not None: - pulumi.set(__self__, "trigger", trigger) + _setter("trigger", trigger) if webhook_notifications is not None: - pulumi.set(__self__, "webhook_notifications", webhook_notifications) + _setter("webhook_notifications", webhook_notifications) @property @pulumi.getter(name="alwaysRunning") @@ -582,84 +659,163 @@ def __init__(__self__, *, :param pulumi.Input[bool] retry_on_timeout: (Bool) An optional policy to specify whether to retry a job when it times out. The default behavior is to not retry on timeout. :param pulumi.Input['JobScheduleArgs'] schedule: (List) An optional periodic schedule for this job. The default behavior is that the job runs when triggered by clicking Run Now in the Jobs UI or sending an API request to runNow. This field is a block and is documented below. :param pulumi.Input[int] timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. - :param pulumi.Input[str] url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + :param pulumi.Input[str] url: URL of the job on the given workspace :param pulumi.Input['JobWebhookNotificationsArgs'] webhook_notifications: (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this job begins, completes and fails. The default behavior is to not send any notifications. This field is a block and is documented below. """ + _JobState._configure( + lambda key, value: pulumi.set(__self__, key, value), + always_running=always_running, + computes=computes, + continuous=continuous, + control_run_state=control_run_state, + dbt_task=dbt_task, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + format=format, + git_source=git_source, + health=health, + job_clusters=job_clusters, + libraries=libraries, + max_concurrent_runs=max_concurrent_runs, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + name=name, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + parameters=parameters, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + queue=queue, + retry_on_timeout=retry_on_timeout, + run_as=run_as, + run_job_task=run_job_task, + schedule=schedule, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + tags=tags, + tasks=tasks, + timeout_seconds=timeout_seconds, + trigger=trigger, + url=url, + webhook_notifications=webhook_notifications, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + always_running: Optional[pulumi.Input[bool]] = None, + computes: Optional[pulumi.Input[Sequence[pulumi.Input['JobComputeArgs']]]] = None, + continuous: Optional[pulumi.Input['JobContinuousArgs']] = None, + control_run_state: Optional[pulumi.Input[bool]] = None, + dbt_task: Optional[pulumi.Input['JobDbtTaskArgs']] = None, + email_notifications: Optional[pulumi.Input['JobEmailNotificationsArgs']] = None, + existing_cluster_id: Optional[pulumi.Input[str]] = None, + format: Optional[pulumi.Input[str]] = None, + git_source: Optional[pulumi.Input['JobGitSourceArgs']] = None, + health: Optional[pulumi.Input['JobHealthArgs']] = None, + job_clusters: Optional[pulumi.Input[Sequence[pulumi.Input['JobJobClusterArgs']]]] = None, + libraries: Optional[pulumi.Input[Sequence[pulumi.Input['JobLibraryArgs']]]] = None, + max_concurrent_runs: Optional[pulumi.Input[int]] = None, + max_retries: Optional[pulumi.Input[int]] = None, + min_retry_interval_millis: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + new_cluster: Optional[pulumi.Input['JobNewClusterArgs']] = None, + notebook_task: Optional[pulumi.Input['JobNotebookTaskArgs']] = None, + notification_settings: Optional[pulumi.Input['JobNotificationSettingsArgs']] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input['JobParameterArgs']]]] = None, + pipeline_task: Optional[pulumi.Input['JobPipelineTaskArgs']] = None, + python_wheel_task: Optional[pulumi.Input['JobPythonWheelTaskArgs']] = None, + queue: Optional[pulumi.Input['JobQueueArgs']] = None, + retry_on_timeout: Optional[pulumi.Input[bool]] = None, + run_as: Optional[pulumi.Input['JobRunAsArgs']] = None, + run_job_task: Optional[pulumi.Input['JobRunJobTaskArgs']] = None, + schedule: Optional[pulumi.Input['JobScheduleArgs']] = None, + spark_jar_task: Optional[pulumi.Input['JobSparkJarTaskArgs']] = None, + spark_python_task: Optional[pulumi.Input['JobSparkPythonTaskArgs']] = None, + spark_submit_task: Optional[pulumi.Input['JobSparkSubmitTaskArgs']] = None, + tags: Optional[pulumi.Input[Mapping[str, Any]]] = None, + tasks: Optional[pulumi.Input[Sequence[pulumi.Input['JobTaskArgs']]]] = None, + timeout_seconds: Optional[pulumi.Input[int]] = None, + trigger: Optional[pulumi.Input['JobTriggerArgs']] = None, + url: Optional[pulumi.Input[str]] = None, + webhook_notifications: Optional[pulumi.Input['JobWebhookNotificationsArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if always_running is not None: warnings.warn("""always_running will be replaced by control_run_state in the next major release.""", DeprecationWarning) pulumi.log.warn("""always_running is deprecated: always_running will be replaced by control_run_state in the next major release.""") if always_running is not None: - pulumi.set(__self__, "always_running", always_running) + _setter("always_running", always_running) if computes is not None: - pulumi.set(__self__, "computes", computes) + _setter("computes", computes) if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if control_run_state is not None: - pulumi.set(__self__, "control_run_state", control_run_state) + _setter("control_run_state", control_run_state) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if format is not None: - pulumi.set(__self__, "format", format) + _setter("format", format) if git_source is not None: - pulumi.set(__self__, "git_source", git_source) + _setter("git_source", git_source) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_clusters is not None: - pulumi.set(__self__, "job_clusters", job_clusters) + _setter("job_clusters", job_clusters) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_concurrent_runs is not None: - pulumi.set(__self__, "max_concurrent_runs", max_concurrent_runs) + _setter("max_concurrent_runs", max_concurrent_runs) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if queue is not None: - pulumi.set(__self__, "queue", queue) + _setter("queue", queue) if retry_on_timeout is not None: - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + _setter("retry_on_timeout", retry_on_timeout) if run_as is not None: - pulumi.set(__self__, "run_as", run_as) + _setter("run_as", run_as) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if schedule is not None: - pulumi.set(__self__, "schedule", schedule) + _setter("schedule", schedule) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if tasks is not None: - pulumi.set(__self__, "tasks", tasks) + _setter("tasks", tasks) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) if trigger is not None: - pulumi.set(__self__, "trigger", trigger) + _setter("trigger", trigger) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) if webhook_notifications is not None: - pulumi.set(__self__, "webhook_notifications", webhook_notifications) + _setter("webhook_notifications", webhook_notifications) @property @pulumi.getter(name="alwaysRunning") @@ -1024,7 +1180,7 @@ def trigger(self, value: Optional[pulumi.Input['JobTriggerArgs']]): @pulumi.getter def url(self) -> Optional[pulumi.Input[str]]: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -1144,6 +1300,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + JobArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -1193,18 +1353,40 @@ def _internal_init(__self__, raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = JobArgs.__new__(JobArgs) - if always_running is not None and not opts.urn: - warnings.warn("""always_running will be replaced by control_run_state in the next major release.""", DeprecationWarning) - pulumi.log.warn("""always_running is deprecated: always_running will be replaced by control_run_state in the next major release.""") __props__.__dict__["always_running"] = always_running __props__.__dict__["computes"] = computes + if continuous is not None and not isinstance(continuous, JobContinuousArgs): + continuous = continuous or {} + def _setter(key, value): + continuous[key] = value + JobContinuousArgs._configure(_setter, **continuous) __props__.__dict__["continuous"] = continuous __props__.__dict__["control_run_state"] = control_run_state + if dbt_task is not None and not isinstance(dbt_task, JobDbtTaskArgs): + dbt_task = dbt_task or {} + def _setter(key, value): + dbt_task[key] = value + JobDbtTaskArgs._configure(_setter, **dbt_task) __props__.__dict__["dbt_task"] = dbt_task + if email_notifications is not None and not isinstance(email_notifications, JobEmailNotificationsArgs): + email_notifications = email_notifications or {} + def _setter(key, value): + email_notifications[key] = value + JobEmailNotificationsArgs._configure(_setter, **email_notifications) __props__.__dict__["email_notifications"] = email_notifications __props__.__dict__["existing_cluster_id"] = existing_cluster_id __props__.__dict__["format"] = format + if git_source is not None and not isinstance(git_source, JobGitSourceArgs): + git_source = git_source or {} + def _setter(key, value): + git_source[key] = value + JobGitSourceArgs._configure(_setter, **git_source) __props__.__dict__["git_source"] = git_source + if health is not None and not isinstance(health, JobHealthArgs): + health = health or {} + def _setter(key, value): + health[key] = value + JobHealthArgs._configure(_setter, **health) __props__.__dict__["health"] = health __props__.__dict__["job_clusters"] = job_clusters __props__.__dict__["libraries"] = libraries @@ -1212,24 +1394,94 @@ def _internal_init(__self__, __props__.__dict__["max_retries"] = max_retries __props__.__dict__["min_retry_interval_millis"] = min_retry_interval_millis __props__.__dict__["name"] = name + if new_cluster is not None and not isinstance(new_cluster, JobNewClusterArgs): + new_cluster = new_cluster or {} + def _setter(key, value): + new_cluster[key] = value + JobNewClusterArgs._configure(_setter, **new_cluster) __props__.__dict__["new_cluster"] = new_cluster + if notebook_task is not None and not isinstance(notebook_task, JobNotebookTaskArgs): + notebook_task = notebook_task or {} + def _setter(key, value): + notebook_task[key] = value + JobNotebookTaskArgs._configure(_setter, **notebook_task) __props__.__dict__["notebook_task"] = notebook_task + if notification_settings is not None and not isinstance(notification_settings, JobNotificationSettingsArgs): + notification_settings = notification_settings or {} + def _setter(key, value): + notification_settings[key] = value + JobNotificationSettingsArgs._configure(_setter, **notification_settings) __props__.__dict__["notification_settings"] = notification_settings __props__.__dict__["parameters"] = parameters + if pipeline_task is not None and not isinstance(pipeline_task, JobPipelineTaskArgs): + pipeline_task = pipeline_task or {} + def _setter(key, value): + pipeline_task[key] = value + JobPipelineTaskArgs._configure(_setter, **pipeline_task) __props__.__dict__["pipeline_task"] = pipeline_task + if python_wheel_task is not None and not isinstance(python_wheel_task, JobPythonWheelTaskArgs): + python_wheel_task = python_wheel_task or {} + def _setter(key, value): + python_wheel_task[key] = value + JobPythonWheelTaskArgs._configure(_setter, **python_wheel_task) __props__.__dict__["python_wheel_task"] = python_wheel_task + if queue is not None and not isinstance(queue, JobQueueArgs): + queue = queue or {} + def _setter(key, value): + queue[key] = value + JobQueueArgs._configure(_setter, **queue) __props__.__dict__["queue"] = queue __props__.__dict__["retry_on_timeout"] = retry_on_timeout + if run_as is not None and not isinstance(run_as, JobRunAsArgs): + run_as = run_as or {} + def _setter(key, value): + run_as[key] = value + JobRunAsArgs._configure(_setter, **run_as) __props__.__dict__["run_as"] = run_as + if run_job_task is not None and not isinstance(run_job_task, JobRunJobTaskArgs): + run_job_task = run_job_task or {} + def _setter(key, value): + run_job_task[key] = value + JobRunJobTaskArgs._configure(_setter, **run_job_task) __props__.__dict__["run_job_task"] = run_job_task + if schedule is not None and not isinstance(schedule, JobScheduleArgs): + schedule = schedule or {} + def _setter(key, value): + schedule[key] = value + JobScheduleArgs._configure(_setter, **schedule) __props__.__dict__["schedule"] = schedule + if spark_jar_task is not None and not isinstance(spark_jar_task, JobSparkJarTaskArgs): + spark_jar_task = spark_jar_task or {} + def _setter(key, value): + spark_jar_task[key] = value + JobSparkJarTaskArgs._configure(_setter, **spark_jar_task) __props__.__dict__["spark_jar_task"] = spark_jar_task + if spark_python_task is not None and not isinstance(spark_python_task, JobSparkPythonTaskArgs): + spark_python_task = spark_python_task or {} + def _setter(key, value): + spark_python_task[key] = value + JobSparkPythonTaskArgs._configure(_setter, **spark_python_task) __props__.__dict__["spark_python_task"] = spark_python_task + if spark_submit_task is not None and not isinstance(spark_submit_task, JobSparkSubmitTaskArgs): + spark_submit_task = spark_submit_task or {} + def _setter(key, value): + spark_submit_task[key] = value + JobSparkSubmitTaskArgs._configure(_setter, **spark_submit_task) __props__.__dict__["spark_submit_task"] = spark_submit_task __props__.__dict__["tags"] = tags __props__.__dict__["tasks"] = tasks __props__.__dict__["timeout_seconds"] = timeout_seconds + if trigger is not None and not isinstance(trigger, JobTriggerArgs): + trigger = trigger or {} + def _setter(key, value): + trigger[key] = value + JobTriggerArgs._configure(_setter, **trigger) __props__.__dict__["trigger"] = trigger + if webhook_notifications is not None and not isinstance(webhook_notifications, JobWebhookNotificationsArgs): + webhook_notifications = webhook_notifications or {} + def _setter(key, value): + webhook_notifications[key] = value + JobWebhookNotificationsArgs._configure(_setter, **webhook_notifications) __props__.__dict__["webhook_notifications"] = webhook_notifications __props__.__dict__["url"] = None super(Job, __self__).__init__( @@ -1305,7 +1557,7 @@ def get(resource_name: str, :param pulumi.Input[bool] retry_on_timeout: (Bool) An optional policy to specify whether to retry a job when it times out. The default behavior is to not retry on timeout. :param pulumi.Input[pulumi.InputType['JobScheduleArgs']] schedule: (List) An optional periodic schedule for this job. The default behavior is that the job runs when triggered by clicking Run Now in the Jobs UI or sending an API request to runNow. This field is a block and is documented below. :param pulumi.Input[int] timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. - :param pulumi.Input[str] url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + :param pulumi.Input[str] url: URL of the job on the given workspace :param pulumi.Input[pulumi.InputType['JobWebhookNotificationsArgs']] webhook_notifications: (List) An optional set of system destinations (for example, webhook destinations or Slack) to be notified when runs of this job begins, completes and fails. The default behavior is to not send any notifications. This field is a block and is documented below. """ opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id)) @@ -1577,7 +1829,7 @@ def trigger(self) -> pulumi.Output[Optional['outputs.JobTrigger']]: @pulumi.getter def url(self) -> pulumi.Output[str]: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") diff --git a/sdk/python/pulumi_databricks/library.py b/sdk/python/pulumi_databricks/library.py index 47581aac..5a3b0f72 100644 --- a/sdk/python/pulumi_databricks/library.py +++ b/sdk/python/pulumi_databricks/library.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -26,19 +26,40 @@ def __init__(__self__, *, """ The set of arguments for constructing a Library resource. """ - pulumi.set(__self__, "cluster_id", cluster_id) + LibraryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cluster_id=cluster_id, + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cluster_id: pulumi.Input[str], + cran: Optional[pulumi.Input['LibraryCranArgs']] = None, + egg: Optional[pulumi.Input[str]] = None, + jar: Optional[pulumi.Input[str]] = None, + maven: Optional[pulumi.Input['LibraryMavenArgs']] = None, + pypi: Optional[pulumi.Input['LibraryPypiArgs']] = None, + whl: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("cluster_id", cluster_id) if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter(name="clusterId") @@ -117,20 +138,41 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering Library resources. """ + _LibraryState._configure( + lambda key, value: pulumi.set(__self__, key, value), + cluster_id=cluster_id, + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cluster_id: Optional[pulumi.Input[str]] = None, + cran: Optional[pulumi.Input['LibraryCranArgs']] = None, + egg: Optional[pulumi.Input[str]] = None, + jar: Optional[pulumi.Input[str]] = None, + maven: Optional[pulumi.Input['LibraryMavenArgs']] = None, + pypi: Optional[pulumi.Input['LibraryPypiArgs']] = None, + whl: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter(name="clusterId") @@ -496,6 +538,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + LibraryArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -520,10 +566,25 @@ def _internal_init(__self__, if cluster_id is None and not opts.urn: raise TypeError("Missing required property 'cluster_id'") __props__.__dict__["cluster_id"] = cluster_id + if cran is not None and not isinstance(cran, LibraryCranArgs): + cran = cran or {} + def _setter(key, value): + cran[key] = value + LibraryCranArgs._configure(_setter, **cran) __props__.__dict__["cran"] = cran __props__.__dict__["egg"] = egg __props__.__dict__["jar"] = jar + if maven is not None and not isinstance(maven, LibraryMavenArgs): + maven = maven or {} + def _setter(key, value): + maven[key] = value + LibraryMavenArgs._configure(_setter, **maven) __props__.__dict__["maven"] = maven + if pypi is not None and not isinstance(pypi, LibraryPypiArgs): + pypi = pypi or {} + def _setter(key, value): + pypi[key] = value + LibraryPypiArgs._configure(_setter, **pypi) __props__.__dict__["pypi"] = pypi __props__.__dict__["whl"] = whl super(Library, __self__).__init__( diff --git a/sdk/python/pulumi_databricks/metastore.py b/sdk/python/pulumi_databricks/metastore.py index fb021f63..6831234b 100644 --- a/sdk/python/pulumi_databricks/metastore.py +++ b/sdk/python/pulumi_databricks/metastore.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MetastoreArgs', 'Metastore'] @@ -42,39 +42,80 @@ def __init__(__self__, *, :param pulumi.Input[str] owner: Username/groupname/sp application_id of the metastore owner. :param pulumi.Input[str] region: The region of the metastore """ - pulumi.set(__self__, "storage_root", storage_root) + MetastoreArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + storage_root=storage_root, + cloud=cloud, + created_at=created_at, + created_by=created_by, + default_data_access_config_id=default_data_access_config_id, + delta_sharing_organization_name=delta_sharing_organization_name, + delta_sharing_recipient_token_lifetime_in_seconds=delta_sharing_recipient_token_lifetime_in_seconds, + delta_sharing_scope=delta_sharing_scope, + force_destroy=force_destroy, + global_metastore_id=global_metastore_id, + metastore_id=metastore_id, + name=name, + owner=owner, + region=region, + storage_root_credential_id=storage_root_credential_id, + updated_at=updated_at, + updated_by=updated_by, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + storage_root: pulumi.Input[str], + cloud: Optional[pulumi.Input[str]] = None, + created_at: Optional[pulumi.Input[int]] = None, + created_by: Optional[pulumi.Input[str]] = None, + default_data_access_config_id: Optional[pulumi.Input[str]] = None, + delta_sharing_organization_name: Optional[pulumi.Input[str]] = None, + delta_sharing_recipient_token_lifetime_in_seconds: Optional[pulumi.Input[int]] = None, + delta_sharing_scope: Optional[pulumi.Input[str]] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + global_metastore_id: Optional[pulumi.Input[str]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + storage_root_credential_id: Optional[pulumi.Input[str]] = None, + updated_at: Optional[pulumi.Input[int]] = None, + updated_by: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("storage_root", storage_root) if cloud is not None: - pulumi.set(__self__, "cloud", cloud) + _setter("cloud", cloud) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if default_data_access_config_id is not None: - pulumi.set(__self__, "default_data_access_config_id", default_data_access_config_id) + _setter("default_data_access_config_id", default_data_access_config_id) if delta_sharing_organization_name is not None: - pulumi.set(__self__, "delta_sharing_organization_name", delta_sharing_organization_name) + _setter("delta_sharing_organization_name", delta_sharing_organization_name) if delta_sharing_recipient_token_lifetime_in_seconds is not None: - pulumi.set(__self__, "delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) + _setter("delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) if delta_sharing_scope is not None: - pulumi.set(__self__, "delta_sharing_scope", delta_sharing_scope) + _setter("delta_sharing_scope", delta_sharing_scope) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if global_metastore_id is not None: - pulumi.set(__self__, "global_metastore_id", global_metastore_id) + _setter("global_metastore_id", global_metastore_id) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) if storage_root_credential_id is not None: - pulumi.set(__self__, "storage_root_credential_id", storage_root_credential_id) + _setter("storage_root_credential_id", storage_root_credential_id) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) if updated_by is not None: - pulumi.set(__self__, "updated_by", updated_by) + _setter("updated_by", updated_by) @property @pulumi.getter(name="storageRoot") @@ -285,40 +326,81 @@ def __init__(__self__, *, :param pulumi.Input[str] region: The region of the metastore :param pulumi.Input[str] storage_root: Path on cloud storage account, where managed `Table` are stored. Change forces creation of a new resource. """ + _MetastoreState._configure( + lambda key, value: pulumi.set(__self__, key, value), + cloud=cloud, + created_at=created_at, + created_by=created_by, + default_data_access_config_id=default_data_access_config_id, + delta_sharing_organization_name=delta_sharing_organization_name, + delta_sharing_recipient_token_lifetime_in_seconds=delta_sharing_recipient_token_lifetime_in_seconds, + delta_sharing_scope=delta_sharing_scope, + force_destroy=force_destroy, + global_metastore_id=global_metastore_id, + metastore_id=metastore_id, + name=name, + owner=owner, + region=region, + storage_root=storage_root, + storage_root_credential_id=storage_root_credential_id, + updated_at=updated_at, + updated_by=updated_by, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cloud: Optional[pulumi.Input[str]] = None, + created_at: Optional[pulumi.Input[int]] = None, + created_by: Optional[pulumi.Input[str]] = None, + default_data_access_config_id: Optional[pulumi.Input[str]] = None, + delta_sharing_organization_name: Optional[pulumi.Input[str]] = None, + delta_sharing_recipient_token_lifetime_in_seconds: Optional[pulumi.Input[int]] = None, + delta_sharing_scope: Optional[pulumi.Input[str]] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + global_metastore_id: Optional[pulumi.Input[str]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None, + storage_root: Optional[pulumi.Input[str]] = None, + storage_root_credential_id: Optional[pulumi.Input[str]] = None, + updated_at: Optional[pulumi.Input[int]] = None, + updated_by: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cloud is not None: - pulumi.set(__self__, "cloud", cloud) + _setter("cloud", cloud) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if default_data_access_config_id is not None: - pulumi.set(__self__, "default_data_access_config_id", default_data_access_config_id) + _setter("default_data_access_config_id", default_data_access_config_id) if delta_sharing_organization_name is not None: - pulumi.set(__self__, "delta_sharing_organization_name", delta_sharing_organization_name) + _setter("delta_sharing_organization_name", delta_sharing_organization_name) if delta_sharing_recipient_token_lifetime_in_seconds is not None: - pulumi.set(__self__, "delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) + _setter("delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) if delta_sharing_scope is not None: - pulumi.set(__self__, "delta_sharing_scope", delta_sharing_scope) + _setter("delta_sharing_scope", delta_sharing_scope) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if global_metastore_id is not None: - pulumi.set(__self__, "global_metastore_id", global_metastore_id) + _setter("global_metastore_id", global_metastore_id) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) if storage_root is not None: - pulumi.set(__self__, "storage_root", storage_root) + _setter("storage_root", storage_root) if storage_root_credential_id is not None: - pulumi.set(__self__, "storage_root_credential_id", storage_root_credential_id) + _setter("storage_root_credential_id", storage_root_credential_id) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) if updated_by is not None: - pulumi.set(__self__, "updated_by", updated_by) + _setter("updated_by", updated_by) @property @pulumi.getter @@ -574,6 +656,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MetastoreArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/metastore_assignment.py b/sdk/python/pulumi_databricks/metastore_assignment.py index 213c6c85..602b2188 100644 --- a/sdk/python/pulumi_databricks/metastore_assignment.py +++ b/sdk/python/pulumi_databricks/metastore_assignment.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MetastoreAssignmentArgs', 'MetastoreAssignment'] @@ -23,10 +23,23 @@ def __init__(__self__, *, :param pulumi.Input[int] workspace_id: id of the workspace for the assignment :param pulumi.Input[str] default_catalog_name: Default catalog used for this assignment, default to `hive_metastore` """ - pulumi.set(__self__, "metastore_id", metastore_id) - pulumi.set(__self__, "workspace_id", workspace_id) + MetastoreAssignmentArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + metastore_id=metastore_id, + workspace_id=workspace_id, + default_catalog_name=default_catalog_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metastore_id: pulumi.Input[str], + workspace_id: pulumi.Input[int], + default_catalog_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("metastore_id", metastore_id) + _setter("workspace_id", workspace_id) if default_catalog_name is not None: - pulumi.set(__self__, "default_catalog_name", default_catalog_name) + _setter("default_catalog_name", default_catalog_name) @property @pulumi.getter(name="metastoreId") @@ -77,12 +90,25 @@ def __init__(__self__, *, :param pulumi.Input[str] metastore_id: Unique identifier of the parent Metastore :param pulumi.Input[int] workspace_id: id of the workspace for the assignment """ + _MetastoreAssignmentState._configure( + lambda key, value: pulumi.set(__self__, key, value), + default_catalog_name=default_catalog_name, + metastore_id=metastore_id, + workspace_id=workspace_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default_catalog_name: Optional[pulumi.Input[str]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + workspace_id: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if default_catalog_name is not None: - pulumi.set(__self__, "default_catalog_name", default_catalog_name) + _setter("default_catalog_name", default_catalog_name) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if workspace_id is not None: - pulumi.set(__self__, "workspace_id", workspace_id) + _setter("workspace_id", workspace_id) @property @pulumi.getter(name="defaultCatalogName") @@ -206,6 +232,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MetastoreAssignmentArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/metastore_data_access.py b/sdk/python/pulumi_databricks/metastore_data_access.py index fbed4572..9996a063 100644 --- a/sdk/python/pulumi_databricks/metastore_data_access.py +++ b/sdk/python/pulumi_databricks/metastore_data_access.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -37,29 +37,60 @@ def __init__(__self__, *, :param pulumi.Input[str] name: Name of Data Access Configuration, which must be unique within the databricks_metastore. Change forces creation of a new resource. :param pulumi.Input[str] owner: Username/groupname/sp application_id of the data access configuration owner. """ - pulumi.set(__self__, "metastore_id", metastore_id) + MetastoreDataAccessArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + metastore_id=metastore_id, + aws_iam_role=aws_iam_role, + azure_managed_identity=azure_managed_identity, + azure_service_principal=azure_service_principal, + comment=comment, + databricks_gcp_service_account=databricks_gcp_service_account, + force_destroy=force_destroy, + gcp_service_account_key=gcp_service_account_key, + is_default=is_default, + name=name, + owner=owner, + read_only=read_only, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metastore_id: pulumi.Input[str], + aws_iam_role: Optional[pulumi.Input['MetastoreDataAccessAwsIamRoleArgs']] = None, + azure_managed_identity: Optional[pulumi.Input['MetastoreDataAccessAzureManagedIdentityArgs']] = None, + azure_service_principal: Optional[pulumi.Input['MetastoreDataAccessAzureServicePrincipalArgs']] = None, + comment: Optional[pulumi.Input[str]] = None, + databricks_gcp_service_account: Optional[pulumi.Input['MetastoreDataAccessDatabricksGcpServiceAccountArgs']] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + gcp_service_account_key: Optional[pulumi.Input['MetastoreDataAccessGcpServiceAccountKeyArgs']] = None, + is_default: Optional[pulumi.Input[bool]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("metastore_id", metastore_id) if aws_iam_role is not None: - pulumi.set(__self__, "aws_iam_role", aws_iam_role) + _setter("aws_iam_role", aws_iam_role) if azure_managed_identity is not None: - pulumi.set(__self__, "azure_managed_identity", azure_managed_identity) + _setter("azure_managed_identity", azure_managed_identity) if azure_service_principal is not None: - pulumi.set(__self__, "azure_service_principal", azure_service_principal) + _setter("azure_service_principal", azure_service_principal) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if databricks_gcp_service_account is not None: - pulumi.set(__self__, "databricks_gcp_service_account", databricks_gcp_service_account) + _setter("databricks_gcp_service_account", databricks_gcp_service_account) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if gcp_service_account_key is not None: - pulumi.set(__self__, "gcp_service_account_key", gcp_service_account_key) + _setter("gcp_service_account_key", gcp_service_account_key) if is_default is not None: - pulumi.set(__self__, "is_default", is_default) + _setter("is_default", is_default) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) @property @pulumi.getter(name="metastoreId") @@ -208,30 +239,61 @@ def __init__(__self__, *, :param pulumi.Input[str] name: Name of Data Access Configuration, which must be unique within the databricks_metastore. Change forces creation of a new resource. :param pulumi.Input[str] owner: Username/groupname/sp application_id of the data access configuration owner. """ + _MetastoreDataAccessState._configure( + lambda key, value: pulumi.set(__self__, key, value), + aws_iam_role=aws_iam_role, + azure_managed_identity=azure_managed_identity, + azure_service_principal=azure_service_principal, + comment=comment, + databricks_gcp_service_account=databricks_gcp_service_account, + force_destroy=force_destroy, + gcp_service_account_key=gcp_service_account_key, + is_default=is_default, + metastore_id=metastore_id, + name=name, + owner=owner, + read_only=read_only, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + aws_iam_role: Optional[pulumi.Input['MetastoreDataAccessAwsIamRoleArgs']] = None, + azure_managed_identity: Optional[pulumi.Input['MetastoreDataAccessAzureManagedIdentityArgs']] = None, + azure_service_principal: Optional[pulumi.Input['MetastoreDataAccessAzureServicePrincipalArgs']] = None, + comment: Optional[pulumi.Input[str]] = None, + databricks_gcp_service_account: Optional[pulumi.Input['MetastoreDataAccessDatabricksGcpServiceAccountArgs']] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + gcp_service_account_key: Optional[pulumi.Input['MetastoreDataAccessGcpServiceAccountKeyArgs']] = None, + is_default: Optional[pulumi.Input[bool]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if aws_iam_role is not None: - pulumi.set(__self__, "aws_iam_role", aws_iam_role) + _setter("aws_iam_role", aws_iam_role) if azure_managed_identity is not None: - pulumi.set(__self__, "azure_managed_identity", azure_managed_identity) + _setter("azure_managed_identity", azure_managed_identity) if azure_service_principal is not None: - pulumi.set(__self__, "azure_service_principal", azure_service_principal) + _setter("azure_service_principal", azure_service_principal) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if databricks_gcp_service_account is not None: - pulumi.set(__self__, "databricks_gcp_service_account", databricks_gcp_service_account) + _setter("databricks_gcp_service_account", databricks_gcp_service_account) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if gcp_service_account_key is not None: - pulumi.set(__self__, "gcp_service_account_key", gcp_service_account_key) + _setter("gcp_service_account_key", gcp_service_account_key) if is_default is not None: - pulumi.set(__self__, "is_default", is_default) + _setter("is_default", is_default) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) @property @pulumi.getter(name="awsIamRole") @@ -421,6 +483,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MetastoreDataAccessArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -447,12 +513,37 @@ def _internal_init(__self__, raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = MetastoreDataAccessArgs.__new__(MetastoreDataAccessArgs) + if aws_iam_role is not None and not isinstance(aws_iam_role, MetastoreDataAccessAwsIamRoleArgs): + aws_iam_role = aws_iam_role or {} + def _setter(key, value): + aws_iam_role[key] = value + MetastoreDataAccessAwsIamRoleArgs._configure(_setter, **aws_iam_role) __props__.__dict__["aws_iam_role"] = aws_iam_role + if azure_managed_identity is not None and not isinstance(azure_managed_identity, MetastoreDataAccessAzureManagedIdentityArgs): + azure_managed_identity = azure_managed_identity or {} + def _setter(key, value): + azure_managed_identity[key] = value + MetastoreDataAccessAzureManagedIdentityArgs._configure(_setter, **azure_managed_identity) __props__.__dict__["azure_managed_identity"] = azure_managed_identity + if azure_service_principal is not None and not isinstance(azure_service_principal, MetastoreDataAccessAzureServicePrincipalArgs): + azure_service_principal = azure_service_principal or {} + def _setter(key, value): + azure_service_principal[key] = value + MetastoreDataAccessAzureServicePrincipalArgs._configure(_setter, **azure_service_principal) __props__.__dict__["azure_service_principal"] = azure_service_principal __props__.__dict__["comment"] = comment + if databricks_gcp_service_account is not None and not isinstance(databricks_gcp_service_account, MetastoreDataAccessDatabricksGcpServiceAccountArgs): + databricks_gcp_service_account = databricks_gcp_service_account or {} + def _setter(key, value): + databricks_gcp_service_account[key] = value + MetastoreDataAccessDatabricksGcpServiceAccountArgs._configure(_setter, **databricks_gcp_service_account) __props__.__dict__["databricks_gcp_service_account"] = databricks_gcp_service_account __props__.__dict__["force_destroy"] = force_destroy + if gcp_service_account_key is not None and not isinstance(gcp_service_account_key, MetastoreDataAccessGcpServiceAccountKeyArgs): + gcp_service_account_key = gcp_service_account_key or {} + def _setter(key, value): + gcp_service_account_key[key] = value + MetastoreDataAccessGcpServiceAccountKeyArgs._configure(_setter, **gcp_service_account_key) __props__.__dict__["gcp_service_account_key"] = gcp_service_account_key __props__.__dict__["is_default"] = is_default if metastore_id is None and not opts.urn: diff --git a/sdk/python/pulumi_databricks/metastore_provider.py b/sdk/python/pulumi_databricks/metastore_provider.py index 08ff64d4..1cf31a3d 100644 --- a/sdk/python/pulumi_databricks/metastore_provider.py +++ b/sdk/python/pulumi_databricks/metastore_provider.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MetastoreProviderArgs', 'MetastoreProvider'] @@ -25,12 +25,27 @@ def __init__(__self__, *, :param pulumi.Input[str] comment: Description about the provider. :param pulumi.Input[str] name: Name of provider. Change forces creation of a new resource. """ - pulumi.set(__self__, "authentication_type", authentication_type) - pulumi.set(__self__, "recipient_profile_str", recipient_profile_str) + MetastoreProviderArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + authentication_type=authentication_type, + recipient_profile_str=recipient_profile_str, + comment=comment, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + authentication_type: pulumi.Input[str], + recipient_profile_str: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("authentication_type", authentication_type) + _setter("recipient_profile_str", recipient_profile_str) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter(name="authenticationType") @@ -95,14 +110,29 @@ def __init__(__self__, *, :param pulumi.Input[str] name: Name of provider. Change forces creation of a new resource. :param pulumi.Input[str] recipient_profile_str: This is the json file that is created from a recipient url. """ + _MetastoreProviderState._configure( + lambda key, value: pulumi.set(__self__, key, value), + authentication_type=authentication_type, + comment=comment, + name=name, + recipient_profile_str=recipient_profile_str, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + authentication_type: Optional[pulumi.Input[str]] = None, + comment: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + recipient_profile_str: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if authentication_type is not None: - pulumi.set(__self__, "authentication_type", authentication_type) + _setter("authentication_type", authentication_type) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if recipient_profile_str is not None: - pulumi.set(__self__, "recipient_profile_str", recipient_profile_str) + _setter("recipient_profile_str", recipient_profile_str) @property @pulumi.getter(name="authenticationType") @@ -250,6 +280,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MetastoreProviderArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mlflow_experiment.py b/sdk/python/pulumi_databricks/mlflow_experiment.py index b2de3602..7855e949 100644 --- a/sdk/python/pulumi_databricks/mlflow_experiment.py +++ b/sdk/python/pulumi_databricks/mlflow_experiment.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MlflowExperimentArgs', 'MlflowExperiment'] @@ -27,20 +27,41 @@ def __init__(__self__, *, :param pulumi.Input[str] description: The description of the MLflow experiment. :param pulumi.Input[str] name: Name of MLflow experiment. It must be an absolute path within the Databricks workspace, e.g. `/Users//my-experiment`. For more information about changes to experiment naming conventions, see [mlflow docs](https://docs.databricks.com/applications/mlflow/experiments.html#experiment-migration). """ + MlflowExperimentArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + artifact_location=artifact_location, + creation_time=creation_time, + description=description, + experiment_id=experiment_id, + last_update_time=last_update_time, + lifecycle_stage=lifecycle_stage, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + artifact_location: Optional[pulumi.Input[str]] = None, + creation_time: Optional[pulumi.Input[int]] = None, + description: Optional[pulumi.Input[str]] = None, + experiment_id: Optional[pulumi.Input[str]] = None, + last_update_time: Optional[pulumi.Input[int]] = None, + lifecycle_stage: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if artifact_location is not None: - pulumi.set(__self__, "artifact_location", artifact_location) + _setter("artifact_location", artifact_location) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if experiment_id is not None: - pulumi.set(__self__, "experiment_id", experiment_id) + _setter("experiment_id", experiment_id) if last_update_time is not None: - pulumi.set(__self__, "last_update_time", last_update_time) + _setter("last_update_time", last_update_time) if lifecycle_stage is not None: - pulumi.set(__self__, "lifecycle_stage", lifecycle_stage) + _setter("lifecycle_stage", lifecycle_stage) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter(name="artifactLocation") @@ -131,20 +152,41 @@ def __init__(__self__, *, :param pulumi.Input[str] description: The description of the MLflow experiment. :param pulumi.Input[str] name: Name of MLflow experiment. It must be an absolute path within the Databricks workspace, e.g. `/Users//my-experiment`. For more information about changes to experiment naming conventions, see [mlflow docs](https://docs.databricks.com/applications/mlflow/experiments.html#experiment-migration). """ + _MlflowExperimentState._configure( + lambda key, value: pulumi.set(__self__, key, value), + artifact_location=artifact_location, + creation_time=creation_time, + description=description, + experiment_id=experiment_id, + last_update_time=last_update_time, + lifecycle_stage=lifecycle_stage, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + artifact_location: Optional[pulumi.Input[str]] = None, + creation_time: Optional[pulumi.Input[int]] = None, + description: Optional[pulumi.Input[str]] = None, + experiment_id: Optional[pulumi.Input[str]] = None, + last_update_time: Optional[pulumi.Input[int]] = None, + lifecycle_stage: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if artifact_location is not None: - pulumi.set(__self__, "artifact_location", artifact_location) + _setter("artifact_location", artifact_location) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if experiment_id is not None: - pulumi.set(__self__, "experiment_id", experiment_id) + _setter("experiment_id", experiment_id) if last_update_time is not None: - pulumi.set(__self__, "last_update_time", last_update_time) + _setter("last_update_time", last_update_time) if lifecycle_stage is not None: - pulumi.set(__self__, "lifecycle_stage", lifecycle_stage) + _setter("lifecycle_stage", lifecycle_stage) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter(name="artifactLocation") @@ -328,6 +370,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MlflowExperimentArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mlflow_model.py b/sdk/python/pulumi_databricks/mlflow_model.py index 390983aa..3cfafcb5 100644 --- a/sdk/python/pulumi_databricks/mlflow_model.py +++ b/sdk/python/pulumi_databricks/mlflow_model.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -28,18 +28,37 @@ def __init__(__self__, *, :param pulumi.Input[str] name: Name of MLflow model. Change of name triggers new resource. :param pulumi.Input[Sequence[pulumi.Input['MlflowModelTagArgs']]] tags: Tags for the MLflow model. """ + MlflowModelArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + creation_timestamp=creation_timestamp, + description=description, + last_updated_timestamp=last_updated_timestamp, + name=name, + tags=tags, + user_id=user_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + creation_timestamp: Optional[pulumi.Input[int]] = None, + description: Optional[pulumi.Input[str]] = None, + last_updated_timestamp: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input['MlflowModelTagArgs']]]] = None, + user_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if creation_timestamp is not None: - pulumi.set(__self__, "creation_timestamp", creation_timestamp) + _setter("creation_timestamp", creation_timestamp) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if last_updated_timestamp is not None: - pulumi.set(__self__, "last_updated_timestamp", last_updated_timestamp) + _setter("last_updated_timestamp", last_updated_timestamp) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if user_id is not None: - pulumi.set(__self__, "user_id", user_id) + _setter("user_id", user_id) @property @pulumi.getter(name="creationTimestamp") @@ -121,20 +140,41 @@ def __init__(__self__, *, :param pulumi.Input[str] name: Name of MLflow model. Change of name triggers new resource. :param pulumi.Input[Sequence[pulumi.Input['MlflowModelTagArgs']]] tags: Tags for the MLflow model. """ + _MlflowModelState._configure( + lambda key, value: pulumi.set(__self__, key, value), + creation_timestamp=creation_timestamp, + description=description, + last_updated_timestamp=last_updated_timestamp, + name=name, + registered_model_id=registered_model_id, + tags=tags, + user_id=user_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + creation_timestamp: Optional[pulumi.Input[int]] = None, + description: Optional[pulumi.Input[str]] = None, + last_updated_timestamp: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + registered_model_id: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input['MlflowModelTagArgs']]]] = None, + user_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if creation_timestamp is not None: - pulumi.set(__self__, "creation_timestamp", creation_timestamp) + _setter("creation_timestamp", creation_timestamp) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if last_updated_timestamp is not None: - pulumi.set(__self__, "last_updated_timestamp", last_updated_timestamp) + _setter("last_updated_timestamp", last_updated_timestamp) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if registered_model_id is not None: - pulumi.set(__self__, "registered_model_id", registered_model_id) + _setter("registered_model_id", registered_model_id) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if user_id is not None: - pulumi.set(__self__, "user_id", user_id) + _setter("user_id", user_id) @property @pulumi.getter(name="creationTimestamp") @@ -335,6 +375,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MlflowModelArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mlflow_webhook.py b/sdk/python/pulumi_databricks/mlflow_webhook.py index be979967..68699f96 100644 --- a/sdk/python/pulumi_databricks/mlflow_webhook.py +++ b/sdk/python/pulumi_databricks/mlflow_webhook.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -31,17 +31,36 @@ def __init__(__self__, *, :param pulumi.Input[str] model_name: Name of MLflow model for which webhook will be created. If the model name is not specified, a registry-wide webhook is created that listens for the specified events across all versions of all registered models. :param pulumi.Input[str] status: Optional status of webhook. Possible values are `ACTIVE`, `TEST_MODE`, `DISABLED`. Default is `ACTIVE`. """ - pulumi.set(__self__, "events", events) + MlflowWebhookArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + events=events, + description=description, + http_url_spec=http_url_spec, + job_spec=job_spec, + model_name=model_name, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + events: pulumi.Input[Sequence[pulumi.Input[str]]], + description: Optional[pulumi.Input[str]] = None, + http_url_spec: Optional[pulumi.Input['MlflowWebhookHttpUrlSpecArgs']] = None, + job_spec: Optional[pulumi.Input['MlflowWebhookJobSpecArgs']] = None, + model_name: Optional[pulumi.Input[str]] = None, + status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("events", events) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if http_url_spec is not None: - pulumi.set(__self__, "http_url_spec", http_url_spec) + _setter("http_url_spec", http_url_spec) if job_spec is not None: - pulumi.set(__self__, "job_spec", job_spec) + _setter("job_spec", job_spec) if model_name is not None: - pulumi.set(__self__, "model_name", model_name) + _setter("model_name", model_name) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter @@ -130,18 +149,37 @@ def __init__(__self__, *, :param pulumi.Input[str] model_name: Name of MLflow model for which webhook will be created. If the model name is not specified, a registry-wide webhook is created that listens for the specified events across all versions of all registered models. :param pulumi.Input[str] status: Optional status of webhook. Possible values are `ACTIVE`, `TEST_MODE`, `DISABLED`. Default is `ACTIVE`. """ + _MlflowWebhookState._configure( + lambda key, value: pulumi.set(__self__, key, value), + description=description, + events=events, + http_url_spec=http_url_spec, + job_spec=job_spec, + model_name=model_name, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + description: Optional[pulumi.Input[str]] = None, + events: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + http_url_spec: Optional[pulumi.Input['MlflowWebhookHttpUrlSpecArgs']] = None, + job_spec: Optional[pulumi.Input['MlflowWebhookJobSpecArgs']] = None, + model_name: Optional[pulumi.Input[str]] = None, + status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if events is not None: - pulumi.set(__self__, "events", events) + _setter("events", events) if http_url_spec is not None: - pulumi.set(__self__, "http_url_spec", http_url_spec) + _setter("http_url_spec", http_url_spec) if job_spec is not None: - pulumi.set(__self__, "job_spec", job_spec) + _setter("job_spec", job_spec) if model_name is not None: - pulumi.set(__self__, "model_name", model_name) + _setter("model_name", model_name) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter @@ -323,6 +361,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MlflowWebhookArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -347,7 +389,17 @@ def _internal_init(__self__, if events is None and not opts.urn: raise TypeError("Missing required property 'events'") __props__.__dict__["events"] = events + if http_url_spec is not None and not isinstance(http_url_spec, MlflowWebhookHttpUrlSpecArgs): + http_url_spec = http_url_spec or {} + def _setter(key, value): + http_url_spec[key] = value + MlflowWebhookHttpUrlSpecArgs._configure(_setter, **http_url_spec) __props__.__dict__["http_url_spec"] = http_url_spec + if job_spec is not None and not isinstance(job_spec, MlflowWebhookJobSpecArgs): + job_spec = job_spec or {} + def _setter(key, value): + job_spec[key] = value + MlflowWebhookJobSpecArgs._configure(_setter, **job_spec) __props__.__dict__["job_spec"] = job_spec __props__.__dict__["model_name"] = model_name __props__.__dict__["status"] = status diff --git a/sdk/python/pulumi_databricks/model_serving.py b/sdk/python/pulumi_databricks/model_serving.py index 7f72f217..c8b01632 100644 --- a/sdk/python/pulumi_databricks/model_serving.py +++ b/sdk/python/pulumi_databricks/model_serving.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -24,11 +24,24 @@ def __init__(__self__, *, :param pulumi.Input['ModelServingConfigArgs'] config: The model serving endpoint configuration. :param pulumi.Input[str] name: The name of the model serving endpoint. This field is required and must be unique across a workspace. An endpoint name can consist of alphanumeric characters, dashes, and underscores. NOTE: Changing this name will delete the existing endpoint and create a new endpoint with the update name. """ - pulumi.set(__self__, "config", config) + ModelServingArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + config=config, + name=name, + tags=tags, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + config: pulumi.Input['ModelServingConfigArgs'], + name: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input['ModelServingTagArgs']]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("config", config) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) @property @pulumi.getter @@ -77,14 +90,29 @@ def __init__(__self__, *, :param pulumi.Input[str] name: The name of the model serving endpoint. This field is required and must be unique across a workspace. An endpoint name can consist of alphanumeric characters, dashes, and underscores. NOTE: Changing this name will delete the existing endpoint and create a new endpoint with the update name. :param pulumi.Input[str] serving_endpoint_id: Unique identifier of the serving endpoint primarily used to set permissions and refer to this instance for other operations. """ + _ModelServingState._configure( + lambda key, value: pulumi.set(__self__, key, value), + config=config, + name=name, + serving_endpoint_id=serving_endpoint_id, + tags=tags, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + config: Optional[pulumi.Input['ModelServingConfigArgs']] = None, + name: Optional[pulumi.Input[str]] = None, + serving_endpoint_id: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input['ModelServingTagArgs']]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if config is not None: - pulumi.set(__self__, "config", config) + _setter("config", config) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if serving_endpoint_id is not None: - pulumi.set(__self__, "serving_endpoint_id", serving_endpoint_id) + _setter("serving_endpoint_id", serving_endpoint_id) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) @property @pulumi.getter @@ -288,6 +316,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ModelServingArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -305,6 +337,11 @@ def _internal_init(__self__, raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = ModelServingArgs.__new__(ModelServingArgs) + if config is not None and not isinstance(config, ModelServingConfigArgs): + config = config or {} + def _setter(key, value): + config[key] = value + ModelServingConfigArgs._configure(_setter, **config) if config is None and not opts.urn: raise TypeError("Missing required property 'config'") __props__.__dict__["config"] = config diff --git a/sdk/python/pulumi_databricks/mount.py b/sdk/python/pulumi_databricks/mount.py index 72cb5156..0c3a51bb 100644 --- a/sdk/python/pulumi_databricks/mount.py +++ b/sdk/python/pulumi_databricks/mount.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -30,28 +30,57 @@ def __init__(__self__, *, """ The set of arguments for constructing a Mount resource. """ + MountArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfs=abfs, + adl=adl, + cluster_id=cluster_id, + encryption_type=encryption_type, + extra_configs=extra_configs, + gs=gs, + name=name, + resource_id=resource_id, + s3=s3, + uri=uri, + wasb=wasb, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfs: Optional[pulumi.Input['MountAbfsArgs']] = None, + adl: Optional[pulumi.Input['MountAdlArgs']] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + extra_configs: Optional[pulumi.Input[Mapping[str, Any]]] = None, + gs: Optional[pulumi.Input['MountGsArgs']] = None, + name: Optional[pulumi.Input[str]] = None, + resource_id: Optional[pulumi.Input[str]] = None, + s3: Optional[pulumi.Input['MountS3Args']] = None, + uri: Optional[pulumi.Input[str]] = None, + wasb: Optional[pulumi.Input['MountWasbArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfs is not None: - pulumi.set(__self__, "abfs", abfs) + _setter("abfs", abfs) if adl is not None: - pulumi.set(__self__, "adl", adl) + _setter("adl", adl) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if extra_configs is not None: - pulumi.set(__self__, "extra_configs", extra_configs) + _setter("extra_configs", extra_configs) if gs is not None: - pulumi.set(__self__, "gs", gs) + _setter("gs", gs) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if resource_id is not None: - pulumi.set(__self__, "resource_id", resource_id) + _setter("resource_id", resource_id) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if uri is not None: - pulumi.set(__self__, "uri", uri) + _setter("uri", uri) if wasb is not None: - pulumi.set(__self__, "wasb", wasb) + _setter("wasb", wasb) @property @pulumi.getter @@ -172,30 +201,61 @@ def __init__(__self__, *, Input properties used for looking up and filtering Mount resources. :param pulumi.Input[str] source: (String) HDFS-compatible url """ + _MountState._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfs=abfs, + adl=adl, + cluster_id=cluster_id, + encryption_type=encryption_type, + extra_configs=extra_configs, + gs=gs, + name=name, + resource_id=resource_id, + s3=s3, + source=source, + uri=uri, + wasb=wasb, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfs: Optional[pulumi.Input['MountAbfsArgs']] = None, + adl: Optional[pulumi.Input['MountAdlArgs']] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + encryption_type: Optional[pulumi.Input[str]] = None, + extra_configs: Optional[pulumi.Input[Mapping[str, Any]]] = None, + gs: Optional[pulumi.Input['MountGsArgs']] = None, + name: Optional[pulumi.Input[str]] = None, + resource_id: Optional[pulumi.Input[str]] = None, + s3: Optional[pulumi.Input['MountS3Args']] = None, + source: Optional[pulumi.Input[str]] = None, + uri: Optional[pulumi.Input[str]] = None, + wasb: Optional[pulumi.Input['MountWasbArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfs is not None: - pulumi.set(__self__, "abfs", abfs) + _setter("abfs", abfs) if adl is not None: - pulumi.set(__self__, "adl", adl) + _setter("adl", adl) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if extra_configs is not None: - pulumi.set(__self__, "extra_configs", extra_configs) + _setter("extra_configs", extra_configs) if gs is not None: - pulumi.set(__self__, "gs", gs) + _setter("gs", gs) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if resource_id is not None: - pulumi.set(__self__, "resource_id", resource_id) + _setter("resource_id", resource_id) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) if uri is not None: - pulumi.set(__self__, "uri", uri) + _setter("uri", uri) if wasb is not None: - pulumi.set(__self__, "wasb", wasb) + _setter("wasb", wasb) @property @pulumi.getter @@ -355,6 +415,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MountArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -380,16 +444,41 @@ def _internal_init(__self__, raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = MountArgs.__new__(MountArgs) + if abfs is not None and not isinstance(abfs, MountAbfsArgs): + abfs = abfs or {} + def _setter(key, value): + abfs[key] = value + MountAbfsArgs._configure(_setter, **abfs) __props__.__dict__["abfs"] = abfs + if adl is not None and not isinstance(adl, MountAdlArgs): + adl = adl or {} + def _setter(key, value): + adl[key] = value + MountAdlArgs._configure(_setter, **adl) __props__.__dict__["adl"] = adl __props__.__dict__["cluster_id"] = cluster_id __props__.__dict__["encryption_type"] = encryption_type __props__.__dict__["extra_configs"] = extra_configs + if gs is not None and not isinstance(gs, MountGsArgs): + gs = gs or {} + def _setter(key, value): + gs[key] = value + MountGsArgs._configure(_setter, **gs) __props__.__dict__["gs"] = gs __props__.__dict__["name"] = name __props__.__dict__["resource_id"] = resource_id + if s3 is not None and not isinstance(s3, MountS3Args): + s3 = s3 or {} + def _setter(key, value): + s3[key] = value + MountS3Args._configure(_setter, **s3) __props__.__dict__["s3"] = s3 __props__.__dict__["uri"] = uri + if wasb is not None and not isinstance(wasb, MountWasbArgs): + wasb = wasb or {} + def _setter(key, value): + wasb[key] = value + MountWasbArgs._configure(_setter, **wasb) __props__.__dict__["wasb"] = wasb __props__.__dict__["source"] = None super(Mount, __self__).__init__( diff --git a/sdk/python/pulumi_databricks/mws_credentials.py b/sdk/python/pulumi_databricks/mws_credentials.py index a00e6e6d..f9a941a7 100644 --- a/sdk/python/pulumi_databricks/mws_credentials.py +++ b/sdk/python/pulumi_databricks/mws_credentials.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MwsCredentialsArgs', 'MwsCredentials'] @@ -23,9 +23,22 @@ def __init__(__self__, *, :param pulumi.Input[str] credentials_name: name of credentials to register :param pulumi.Input[str] role_arn: ARN of cross-account role """ - pulumi.set(__self__, "account_id", account_id) - pulumi.set(__self__, "credentials_name", credentials_name) - pulumi.set(__self__, "role_arn", role_arn) + MwsCredentialsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + credentials_name=credentials_name, + role_arn=role_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: pulumi.Input[str], + credentials_name: pulumi.Input[str], + role_arn: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("account_id", account_id) + _setter("credentials_name", credentials_name) + _setter("role_arn", role_arn) @property @pulumi.getter(name="accountId") @@ -81,18 +94,37 @@ def __init__(__self__, *, :param pulumi.Input[str] credentials_name: name of credentials to register :param pulumi.Input[str] role_arn: ARN of cross-account role """ + _MwsCredentialsState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + creation_time=creation_time, + credentials_id=credentials_id, + credentials_name=credentials_name, + external_id=external_id, + role_arn=role_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + creation_time: Optional[pulumi.Input[int]] = None, + credentials_id: Optional[pulumi.Input[str]] = None, + credentials_name: Optional[pulumi.Input[str]] = None, + external_id: Optional[pulumi.Input[str]] = None, + role_arn: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if credentials_id is not None: - pulumi.set(__self__, "credentials_id", credentials_id) + _setter("credentials_id", credentials_id) if credentials_name is not None: - pulumi.set(__self__, "credentials_name", credentials_name) + _setter("credentials_name", credentials_name) if external_id is not None: - pulumi.set(__self__, "external_id", external_id) + _setter("external_id", external_id) if role_arn is not None: - pulumi.set(__self__, "role_arn", role_arn) + _setter("role_arn", role_arn) @property @pulumi.getter(name="accountId") @@ -273,6 +305,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsCredentialsArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mws_customer_managed_keys.py b/sdk/python/pulumi_databricks/mws_customer_managed_keys.py index e3ede4ca..ea46aa68 100644 --- a/sdk/python/pulumi_databricks/mws_customer_managed_keys.py +++ b/sdk/python/pulumi_databricks/mws_customer_managed_keys.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -31,16 +31,35 @@ def __init__(__self__, *, :param pulumi.Input[str] customer_managed_key_id: (String) ID of the encryption key configuration object. :param pulumi.Input['MwsCustomerManagedKeysGcpKeyInfoArgs'] gcp_key_info: This field is a block and is documented below. This conflicts with `aws_key_info` """ - pulumi.set(__self__, "account_id", account_id) - pulumi.set(__self__, "use_cases", use_cases) + MwsCustomerManagedKeysArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + use_cases=use_cases, + aws_key_info=aws_key_info, + creation_time=creation_time, + customer_managed_key_id=customer_managed_key_id, + gcp_key_info=gcp_key_info, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: pulumi.Input[str], + use_cases: pulumi.Input[Sequence[pulumi.Input[str]]], + aws_key_info: Optional[pulumi.Input['MwsCustomerManagedKeysAwsKeyInfoArgs']] = None, + creation_time: Optional[pulumi.Input[int]] = None, + customer_managed_key_id: Optional[pulumi.Input[str]] = None, + gcp_key_info: Optional[pulumi.Input['MwsCustomerManagedKeysGcpKeyInfoArgs']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("account_id", account_id) + _setter("use_cases", use_cases) if aws_key_info is not None: - pulumi.set(__self__, "aws_key_info", aws_key_info) + _setter("aws_key_info", aws_key_info) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if customer_managed_key_id is not None: - pulumi.set(__self__, "customer_managed_key_id", customer_managed_key_id) + _setter("customer_managed_key_id", customer_managed_key_id) if gcp_key_info is not None: - pulumi.set(__self__, "gcp_key_info", gcp_key_info) + _setter("gcp_key_info", gcp_key_info) @property @pulumi.getter(name="accountId") @@ -133,18 +152,37 @@ def __init__(__self__, *, :param pulumi.Input['MwsCustomerManagedKeysGcpKeyInfoArgs'] gcp_key_info: This field is a block and is documented below. This conflicts with `aws_key_info` :param pulumi.Input[Sequence[pulumi.Input[str]]] use_cases: *(since v0.3.4)* List of use cases for which this key will be used. *If you've used the resource before, please add `use_cases = ["MANAGED_SERVICES"]` to keep the previous behaviour.* Possible values are: """ + _MwsCustomerManagedKeysState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + aws_key_info=aws_key_info, + creation_time=creation_time, + customer_managed_key_id=customer_managed_key_id, + gcp_key_info=gcp_key_info, + use_cases=use_cases, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + aws_key_info: Optional[pulumi.Input['MwsCustomerManagedKeysAwsKeyInfoArgs']] = None, + creation_time: Optional[pulumi.Input[int]] = None, + customer_managed_key_id: Optional[pulumi.Input[str]] = None, + gcp_key_info: Optional[pulumi.Input['MwsCustomerManagedKeysGcpKeyInfoArgs']] = None, + use_cases: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if aws_key_info is not None: - pulumi.set(__self__, "aws_key_info", aws_key_info) + _setter("aws_key_info", aws_key_info) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if customer_managed_key_id is not None: - pulumi.set(__self__, "customer_managed_key_id", customer_managed_key_id) + _setter("customer_managed_key_id", customer_managed_key_id) if gcp_key_info is not None: - pulumi.set(__self__, "gcp_key_info", gcp_key_info) + _setter("gcp_key_info", gcp_key_info) if use_cases is not None: - pulumi.set(__self__, "use_cases", use_cases) + _setter("use_cases", use_cases) @property @pulumi.getter(name="accountId") @@ -636,6 +674,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsCustomerManagedKeysArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -659,9 +701,19 @@ def _internal_init(__self__, if account_id is None and not opts.urn: raise TypeError("Missing required property 'account_id'") __props__.__dict__["account_id"] = account_id + if aws_key_info is not None and not isinstance(aws_key_info, MwsCustomerManagedKeysAwsKeyInfoArgs): + aws_key_info = aws_key_info or {} + def _setter(key, value): + aws_key_info[key] = value + MwsCustomerManagedKeysAwsKeyInfoArgs._configure(_setter, **aws_key_info) __props__.__dict__["aws_key_info"] = aws_key_info __props__.__dict__["creation_time"] = creation_time __props__.__dict__["customer_managed_key_id"] = customer_managed_key_id + if gcp_key_info is not None and not isinstance(gcp_key_info, MwsCustomerManagedKeysGcpKeyInfoArgs): + gcp_key_info = gcp_key_info or {} + def _setter(key, value): + gcp_key_info[key] = value + MwsCustomerManagedKeysGcpKeyInfoArgs._configure(_setter, **gcp_key_info) __props__.__dict__["gcp_key_info"] = gcp_key_info if use_cases is None and not opts.urn: raise TypeError("Missing required property 'use_cases'") diff --git a/sdk/python/pulumi_databricks/mws_log_delivery.py b/sdk/python/pulumi_databricks/mws_log_delivery.py index 9c917bf2..8bbc1764 100644 --- a/sdk/python/pulumi_databricks/mws_log_delivery.py +++ b/sdk/python/pulumi_databricks/mws_log_delivery.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MwsLogDeliveryArgs', 'MwsLogDelivery'] @@ -39,23 +39,52 @@ def __init__(__self__, *, :param pulumi.Input[str] status: Status of log delivery configuration. Set to ENABLED or DISABLED. Defaults to ENABLED. This is the only field you can update. :param pulumi.Input[Sequence[pulumi.Input[int]]] workspace_ids_filters: By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces. """ - pulumi.set(__self__, "account_id", account_id) - pulumi.set(__self__, "credentials_id", credentials_id) - pulumi.set(__self__, "log_type", log_type) - pulumi.set(__self__, "output_format", output_format) - pulumi.set(__self__, "storage_configuration_id", storage_configuration_id) + MwsLogDeliveryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + credentials_id=credentials_id, + log_type=log_type, + output_format=output_format, + storage_configuration_id=storage_configuration_id, + config_id=config_id, + config_name=config_name, + delivery_path_prefix=delivery_path_prefix, + delivery_start_time=delivery_start_time, + status=status, + workspace_ids_filters=workspace_ids_filters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: pulumi.Input[str], + credentials_id: pulumi.Input[str], + log_type: pulumi.Input[str], + output_format: pulumi.Input[str], + storage_configuration_id: pulumi.Input[str], + config_id: Optional[pulumi.Input[str]] = None, + config_name: Optional[pulumi.Input[str]] = None, + delivery_path_prefix: Optional[pulumi.Input[str]] = None, + delivery_start_time: Optional[pulumi.Input[str]] = None, + status: Optional[pulumi.Input[str]] = None, + workspace_ids_filters: Optional[pulumi.Input[Sequence[pulumi.Input[int]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("account_id", account_id) + _setter("credentials_id", credentials_id) + _setter("log_type", log_type) + _setter("output_format", output_format) + _setter("storage_configuration_id", storage_configuration_id) if config_id is not None: - pulumi.set(__self__, "config_id", config_id) + _setter("config_id", config_id) if config_name is not None: - pulumi.set(__self__, "config_name", config_name) + _setter("config_name", config_name) if delivery_path_prefix is not None: - pulumi.set(__self__, "delivery_path_prefix", delivery_path_prefix) + _setter("delivery_path_prefix", delivery_path_prefix) if delivery_start_time is not None: - pulumi.set(__self__, "delivery_start_time", delivery_start_time) + _setter("delivery_start_time", delivery_start_time) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) if workspace_ids_filters is not None: - pulumi.set(__self__, "workspace_ids_filters", workspace_ids_filters) + _setter("workspace_ids_filters", workspace_ids_filters) @property @pulumi.getter(name="accountId") @@ -218,28 +247,57 @@ def __init__(__self__, *, :param pulumi.Input[str] storage_configuration_id: The ID for a Databricks storage configuration that represents the S3 bucket with bucket policy as described in the main billable usage documentation page. :param pulumi.Input[Sequence[pulumi.Input[int]]] workspace_ids_filters: By default, this log configuration applies to all workspaces associated with your account ID. If your account is on the E2 version of the platform or on a select custom plan that allows multiple workspaces per account, you may have multiple workspaces associated with your account ID. You can optionally set the field as mentioned earlier to an array of workspace IDs. If you plan to use different log delivery configurations for several workspaces, set this explicitly rather than leaving it blank. If you leave this blank and your account ID gets additional workspaces in the future, this configuration will also apply to the new workspaces. """ + _MwsLogDeliveryState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + config_id=config_id, + config_name=config_name, + credentials_id=credentials_id, + delivery_path_prefix=delivery_path_prefix, + delivery_start_time=delivery_start_time, + log_type=log_type, + output_format=output_format, + status=status, + storage_configuration_id=storage_configuration_id, + workspace_ids_filters=workspace_ids_filters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + config_id: Optional[pulumi.Input[str]] = None, + config_name: Optional[pulumi.Input[str]] = None, + credentials_id: Optional[pulumi.Input[str]] = None, + delivery_path_prefix: Optional[pulumi.Input[str]] = None, + delivery_start_time: Optional[pulumi.Input[str]] = None, + log_type: Optional[pulumi.Input[str]] = None, + output_format: Optional[pulumi.Input[str]] = None, + status: Optional[pulumi.Input[str]] = None, + storage_configuration_id: Optional[pulumi.Input[str]] = None, + workspace_ids_filters: Optional[pulumi.Input[Sequence[pulumi.Input[int]]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if config_id is not None: - pulumi.set(__self__, "config_id", config_id) + _setter("config_id", config_id) if config_name is not None: - pulumi.set(__self__, "config_name", config_name) + _setter("config_name", config_name) if credentials_id is not None: - pulumi.set(__self__, "credentials_id", credentials_id) + _setter("credentials_id", credentials_id) if delivery_path_prefix is not None: - pulumi.set(__self__, "delivery_path_prefix", delivery_path_prefix) + _setter("delivery_path_prefix", delivery_path_prefix) if delivery_start_time is not None: - pulumi.set(__self__, "delivery_start_time", delivery_start_time) + _setter("delivery_start_time", delivery_start_time) if log_type is not None: - pulumi.set(__self__, "log_type", log_type) + _setter("log_type", log_type) if output_format is not None: - pulumi.set(__self__, "output_format", output_format) + _setter("output_format", output_format) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) if storage_configuration_id is not None: - pulumi.set(__self__, "storage_configuration_id", storage_configuration_id) + _setter("storage_configuration_id", storage_configuration_id) if workspace_ids_filters is not None: - pulumi.set(__self__, "workspace_ids_filters", workspace_ids_filters) + _setter("workspace_ids_filters", workspace_ids_filters) @property @pulumi.getter(name="accountId") @@ -545,6 +603,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsLogDeliveryArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mws_networks.py b/sdk/python/pulumi_databricks/mws_networks.py index 8e362e73..8e150aee 100644 --- a/sdk/python/pulumi_databricks/mws_networks.py +++ b/sdk/python/pulumi_databricks/mws_networks.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -41,28 +41,59 @@ def __init__(__self__, *, :param pulumi.Input[str] vpc_status: (String) VPC attachment status :param pulumi.Input[int] workspace_id: (Integer) id of associated workspace """ - pulumi.set(__self__, "account_id", account_id) - pulumi.set(__self__, "network_name", network_name) + MwsNetworksArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + network_name=network_name, + creation_time=creation_time, + error_messages=error_messages, + gcp_network_info=gcp_network_info, + network_id=network_id, + security_group_ids=security_group_ids, + subnet_ids=subnet_ids, + vpc_endpoints=vpc_endpoints, + vpc_id=vpc_id, + vpc_status=vpc_status, + workspace_id=workspace_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: pulumi.Input[str], + network_name: pulumi.Input[str], + creation_time: Optional[pulumi.Input[int]] = None, + error_messages: Optional[pulumi.Input[Sequence[pulumi.Input['MwsNetworksErrorMessageArgs']]]] = None, + gcp_network_info: Optional[pulumi.Input['MwsNetworksGcpNetworkInfoArgs']] = None, + network_id: Optional[pulumi.Input[str]] = None, + security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + subnet_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + vpc_endpoints: Optional[pulumi.Input['MwsNetworksVpcEndpointsArgs']] = None, + vpc_id: Optional[pulumi.Input[str]] = None, + vpc_status: Optional[pulumi.Input[str]] = None, + workspace_id: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("account_id", account_id) + _setter("network_name", network_name) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if error_messages is not None: - pulumi.set(__self__, "error_messages", error_messages) + _setter("error_messages", error_messages) if gcp_network_info is not None: - pulumi.set(__self__, "gcp_network_info", gcp_network_info) + _setter("gcp_network_info", gcp_network_info) if network_id is not None: - pulumi.set(__self__, "network_id", network_id) + _setter("network_id", network_id) if security_group_ids is not None: - pulumi.set(__self__, "security_group_ids", security_group_ids) + _setter("security_group_ids", security_group_ids) if subnet_ids is not None: - pulumi.set(__self__, "subnet_ids", subnet_ids) + _setter("subnet_ids", subnet_ids) if vpc_endpoints is not None: - pulumi.set(__self__, "vpc_endpoints", vpc_endpoints) + _setter("vpc_endpoints", vpc_endpoints) if vpc_id is not None: - pulumi.set(__self__, "vpc_id", vpc_id) + _setter("vpc_id", vpc_id) if vpc_status is not None: - pulumi.set(__self__, "vpc_status", vpc_status) + _setter("vpc_status", vpc_status) if workspace_id is not None: - pulumi.set(__self__, "workspace_id", workspace_id) + _setter("workspace_id", workspace_id) @property @pulumi.getter(name="accountId") @@ -231,30 +262,61 @@ def __init__(__self__, *, :param pulumi.Input[str] vpc_status: (String) VPC attachment status :param pulumi.Input[int] workspace_id: (Integer) id of associated workspace """ + _MwsNetworksState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + creation_time=creation_time, + error_messages=error_messages, + gcp_network_info=gcp_network_info, + network_id=network_id, + network_name=network_name, + security_group_ids=security_group_ids, + subnet_ids=subnet_ids, + vpc_endpoints=vpc_endpoints, + vpc_id=vpc_id, + vpc_status=vpc_status, + workspace_id=workspace_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + creation_time: Optional[pulumi.Input[int]] = None, + error_messages: Optional[pulumi.Input[Sequence[pulumi.Input['MwsNetworksErrorMessageArgs']]]] = None, + gcp_network_info: Optional[pulumi.Input['MwsNetworksGcpNetworkInfoArgs']] = None, + network_id: Optional[pulumi.Input[str]] = None, + network_name: Optional[pulumi.Input[str]] = None, + security_group_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + subnet_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + vpc_endpoints: Optional[pulumi.Input['MwsNetworksVpcEndpointsArgs']] = None, + vpc_id: Optional[pulumi.Input[str]] = None, + vpc_status: Optional[pulumi.Input[str]] = None, + workspace_id: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if error_messages is not None: - pulumi.set(__self__, "error_messages", error_messages) + _setter("error_messages", error_messages) if gcp_network_info is not None: - pulumi.set(__self__, "gcp_network_info", gcp_network_info) + _setter("gcp_network_info", gcp_network_info) if network_id is not None: - pulumi.set(__self__, "network_id", network_id) + _setter("network_id", network_id) if network_name is not None: - pulumi.set(__self__, "network_name", network_name) + _setter("network_name", network_name) if security_group_ids is not None: - pulumi.set(__self__, "security_group_ids", security_group_ids) + _setter("security_group_ids", security_group_ids) if subnet_ids is not None: - pulumi.set(__self__, "subnet_ids", subnet_ids) + _setter("subnet_ids", subnet_ids) if vpc_endpoints is not None: - pulumi.set(__self__, "vpc_endpoints", vpc_endpoints) + _setter("vpc_endpoints", vpc_endpoints) if vpc_id is not None: - pulumi.set(__self__, "vpc_id", vpc_id) + _setter("vpc_id", vpc_id) if vpc_status is not None: - pulumi.set(__self__, "vpc_status", vpc_status) + _setter("vpc_status", vpc_status) if workspace_id is not None: - pulumi.set(__self__, "workspace_id", workspace_id) + _setter("workspace_id", workspace_id) @property @pulumi.getter(name="accountId") @@ -726,6 +788,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsNetworksArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -757,6 +823,11 @@ def _internal_init(__self__, __props__.__dict__["account_id"] = None if account_id is None else pulumi.Output.secret(account_id) __props__.__dict__["creation_time"] = creation_time __props__.__dict__["error_messages"] = error_messages + if gcp_network_info is not None and not isinstance(gcp_network_info, MwsNetworksGcpNetworkInfoArgs): + gcp_network_info = gcp_network_info or {} + def _setter(key, value): + gcp_network_info[key] = value + MwsNetworksGcpNetworkInfoArgs._configure(_setter, **gcp_network_info) __props__.__dict__["gcp_network_info"] = gcp_network_info __props__.__dict__["network_id"] = network_id if network_name is None and not opts.urn: @@ -764,6 +835,11 @@ def _internal_init(__self__, __props__.__dict__["network_name"] = network_name __props__.__dict__["security_group_ids"] = security_group_ids __props__.__dict__["subnet_ids"] = subnet_ids + if vpc_endpoints is not None and not isinstance(vpc_endpoints, MwsNetworksVpcEndpointsArgs): + vpc_endpoints = vpc_endpoints or {} + def _setter(key, value): + vpc_endpoints[key] = value + MwsNetworksVpcEndpointsArgs._configure(_setter, **vpc_endpoints) __props__.__dict__["vpc_endpoints"] = vpc_endpoints __props__.__dict__["vpc_id"] = vpc_id __props__.__dict__["vpc_status"] = vpc_status diff --git a/sdk/python/pulumi_databricks/mws_permission_assignment.py b/sdk/python/pulumi_databricks/mws_permission_assignment.py index d3be2e2c..cfcbfefa 100644 --- a/sdk/python/pulumi_databricks/mws_permission_assignment.py +++ b/sdk/python/pulumi_databricks/mws_permission_assignment.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MwsPermissionAssignmentArgs', 'MwsPermissionAssignment'] @@ -25,9 +25,22 @@ def __init__(__self__, *, :param pulumi.Input[int] principal_id: Databricks ID of the user, service principal, or group. The principal ID can be retrieved using the SCIM API, or using databricks_user, ServicePrincipal or Group data sources. :param pulumi.Input[int] workspace_id: Databricks workspace ID. """ - pulumi.set(__self__, "permissions", permissions) - pulumi.set(__self__, "principal_id", principal_id) - pulumi.set(__self__, "workspace_id", workspace_id) + MwsPermissionAssignmentArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + permissions=permissions, + principal_id=principal_id, + workspace_id=workspace_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permissions: pulumi.Input[Sequence[pulumi.Input[str]]], + principal_id: pulumi.Input[int], + workspace_id: pulumi.Input[int], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("permissions", permissions) + _setter("principal_id", principal_id) + _setter("workspace_id", workspace_id) @property @pulumi.getter @@ -82,12 +95,25 @@ def __init__(__self__, *, :param pulumi.Input[int] principal_id: Databricks ID of the user, service principal, or group. The principal ID can be retrieved using the SCIM API, or using databricks_user, ServicePrincipal or Group data sources. :param pulumi.Input[int] workspace_id: Databricks workspace ID. """ + _MwsPermissionAssignmentState._configure( + lambda key, value: pulumi.set(__self__, key, value), + permissions=permissions, + principal_id=principal_id, + workspace_id=workspace_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permissions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + principal_id: Optional[pulumi.Input[int]] = None, + workspace_id: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if permissions is not None: - pulumi.set(__self__, "permissions", permissions) + _setter("permissions", permissions) if principal_id is not None: - pulumi.set(__self__, "principal_id", principal_id) + _setter("principal_id", principal_id) if workspace_id is not None: - pulumi.set(__self__, "workspace_id", workspace_id) + _setter("workspace_id", workspace_id) @property @pulumi.getter @@ -281,6 +307,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsPermissionAssignmentArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mws_private_access_settings.py b/sdk/python/pulumi_databricks/mws_private_access_settings.py index 35aac3ad..8c55d088 100644 --- a/sdk/python/pulumi_databricks/mws_private_access_settings.py +++ b/sdk/python/pulumi_databricks/mws_private_access_settings.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MwsPrivateAccessSettingsArgs', 'MwsPrivateAccessSettings'] @@ -33,20 +33,43 @@ def __init__(__self__, *, :param pulumi.Input[bool] public_access_enabled: If `true`, the MwsWorkspaces can be accessed over the MwsVpcEndpoint as well as over the public network. In such a case, you could also configure an IpAccessList for the workspace, to restrict the source networks that could be used to access it over the public network. If `false`, the workspace can be accessed only over VPC endpoints, and not over the public network. :param pulumi.Input[str] status: (AWS only) Status of Private Access Settings """ - pulumi.set(__self__, "private_access_settings_name", private_access_settings_name) - pulumi.set(__self__, "region", region) + MwsPrivateAccessSettingsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + private_access_settings_name=private_access_settings_name, + region=region, + account_id=account_id, + allowed_vpc_endpoint_ids=allowed_vpc_endpoint_ids, + private_access_level=private_access_level, + private_access_settings_id=private_access_settings_id, + public_access_enabled=public_access_enabled, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + private_access_settings_name: pulumi.Input[str], + region: pulumi.Input[str], + account_id: Optional[pulumi.Input[str]] = None, + allowed_vpc_endpoint_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + private_access_level: Optional[pulumi.Input[str]] = None, + private_access_settings_id: Optional[pulumi.Input[str]] = None, + public_access_enabled: Optional[pulumi.Input[bool]] = None, + status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("private_access_settings_name", private_access_settings_name) + _setter("region", region) if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if allowed_vpc_endpoint_ids is not None: - pulumi.set(__self__, "allowed_vpc_endpoint_ids", allowed_vpc_endpoint_ids) + _setter("allowed_vpc_endpoint_ids", allowed_vpc_endpoint_ids) if private_access_level is not None: - pulumi.set(__self__, "private_access_level", private_access_level) + _setter("private_access_level", private_access_level) if private_access_settings_id is not None: - pulumi.set(__self__, "private_access_settings_id", private_access_settings_id) + _setter("private_access_settings_id", private_access_settings_id) if public_access_enabled is not None: - pulumi.set(__self__, "public_access_enabled", public_access_enabled) + _setter("public_access_enabled", public_access_enabled) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter(name="privateAccessSettingsName") @@ -167,22 +190,45 @@ def __init__(__self__, *, :param pulumi.Input[str] region: Region of AWS VPC or the Google Cloud VPC network :param pulumi.Input[str] status: (AWS only) Status of Private Access Settings """ + _MwsPrivateAccessSettingsState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + allowed_vpc_endpoint_ids=allowed_vpc_endpoint_ids, + private_access_level=private_access_level, + private_access_settings_id=private_access_settings_id, + private_access_settings_name=private_access_settings_name, + public_access_enabled=public_access_enabled, + region=region, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + allowed_vpc_endpoint_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + private_access_level: Optional[pulumi.Input[str]] = None, + private_access_settings_id: Optional[pulumi.Input[str]] = None, + private_access_settings_name: Optional[pulumi.Input[str]] = None, + public_access_enabled: Optional[pulumi.Input[bool]] = None, + region: Optional[pulumi.Input[str]] = None, + status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if allowed_vpc_endpoint_ids is not None: - pulumi.set(__self__, "allowed_vpc_endpoint_ids", allowed_vpc_endpoint_ids) + _setter("allowed_vpc_endpoint_ids", allowed_vpc_endpoint_ids) if private_access_level is not None: - pulumi.set(__self__, "private_access_level", private_access_level) + _setter("private_access_level", private_access_level) if private_access_settings_id is not None: - pulumi.set(__self__, "private_access_settings_id", private_access_settings_id) + _setter("private_access_settings_id", private_access_settings_id) if private_access_settings_name is not None: - pulumi.set(__self__, "private_access_settings_name", private_access_settings_name) + _setter("private_access_settings_name", private_access_settings_name) if public_access_enabled is not None: - pulumi.set(__self__, "public_access_enabled", public_access_enabled) + _setter("public_access_enabled", public_access_enabled) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter(name="accountId") @@ -484,6 +530,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsPrivateAccessSettingsArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mws_storage_configurations.py b/sdk/python/pulumi_databricks/mws_storage_configurations.py index bf6ba6cd..48a37f4b 100644 --- a/sdk/python/pulumi_databricks/mws_storage_configurations.py +++ b/sdk/python/pulumi_databricks/mws_storage_configurations.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['MwsStorageConfigurationsArgs', 'MwsStorageConfigurations'] @@ -23,9 +23,22 @@ def __init__(__self__, *, :param pulumi.Input[str] bucket_name: name of AWS S3 bucket :param pulumi.Input[str] storage_configuration_name: name under which this storage configuration is stored """ - pulumi.set(__self__, "account_id", account_id) - pulumi.set(__self__, "bucket_name", bucket_name) - pulumi.set(__self__, "storage_configuration_name", storage_configuration_name) + MwsStorageConfigurationsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + bucket_name=bucket_name, + storage_configuration_name=storage_configuration_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: pulumi.Input[str], + bucket_name: pulumi.Input[str], + storage_configuration_name: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("account_id", account_id) + _setter("bucket_name", bucket_name) + _setter("storage_configuration_name", storage_configuration_name) @property @pulumi.getter(name="accountId") @@ -79,16 +92,33 @@ def __init__(__self__, *, :param pulumi.Input[str] storage_configuration_id: (String) id of storage config to be used for `databricks_mws_workspace` resource. :param pulumi.Input[str] storage_configuration_name: name under which this storage configuration is stored """ + _MwsStorageConfigurationsState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + bucket_name=bucket_name, + creation_time=creation_time, + storage_configuration_id=storage_configuration_id, + storage_configuration_name=storage_configuration_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + bucket_name: Optional[pulumi.Input[str]] = None, + creation_time: Optional[pulumi.Input[int]] = None, + storage_configuration_id: Optional[pulumi.Input[str]] = None, + storage_configuration_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if bucket_name is not None: - pulumi.set(__self__, "bucket_name", bucket_name) + _setter("bucket_name", bucket_name) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if storage_configuration_id is not None: - pulumi.set(__self__, "storage_configuration_id", storage_configuration_id) + _setter("storage_configuration_id", storage_configuration_id) if storage_configuration_name is not None: - pulumi.set(__self__, "storage_configuration_name", storage_configuration_name) + _setter("storage_configuration_name", storage_configuration_name) @property @pulumi.getter(name="accountId") @@ -255,6 +285,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsStorageConfigurationsArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/mws_vpc_endpoint.py b/sdk/python/pulumi_databricks/mws_vpc_endpoint.py index 7a098f51..4d141ffb 100644 --- a/sdk/python/pulumi_databricks/mws_vpc_endpoint.py +++ b/sdk/python/pulumi_databricks/mws_vpc_endpoint.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -37,25 +37,52 @@ def __init__(__self__, *, :param pulumi.Input[str] state: (AWS Only) State of VPC Endpoint :param pulumi.Input[str] vpc_endpoint_id: Canonical unique identifier of VPC Endpoint in Databricks Account """ - pulumi.set(__self__, "vpc_endpoint_name", vpc_endpoint_name) + MwsVpcEndpointArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + vpc_endpoint_name=vpc_endpoint_name, + account_id=account_id, + aws_account_id=aws_account_id, + aws_endpoint_service_id=aws_endpoint_service_id, + aws_vpc_endpoint_id=aws_vpc_endpoint_id, + gcp_vpc_endpoint_info=gcp_vpc_endpoint_info, + region=region, + state=state, + use_case=use_case, + vpc_endpoint_id=vpc_endpoint_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + vpc_endpoint_name: pulumi.Input[str], + account_id: Optional[pulumi.Input[str]] = None, + aws_account_id: Optional[pulumi.Input[str]] = None, + aws_endpoint_service_id: Optional[pulumi.Input[str]] = None, + aws_vpc_endpoint_id: Optional[pulumi.Input[str]] = None, + gcp_vpc_endpoint_info: Optional[pulumi.Input['MwsVpcEndpointGcpVpcEndpointInfoArgs']] = None, + region: Optional[pulumi.Input[str]] = None, + state: Optional[pulumi.Input[str]] = None, + use_case: Optional[pulumi.Input[str]] = None, + vpc_endpoint_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("vpc_endpoint_name", vpc_endpoint_name) if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if aws_account_id is not None: - pulumi.set(__self__, "aws_account_id", aws_account_id) + _setter("aws_account_id", aws_account_id) if aws_endpoint_service_id is not None: - pulumi.set(__self__, "aws_endpoint_service_id", aws_endpoint_service_id) + _setter("aws_endpoint_service_id", aws_endpoint_service_id) if aws_vpc_endpoint_id is not None: - pulumi.set(__self__, "aws_vpc_endpoint_id", aws_vpc_endpoint_id) + _setter("aws_vpc_endpoint_id", aws_vpc_endpoint_id) if gcp_vpc_endpoint_info is not None: - pulumi.set(__self__, "gcp_vpc_endpoint_info", gcp_vpc_endpoint_info) + _setter("gcp_vpc_endpoint_info", gcp_vpc_endpoint_info) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) if state is not None: - pulumi.set(__self__, "state", state) + _setter("state", state) if use_case is not None: - pulumi.set(__self__, "use_case", use_case) + _setter("use_case", use_case) if vpc_endpoint_id is not None: - pulumi.set(__self__, "vpc_endpoint_id", vpc_endpoint_id) + _setter("vpc_endpoint_id", vpc_endpoint_id) @property @pulumi.getter(name="vpcEndpointName") @@ -196,26 +223,53 @@ def __init__(__self__, *, :param pulumi.Input[str] vpc_endpoint_id: Canonical unique identifier of VPC Endpoint in Databricks Account :param pulumi.Input[str] vpc_endpoint_name: Name of VPC Endpoint in Databricks Account """ + _MwsVpcEndpointState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + aws_account_id=aws_account_id, + aws_endpoint_service_id=aws_endpoint_service_id, + aws_vpc_endpoint_id=aws_vpc_endpoint_id, + gcp_vpc_endpoint_info=gcp_vpc_endpoint_info, + region=region, + state=state, + use_case=use_case, + vpc_endpoint_id=vpc_endpoint_id, + vpc_endpoint_name=vpc_endpoint_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + aws_account_id: Optional[pulumi.Input[str]] = None, + aws_endpoint_service_id: Optional[pulumi.Input[str]] = None, + aws_vpc_endpoint_id: Optional[pulumi.Input[str]] = None, + gcp_vpc_endpoint_info: Optional[pulumi.Input['MwsVpcEndpointGcpVpcEndpointInfoArgs']] = None, + region: Optional[pulumi.Input[str]] = None, + state: Optional[pulumi.Input[str]] = None, + use_case: Optional[pulumi.Input[str]] = None, + vpc_endpoint_id: Optional[pulumi.Input[str]] = None, + vpc_endpoint_name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if aws_account_id is not None: - pulumi.set(__self__, "aws_account_id", aws_account_id) + _setter("aws_account_id", aws_account_id) if aws_endpoint_service_id is not None: - pulumi.set(__self__, "aws_endpoint_service_id", aws_endpoint_service_id) + _setter("aws_endpoint_service_id", aws_endpoint_service_id) if aws_vpc_endpoint_id is not None: - pulumi.set(__self__, "aws_vpc_endpoint_id", aws_vpc_endpoint_id) + _setter("aws_vpc_endpoint_id", aws_vpc_endpoint_id) if gcp_vpc_endpoint_info is not None: - pulumi.set(__self__, "gcp_vpc_endpoint_info", gcp_vpc_endpoint_info) + _setter("gcp_vpc_endpoint_info", gcp_vpc_endpoint_info) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) if state is not None: - pulumi.set(__self__, "state", state) + _setter("state", state) if use_case is not None: - pulumi.set(__self__, "use_case", use_case) + _setter("use_case", use_case) if vpc_endpoint_id is not None: - pulumi.set(__self__, "vpc_endpoint_id", vpc_endpoint_id) + _setter("vpc_endpoint_id", vpc_endpoint_id) if vpc_endpoint_name is not None: - pulumi.set(__self__, "vpc_endpoint_name", vpc_endpoint_name) + _setter("vpc_endpoint_name", vpc_endpoint_name) @property @pulumi.getter(name="accountId") @@ -385,6 +439,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsVpcEndpointArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -413,6 +471,11 @@ def _internal_init(__self__, __props__.__dict__["aws_account_id"] = aws_account_id __props__.__dict__["aws_endpoint_service_id"] = aws_endpoint_service_id __props__.__dict__["aws_vpc_endpoint_id"] = aws_vpc_endpoint_id + if gcp_vpc_endpoint_info is not None and not isinstance(gcp_vpc_endpoint_info, MwsVpcEndpointGcpVpcEndpointInfoArgs): + gcp_vpc_endpoint_info = gcp_vpc_endpoint_info or {} + def _setter(key, value): + gcp_vpc_endpoint_info[key] = value + MwsVpcEndpointGcpVpcEndpointInfoArgs._configure(_setter, **gcp_vpc_endpoint_info) __props__.__dict__["gcp_vpc_endpoint_info"] = gcp_vpc_endpoint_info __props__.__dict__["region"] = region __props__.__dict__["state"] = state diff --git a/sdk/python/pulumi_databricks/mws_workspaces.py b/sdk/python/pulumi_databricks/mws_workspaces.py index 31b2861d..dc8e09d1 100644 --- a/sdk/python/pulumi_databricks/mws_workspaces.py +++ b/sdk/python/pulumi_databricks/mws_workspaces.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -61,57 +61,114 @@ def __init__(__self__, *, :param pulumi.Input[str] workspace_status_message: (String) updates on workspace status :param pulumi.Input[str] workspace_url: (String) URL of the workspace """ - pulumi.set(__self__, "account_id", account_id) - pulumi.set(__self__, "workspace_name", workspace_name) + MwsWorkspacesArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + workspace_name=workspace_name, + aws_region=aws_region, + cloud=cloud, + cloud_resource_container=cloud_resource_container, + creation_time=creation_time, + credentials_id=credentials_id, + customer_managed_key_id=customer_managed_key_id, + deployment_name=deployment_name, + external_customer_info=external_customer_info, + gcp_managed_network_config=gcp_managed_network_config, + gke_config=gke_config, + is_no_public_ip_enabled=is_no_public_ip_enabled, + location=location, + managed_services_customer_managed_key_id=managed_services_customer_managed_key_id, + network_id=network_id, + pricing_tier=pricing_tier, + private_access_settings_id=private_access_settings_id, + storage_configuration_id=storage_configuration_id, + storage_customer_managed_key_id=storage_customer_managed_key_id, + token=token, + workspace_id=workspace_id, + workspace_status=workspace_status, + workspace_status_message=workspace_status_message, + workspace_url=workspace_url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: pulumi.Input[str], + workspace_name: pulumi.Input[str], + aws_region: Optional[pulumi.Input[str]] = None, + cloud: Optional[pulumi.Input[str]] = None, + cloud_resource_container: Optional[pulumi.Input['MwsWorkspacesCloudResourceContainerArgs']] = None, + creation_time: Optional[pulumi.Input[int]] = None, + credentials_id: Optional[pulumi.Input[str]] = None, + customer_managed_key_id: Optional[pulumi.Input[str]] = None, + deployment_name: Optional[pulumi.Input[str]] = None, + external_customer_info: Optional[pulumi.Input['MwsWorkspacesExternalCustomerInfoArgs']] = None, + gcp_managed_network_config: Optional[pulumi.Input['MwsWorkspacesGcpManagedNetworkConfigArgs']] = None, + gke_config: Optional[pulumi.Input['MwsWorkspacesGkeConfigArgs']] = None, + is_no_public_ip_enabled: Optional[pulumi.Input[bool]] = None, + location: Optional[pulumi.Input[str]] = None, + managed_services_customer_managed_key_id: Optional[pulumi.Input[str]] = None, + network_id: Optional[pulumi.Input[str]] = None, + pricing_tier: Optional[pulumi.Input[str]] = None, + private_access_settings_id: Optional[pulumi.Input[str]] = None, + storage_configuration_id: Optional[pulumi.Input[str]] = None, + storage_customer_managed_key_id: Optional[pulumi.Input[str]] = None, + token: Optional[pulumi.Input['MwsWorkspacesTokenArgs']] = None, + workspace_id: Optional[pulumi.Input[int]] = None, + workspace_status: Optional[pulumi.Input[str]] = None, + workspace_status_message: Optional[pulumi.Input[str]] = None, + workspace_url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("account_id", account_id) + _setter("workspace_name", workspace_name) if aws_region is not None: - pulumi.set(__self__, "aws_region", aws_region) + _setter("aws_region", aws_region) if cloud is not None: - pulumi.set(__self__, "cloud", cloud) + _setter("cloud", cloud) if cloud_resource_container is not None: - pulumi.set(__self__, "cloud_resource_container", cloud_resource_container) + _setter("cloud_resource_container", cloud_resource_container) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if credentials_id is not None: - pulumi.set(__self__, "credentials_id", credentials_id) + _setter("credentials_id", credentials_id) if customer_managed_key_id is not None: warnings.warn("""Use managed_services_customer_managed_key_id instead""", DeprecationWarning) pulumi.log.warn("""customer_managed_key_id is deprecated: Use managed_services_customer_managed_key_id instead""") if customer_managed_key_id is not None: - pulumi.set(__self__, "customer_managed_key_id", customer_managed_key_id) + _setter("customer_managed_key_id", customer_managed_key_id) if deployment_name is not None: - pulumi.set(__self__, "deployment_name", deployment_name) + _setter("deployment_name", deployment_name) if external_customer_info is not None: - pulumi.set(__self__, "external_customer_info", external_customer_info) + _setter("external_customer_info", external_customer_info) if gcp_managed_network_config is not None: - pulumi.set(__self__, "gcp_managed_network_config", gcp_managed_network_config) + _setter("gcp_managed_network_config", gcp_managed_network_config) if gke_config is not None: - pulumi.set(__self__, "gke_config", gke_config) + _setter("gke_config", gke_config) if is_no_public_ip_enabled is not None: - pulumi.set(__self__, "is_no_public_ip_enabled", is_no_public_ip_enabled) + _setter("is_no_public_ip_enabled", is_no_public_ip_enabled) if location is not None: - pulumi.set(__self__, "location", location) + _setter("location", location) if managed_services_customer_managed_key_id is not None: - pulumi.set(__self__, "managed_services_customer_managed_key_id", managed_services_customer_managed_key_id) + _setter("managed_services_customer_managed_key_id", managed_services_customer_managed_key_id) if network_id is not None: - pulumi.set(__self__, "network_id", network_id) + _setter("network_id", network_id) if pricing_tier is not None: - pulumi.set(__self__, "pricing_tier", pricing_tier) + _setter("pricing_tier", pricing_tier) if private_access_settings_id is not None: - pulumi.set(__self__, "private_access_settings_id", private_access_settings_id) + _setter("private_access_settings_id", private_access_settings_id) if storage_configuration_id is not None: - pulumi.set(__self__, "storage_configuration_id", storage_configuration_id) + _setter("storage_configuration_id", storage_configuration_id) if storage_customer_managed_key_id is not None: - pulumi.set(__self__, "storage_customer_managed_key_id", storage_customer_managed_key_id) + _setter("storage_customer_managed_key_id", storage_customer_managed_key_id) if token is not None: - pulumi.set(__self__, "token", token) + _setter("token", token) if workspace_id is not None: - pulumi.set(__self__, "workspace_id", workspace_id) + _setter("workspace_id", workspace_id) if workspace_status is not None: - pulumi.set(__self__, "workspace_status", workspace_status) + _setter("workspace_status", workspace_status) if workspace_status_message is not None: - pulumi.set(__self__, "workspace_status_message", workspace_status_message) + _setter("workspace_status_message", workspace_status_message) if workspace_url is not None: - pulumi.set(__self__, "workspace_url", workspace_url) + _setter("workspace_url", workspace_url) @property @pulumi.getter(name="accountId") @@ -441,59 +498,116 @@ def __init__(__self__, *, :param pulumi.Input[str] workspace_status_message: (String) updates on workspace status :param pulumi.Input[str] workspace_url: (String) URL of the workspace """ + _MwsWorkspacesState._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + aws_region=aws_region, + cloud=cloud, + cloud_resource_container=cloud_resource_container, + creation_time=creation_time, + credentials_id=credentials_id, + customer_managed_key_id=customer_managed_key_id, + deployment_name=deployment_name, + external_customer_info=external_customer_info, + gcp_managed_network_config=gcp_managed_network_config, + gke_config=gke_config, + is_no_public_ip_enabled=is_no_public_ip_enabled, + location=location, + managed_services_customer_managed_key_id=managed_services_customer_managed_key_id, + network_id=network_id, + pricing_tier=pricing_tier, + private_access_settings_id=private_access_settings_id, + storage_configuration_id=storage_configuration_id, + storage_customer_managed_key_id=storage_customer_managed_key_id, + token=token, + workspace_id=workspace_id, + workspace_name=workspace_name, + workspace_status=workspace_status, + workspace_status_message=workspace_status_message, + workspace_url=workspace_url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + aws_region: Optional[pulumi.Input[str]] = None, + cloud: Optional[pulumi.Input[str]] = None, + cloud_resource_container: Optional[pulumi.Input['MwsWorkspacesCloudResourceContainerArgs']] = None, + creation_time: Optional[pulumi.Input[int]] = None, + credentials_id: Optional[pulumi.Input[str]] = None, + customer_managed_key_id: Optional[pulumi.Input[str]] = None, + deployment_name: Optional[pulumi.Input[str]] = None, + external_customer_info: Optional[pulumi.Input['MwsWorkspacesExternalCustomerInfoArgs']] = None, + gcp_managed_network_config: Optional[pulumi.Input['MwsWorkspacesGcpManagedNetworkConfigArgs']] = None, + gke_config: Optional[pulumi.Input['MwsWorkspacesGkeConfigArgs']] = None, + is_no_public_ip_enabled: Optional[pulumi.Input[bool]] = None, + location: Optional[pulumi.Input[str]] = None, + managed_services_customer_managed_key_id: Optional[pulumi.Input[str]] = None, + network_id: Optional[pulumi.Input[str]] = None, + pricing_tier: Optional[pulumi.Input[str]] = None, + private_access_settings_id: Optional[pulumi.Input[str]] = None, + storage_configuration_id: Optional[pulumi.Input[str]] = None, + storage_customer_managed_key_id: Optional[pulumi.Input[str]] = None, + token: Optional[pulumi.Input['MwsWorkspacesTokenArgs']] = None, + workspace_id: Optional[pulumi.Input[int]] = None, + workspace_name: Optional[pulumi.Input[str]] = None, + workspace_status: Optional[pulumi.Input[str]] = None, + workspace_status_message: Optional[pulumi.Input[str]] = None, + workspace_url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if aws_region is not None: - pulumi.set(__self__, "aws_region", aws_region) + _setter("aws_region", aws_region) if cloud is not None: - pulumi.set(__self__, "cloud", cloud) + _setter("cloud", cloud) if cloud_resource_container is not None: - pulumi.set(__self__, "cloud_resource_container", cloud_resource_container) + _setter("cloud_resource_container", cloud_resource_container) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if credentials_id is not None: - pulumi.set(__self__, "credentials_id", credentials_id) + _setter("credentials_id", credentials_id) if customer_managed_key_id is not None: warnings.warn("""Use managed_services_customer_managed_key_id instead""", DeprecationWarning) pulumi.log.warn("""customer_managed_key_id is deprecated: Use managed_services_customer_managed_key_id instead""") if customer_managed_key_id is not None: - pulumi.set(__self__, "customer_managed_key_id", customer_managed_key_id) + _setter("customer_managed_key_id", customer_managed_key_id) if deployment_name is not None: - pulumi.set(__self__, "deployment_name", deployment_name) + _setter("deployment_name", deployment_name) if external_customer_info is not None: - pulumi.set(__self__, "external_customer_info", external_customer_info) + _setter("external_customer_info", external_customer_info) if gcp_managed_network_config is not None: - pulumi.set(__self__, "gcp_managed_network_config", gcp_managed_network_config) + _setter("gcp_managed_network_config", gcp_managed_network_config) if gke_config is not None: - pulumi.set(__self__, "gke_config", gke_config) + _setter("gke_config", gke_config) if is_no_public_ip_enabled is not None: - pulumi.set(__self__, "is_no_public_ip_enabled", is_no_public_ip_enabled) + _setter("is_no_public_ip_enabled", is_no_public_ip_enabled) if location is not None: - pulumi.set(__self__, "location", location) + _setter("location", location) if managed_services_customer_managed_key_id is not None: - pulumi.set(__self__, "managed_services_customer_managed_key_id", managed_services_customer_managed_key_id) + _setter("managed_services_customer_managed_key_id", managed_services_customer_managed_key_id) if network_id is not None: - pulumi.set(__self__, "network_id", network_id) + _setter("network_id", network_id) if pricing_tier is not None: - pulumi.set(__self__, "pricing_tier", pricing_tier) + _setter("pricing_tier", pricing_tier) if private_access_settings_id is not None: - pulumi.set(__self__, "private_access_settings_id", private_access_settings_id) + _setter("private_access_settings_id", private_access_settings_id) if storage_configuration_id is not None: - pulumi.set(__self__, "storage_configuration_id", storage_configuration_id) + _setter("storage_configuration_id", storage_configuration_id) if storage_customer_managed_key_id is not None: - pulumi.set(__self__, "storage_customer_managed_key_id", storage_customer_managed_key_id) + _setter("storage_customer_managed_key_id", storage_customer_managed_key_id) if token is not None: - pulumi.set(__self__, "token", token) + _setter("token", token) if workspace_id is not None: - pulumi.set(__self__, "workspace_id", workspace_id) + _setter("workspace_id", workspace_id) if workspace_name is not None: - pulumi.set(__self__, "workspace_name", workspace_name) + _setter("workspace_name", workspace_name) if workspace_status is not None: - pulumi.set(__self__, "workspace_status", workspace_status) + _setter("workspace_status", workspace_status) if workspace_status_message is not None: - pulumi.set(__self__, "workspace_status_message", workspace_status_message) + _setter("workspace_status_message", workspace_status_message) if workspace_url is not None: - pulumi.set(__self__, "workspace_url", workspace_url) + _setter("workspace_url", workspace_url) @property @pulumi.getter(name="accountId") @@ -852,6 +966,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + MwsWorkspacesArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -896,16 +1014,33 @@ def _internal_init(__self__, __props__.__dict__["account_id"] = None if account_id is None else pulumi.Output.secret(account_id) __props__.__dict__["aws_region"] = aws_region __props__.__dict__["cloud"] = cloud + if cloud_resource_container is not None and not isinstance(cloud_resource_container, MwsWorkspacesCloudResourceContainerArgs): + cloud_resource_container = cloud_resource_container or {} + def _setter(key, value): + cloud_resource_container[key] = value + MwsWorkspacesCloudResourceContainerArgs._configure(_setter, **cloud_resource_container) __props__.__dict__["cloud_resource_container"] = cloud_resource_container __props__.__dict__["creation_time"] = creation_time __props__.__dict__["credentials_id"] = credentials_id - if customer_managed_key_id is not None and not opts.urn: - warnings.warn("""Use managed_services_customer_managed_key_id instead""", DeprecationWarning) - pulumi.log.warn("""customer_managed_key_id is deprecated: Use managed_services_customer_managed_key_id instead""") __props__.__dict__["customer_managed_key_id"] = customer_managed_key_id __props__.__dict__["deployment_name"] = deployment_name + if external_customer_info is not None and not isinstance(external_customer_info, MwsWorkspacesExternalCustomerInfoArgs): + external_customer_info = external_customer_info or {} + def _setter(key, value): + external_customer_info[key] = value + MwsWorkspacesExternalCustomerInfoArgs._configure(_setter, **external_customer_info) __props__.__dict__["external_customer_info"] = external_customer_info + if gcp_managed_network_config is not None and not isinstance(gcp_managed_network_config, MwsWorkspacesGcpManagedNetworkConfigArgs): + gcp_managed_network_config = gcp_managed_network_config or {} + def _setter(key, value): + gcp_managed_network_config[key] = value + MwsWorkspacesGcpManagedNetworkConfigArgs._configure(_setter, **gcp_managed_network_config) __props__.__dict__["gcp_managed_network_config"] = gcp_managed_network_config + if gke_config is not None and not isinstance(gke_config, MwsWorkspacesGkeConfigArgs): + gke_config = gke_config or {} + def _setter(key, value): + gke_config[key] = value + MwsWorkspacesGkeConfigArgs._configure(_setter, **gke_config) __props__.__dict__["gke_config"] = gke_config __props__.__dict__["is_no_public_ip_enabled"] = is_no_public_ip_enabled __props__.__dict__["location"] = location @@ -915,6 +1050,11 @@ def _internal_init(__self__, __props__.__dict__["private_access_settings_id"] = private_access_settings_id __props__.__dict__["storage_configuration_id"] = storage_configuration_id __props__.__dict__["storage_customer_managed_key_id"] = storage_customer_managed_key_id + if token is not None and not isinstance(token, MwsWorkspacesTokenArgs): + token = token or {} + def _setter(key, value): + token[key] = value + MwsWorkspacesTokenArgs._configure(_setter, **token) __props__.__dict__["token"] = token __props__.__dict__["workspace_id"] = workspace_id if workspace_name is None and not opts.urn: diff --git a/sdk/python/pulumi_databricks/notebook.py b/sdk/python/pulumi_databricks/notebook.py index 28bd45c3..00a36ca4 100644 --- a/sdk/python/pulumi_databricks/notebook.py +++ b/sdk/python/pulumi_databricks/notebook.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['NotebookArgs', 'Notebook'] @@ -29,24 +29,47 @@ def __init__(__self__, *, :param pulumi.Input[int] object_id: Unique identifier for a NOTEBOOK :param pulumi.Input[str] source: Path to notebook in source code format on local filesystem. Conflicts with `content_base64`. """ - pulumi.set(__self__, "path", path) + NotebookArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + content_base64=content_base64, + format=format, + language=language, + md5=md5, + object_id=object_id, + object_type=object_type, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + content_base64: Optional[pulumi.Input[str]] = None, + format: Optional[pulumi.Input[str]] = None, + language: Optional[pulumi.Input[str]] = None, + md5: Optional[pulumi.Input[str]] = None, + object_id: Optional[pulumi.Input[int]] = None, + object_type: Optional[pulumi.Input[str]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if format is not None: - pulumi.set(__self__, "format", format) + _setter("format", format) if language is not None: - pulumi.set(__self__, "language", language) + _setter("language", language) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if object_id is not None: - pulumi.set(__self__, "object_id", object_id) + _setter("object_id", object_id) if object_type is not None: warnings.warn("""Always is a notebook""", DeprecationWarning) pulumi.log.warn("""object_type is deprecated: Always is a notebook""") if object_type is not None: - pulumi.set(__self__, "object_type", object_type) + _setter("object_type", object_type) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter @@ -156,27 +179,52 @@ def __init__(__self__, *, :param pulumi.Input[str] source: Path to notebook in source code format on local filesystem. Conflicts with `content_base64`. :param pulumi.Input[str] url: Routable URL of the notebook """ + _NotebookState._configure( + lambda key, value: pulumi.set(__self__, key, value), + content_base64=content_base64, + format=format, + language=language, + md5=md5, + object_id=object_id, + object_type=object_type, + path=path, + source=source, + url=url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + content_base64: Optional[pulumi.Input[str]] = None, + format: Optional[pulumi.Input[str]] = None, + language: Optional[pulumi.Input[str]] = None, + md5: Optional[pulumi.Input[str]] = None, + object_id: Optional[pulumi.Input[int]] = None, + object_type: Optional[pulumi.Input[str]] = None, + path: Optional[pulumi.Input[str]] = None, + source: Optional[pulumi.Input[str]] = None, + url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if format is not None: - pulumi.set(__self__, "format", format) + _setter("format", format) if language is not None: - pulumi.set(__self__, "language", language) + _setter("language", language) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if object_id is not None: - pulumi.set(__self__, "object_id", object_id) + _setter("object_id", object_id) if object_type is not None: warnings.warn("""Always is a notebook""", DeprecationWarning) pulumi.log.warn("""object_type is deprecated: Always is a notebook""") if object_type is not None: - pulumi.set(__self__, "object_type", object_type) + _setter("object_type", object_type) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) @property @pulumi.getter(name="contentBase64") @@ -333,6 +381,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + NotebookArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -360,9 +412,6 @@ def _internal_init(__self__, __props__.__dict__["language"] = language __props__.__dict__["md5"] = md5 __props__.__dict__["object_id"] = object_id - if object_type is not None and not opts.urn: - warnings.warn("""Always is a notebook""", DeprecationWarning) - pulumi.log.warn("""object_type is deprecated: Always is a notebook""") __props__.__dict__["object_type"] = object_type if path is None and not opts.urn: raise TypeError("Missing required property 'path'") diff --git a/sdk/python/pulumi_databricks/obo_token.py b/sdk/python/pulumi_databricks/obo_token.py index 6fabba07..4e18b8c5 100644 --- a/sdk/python/pulumi_databricks/obo_token.py +++ b/sdk/python/pulumi_databricks/obo_token.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['OboTokenArgs', 'OboToken'] @@ -23,11 +23,24 @@ def __init__(__self__, *, :param pulumi.Input[str] comment: Comment that describes the purpose of the token. :param pulumi.Input[int] lifetime_seconds: The number of seconds before the token expires. Token resource is re-created when it expires. If no lifetime is specified, the token remains valid indefinitely. """ - pulumi.set(__self__, "application_id", application_id) + OboTokenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + application_id=application_id, + comment=comment, + lifetime_seconds=lifetime_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + application_id: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + lifetime_seconds: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("application_id", application_id) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if lifetime_seconds is not None: - pulumi.set(__self__, "lifetime_seconds", lifetime_seconds) + _setter("lifetime_seconds", lifetime_seconds) @property @pulumi.getter(name="applicationId") @@ -80,14 +93,29 @@ def __init__(__self__, *, :param pulumi.Input[int] lifetime_seconds: The number of seconds before the token expires. Token resource is re-created when it expires. If no lifetime is specified, the token remains valid indefinitely. :param pulumi.Input[str] token_value: **Sensitive** value of the newly-created token. """ + _OboTokenState._configure( + lambda key, value: pulumi.set(__self__, key, value), + application_id=application_id, + comment=comment, + lifetime_seconds=lifetime_seconds, + token_value=token_value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + application_id: Optional[pulumi.Input[str]] = None, + comment: Optional[pulumi.Input[str]] = None, + lifetime_seconds: Optional[pulumi.Input[int]] = None, + token_value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if application_id is not None: - pulumi.set(__self__, "application_id", application_id) + _setter("application_id", application_id) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if lifetime_seconds is not None: - pulumi.set(__self__, "lifetime_seconds", lifetime_seconds) + _setter("lifetime_seconds", lifetime_seconds) if token_value is not None: - pulumi.set(__self__, "token_value", token_value) + _setter("token_value", token_value) @property @pulumi.getter(name="applicationId") @@ -179,6 +207,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + OboTokenArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/outputs.py b/sdk/python/pulumi_databricks/outputs.py index 3e15c949..6f841d13 100644 --- a/sdk/python/pulumi_databricks/outputs.py +++ b/sdk/python/pulumi_databricks/outputs.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs @@ -466,9 +466,20 @@ def __init__(__self__, *, * `groups/{groupname}` (also exposed as `acl_principal_id` attribute of `Group` resource). * `servicePrincipals/{applicationId}` (also exposed as `acl_principal_id` attribute of `ServicePrincipal` resource). """ - pulumi.set(__self__, "role", role) + AccessControlRuleSetGrantRule._configure( + lambda key, value: pulumi.set(__self__, key, value), + role=role, + principals=principals, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role: str, + principals: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role", role) if principals is not None: - pulumi.set(__self__, "principals", principals) + _setter("principals", principals) @property @pulumi.getter @@ -518,10 +529,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + ClusterAutoscale._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -574,22 +596,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + ClusterAwsAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -657,12 +702,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + ClusterAzureAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -685,10 +743,21 @@ class ClusterClusterLogConf(dict): def __init__(__self__, *, dbfs: Optional['outputs.ClusterClusterLogConfDbfs'] = None, s3: Optional['outputs.ClusterClusterLogConfS3'] = None): + ClusterClusterLogConf._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.ClusterClusterLogConfDbfs'] = None, + s3: Optional['outputs.ClusterClusterLogConfS3'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -705,7 +774,16 @@ def s3(self) -> Optional['outputs.ClusterClusterLogConfS3']: class ClusterClusterLogConfDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + ClusterClusterLogConfDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -746,19 +824,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + ClusterClusterLogConfS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -823,10 +922,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'outputs.ClusterClusterMountInfoNetworkFilesystemInfo', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + ClusterClusterMountInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'outputs.ClusterClusterMountInfoNetworkFilesystemInfo', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -868,9 +980,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + ClusterClusterMountInfoNetworkFilesystemInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -905,9 +1028,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, url: str, basic_auth: Optional['outputs.ClusterDockerImageBasicAuth'] = None): - pulumi.set(__self__, "url", url) + ClusterDockerImage._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.ClusterDockerImageBasicAuth'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -925,8 +1059,19 @@ class ClusterDockerImageBasicAuth(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + ClusterDockerImageBasicAuth._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -973,18 +1118,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + ClusterGcpAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -1030,20 +1194,41 @@ def __init__(__self__, *, s3: Optional['outputs.ClusterInitScriptS3'] = None, volumes: Optional['outputs.ClusterInitScriptVolumes'] = None, workspace: Optional['outputs.ClusterInitScriptWorkspace'] = None): + ClusterInitScript._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.ClusterInitScriptAbfss'] = None, + dbfs: Optional['outputs.ClusterInitScriptDbfs'] = None, + file: Optional['outputs.ClusterInitScriptFile'] = None, + gcs: Optional['outputs.ClusterInitScriptGcs'] = None, + s3: Optional['outputs.ClusterInitScriptS3'] = None, + volumes: Optional['outputs.ClusterInitScriptVolumes'] = None, + workspace: Optional['outputs.ClusterInitScriptWorkspace'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -1088,8 +1273,17 @@ def workspace(self) -> Optional['outputs.ClusterInitScriptWorkspace']: class ClusterInitScriptAbfss(dict): def __init__(__self__, *, destination: Optional[str] = None): + ClusterInitScriptAbfss._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1101,7 +1295,16 @@ def destination(self) -> Optional[str]: class ClusterInitScriptDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + ClusterInitScriptDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -1113,8 +1316,17 @@ def destination(self) -> str: class ClusterInitScriptFile(dict): def __init__(__self__, *, destination: Optional[str] = None): + ClusterInitScriptFile._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1126,8 +1338,17 @@ def destination(self) -> Optional[str]: class ClusterInitScriptGcs(dict): def __init__(__self__, *, destination: Optional[str] = None): + ClusterInitScriptGcs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1168,19 +1389,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + ClusterInitScriptS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -1222,8 +1464,17 @@ def region(self) -> Optional[str]: class ClusterInitScriptVolumes(dict): def __init__(__self__, *, destination: Optional[str] = None): + ClusterInitScriptVolumes._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1235,8 +1486,17 @@ def destination(self) -> Optional[str]: class ClusterInitScriptWorkspace(dict): def __init__(__self__, *, destination: Optional[str] = None): + ClusterInitScriptWorkspace._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -1253,18 +1513,37 @@ def __init__(__self__, *, maven: Optional['outputs.ClusterLibraryMaven'] = None, pypi: Optional['outputs.ClusterLibraryPypi'] = None, whl: Optional[str] = None): + ClusterLibrary._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional['outputs.ClusterLibraryCran'] = None, + egg: Optional[str] = None, + jar: Optional[str] = None, + maven: Optional['outputs.ClusterLibraryMaven'] = None, + pypi: Optional['outputs.ClusterLibraryPypi'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -1302,9 +1581,20 @@ class ClusterLibraryCran(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + ClusterLibraryCran._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -1323,11 +1613,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + ClusterLibraryMaven._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -1350,9 +1653,20 @@ class ClusterLibraryPypi(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + ClusterLibraryPypi._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -1369,7 +1683,16 @@ def repo(self) -> Optional[str]: class ClusterWorkloadType(dict): def __init__(__self__, *, clients: 'outputs.ClusterWorkloadTypeClients'): - pulumi.set(__self__, "clients", clients) + ClusterWorkloadType._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'outputs.ClusterWorkloadTypeClients', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -1382,10 +1705,21 @@ class ClusterWorkloadTypeClients(dict): def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + ClusterWorkloadTypeClients._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -1419,8 +1753,17 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, sse_encryption_details: Optional['outputs.ExternalLocationEncryptionDetailsSseEncryptionDetails'] = None): + ExternalLocationEncryptionDetails._configure( + lambda key, value: pulumi.set(__self__, key, value), + sse_encryption_details=sse_encryption_details, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + sse_encryption_details: Optional['outputs.ExternalLocationEncryptionDetailsSseEncryptionDetails'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if sse_encryption_details is not None: - pulumi.set(__self__, "sse_encryption_details", sse_encryption_details) + _setter("sse_encryption_details", sse_encryption_details) @property @pulumi.getter(name="sseEncryptionDetails") @@ -1450,10 +1793,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, algorithm: Optional[str] = None, aws_kms_key_arn: Optional[str] = None): + ExternalLocationEncryptionDetailsSseEncryptionDetails._configure( + lambda key, value: pulumi.set(__self__, key, value), + algorithm=algorithm, + aws_kms_key_arn=aws_kms_key_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + algorithm: Optional[str] = None, + aws_kms_key_arn: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if algorithm is not None: - pulumi.set(__self__, "algorithm", algorithm) + _setter("algorithm", algorithm) if aws_kms_key_arn is not None: - pulumi.set(__self__, "aws_kms_key_arn", aws_kms_key_arn) + _setter("aws_kms_key_arn", aws_kms_key_arn) @property @pulumi.getter @@ -1471,8 +1825,19 @@ class GrantsGrant(dict): def __init__(__self__, *, principal: str, privileges: Sequence[str]): - pulumi.set(__self__, "principal", principal) - pulumi.set(__self__, "privileges", privileges) + GrantsGrant._configure( + lambda key, value: pulumi.set(__self__, key, value), + principal=principal, + privileges=privileges, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + principal: str, + privileges: Sequence[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("principal", principal) + _setter("privileges", privileges) @property @pulumi.getter @@ -1515,12 +1880,25 @@ def __init__(__self__, *, :param int spot_bid_price_percent: (Integer) The max price for AWS spot instances, as a percentage of the corresponding instance type’s on-demand price. For example, if this field is set to 50, and the instance pool needs a new i3.xlarge spot instance, then the max price is half of the price of on-demand i3.xlarge instances. Similarly, if this field is set to 200, the max price is twice the price of on-demand i3.xlarge instances. If not specified, the *default value is 100*. When spot instances are requested for this instance pool, only spot instances whose max price percentage matches this field are considered. *For safety, this field cannot be greater than 10000.* :param str zone_id: (String) Identifier for the availability zone/datacenter in which the instance pool resides. This string is of the form like `"us-west-2a"`. The provided availability zone must be in the same region as the Databricks deployment. For example, `"us-west-2a"` is not a valid zone ID if the Databricks deployment resides in the `"us-east-1"` region. If not specified, a default zone is used. You can find the list of available zones as well as the default value by using the [List Zones API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistavailablezones). """ + InstancePoolAwsAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -1573,10 +1951,21 @@ def __init__(__self__, *, :param str availability: Availability type used for all nodes. Valid values are `SPOT_AZURE` and `ON_DEMAND_AZURE`. :param float spot_bid_max_price: The max price for Azure spot instances. Use `-1` to specify the lowest price. """ + InstancePoolAzureAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -1626,12 +2015,25 @@ def __init__(__self__, *, :param int disk_count: (Integer) The number of disks to attach to each instance. This feature is only enabled for supported node types. Users can choose up to the limit of the disks supported by the node type. For node types with no local disk, at least one disk needs to be specified. :param int disk_size: (Integer) The size of each disk (in GiB) to attach. """ + InstancePoolDiskSpec._configure( + lambda key, value: pulumi.set(__self__, key, value), + disk_count=disk_count, + disk_size=disk_size, + disk_type=disk_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + disk_count: Optional[int] = None, + disk_size: Optional[int] = None, + disk_type: Optional['outputs.InstancePoolDiskSpecDiskType'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if disk_count is not None: - pulumi.set(__self__, "disk_count", disk_count) + _setter("disk_count", disk_count) if disk_size is not None: - pulumi.set(__self__, "disk_size", disk_size) + _setter("disk_size", disk_size) if disk_type is not None: - pulumi.set(__self__, "disk_type", disk_type) + _setter("disk_type", disk_type) @property @pulumi.getter(name="diskCount") @@ -1679,10 +2081,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, azure_disk_volume_type: Optional[str] = None, ebs_volume_type: Optional[str] = None): + InstancePoolDiskSpecDiskType._configure( + lambda key, value: pulumi.set(__self__, key, value), + azure_disk_volume_type=azure_disk_volume_type, + ebs_volume_type=ebs_volume_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + azure_disk_volume_type: Optional[str] = None, + ebs_volume_type: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if azure_disk_volume_type is not None: - pulumi.set(__self__, "azure_disk_volume_type", azure_disk_volume_type) + _setter("azure_disk_volume_type", azure_disk_volume_type) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) @property @pulumi.getter(name="azureDiskVolumeType") @@ -1723,10 +2136,21 @@ def __init__(__self__, *, :param str gcp_availability: Availability type used for all nodes. Valid values are `PREEMPTIBLE_GCP`, `PREEMPTIBLE_WITH_FALLBACK_GCP` and `ON_DEMAND_GCP`, default: `ON_DEMAND_GCP`. :param int local_ssd_count: Number of local SSD disks (each is 375GB in size) that will be attached to each node of the cluster. """ + InstancePoolGcpAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + gcp_availability=gcp_availability, + local_ssd_count=local_ssd_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gcp_availability: Optional[str] = None, + local_ssd_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if gcp_availability is not None: - pulumi.set(__self__, "gcp_availability", gcp_availability) + _setter("gcp_availability", gcp_availability) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) @property @pulumi.getter(name="gcpAvailability") @@ -1772,11 +2196,24 @@ def __init__(__self__, *, launch_template_overrides: Sequence['outputs.InstancePoolInstancePoolFleetAttributesLaunchTemplateOverride'], fleet_on_demand_option: Optional['outputs.InstancePoolInstancePoolFleetAttributesFleetOnDemandOption'] = None, fleet_spot_option: Optional['outputs.InstancePoolInstancePoolFleetAttributesFleetSpotOption'] = None): - pulumi.set(__self__, "launch_template_overrides", launch_template_overrides) + InstancePoolInstancePoolFleetAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + launch_template_overrides=launch_template_overrides, + fleet_on_demand_option=fleet_on_demand_option, + fleet_spot_option=fleet_spot_option, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + launch_template_overrides: Sequence['outputs.InstancePoolInstancePoolFleetAttributesLaunchTemplateOverride'], + fleet_on_demand_option: Optional['outputs.InstancePoolInstancePoolFleetAttributesFleetOnDemandOption'] = None, + fleet_spot_option: Optional['outputs.InstancePoolInstancePoolFleetAttributesFleetSpotOption'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("launch_template_overrides", launch_template_overrides) if fleet_on_demand_option is not None: - pulumi.set(__self__, "fleet_on_demand_option", fleet_on_demand_option) + _setter("fleet_on_demand_option", fleet_on_demand_option) if fleet_spot_option is not None: - pulumi.set(__self__, "fleet_spot_option", fleet_spot_option) + _setter("fleet_spot_option", fleet_spot_option) @property @pulumi.getter(name="launchTemplateOverrides") @@ -1818,9 +2255,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, allocation_strategy: str, instance_pools_to_use_count: Optional[int] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + InstancePoolInstancePoolFleetAttributesFleetOnDemandOption._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: str, + instance_pools_to_use_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -1857,9 +2305,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, allocation_strategy: str, instance_pools_to_use_count: Optional[int] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + InstancePoolInstancePoolFleetAttributesFleetSpotOption._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: str, + instance_pools_to_use_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -1896,8 +2355,19 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, availability_zone: str, instance_type: str): - pulumi.set(__self__, "availability_zone", availability_zone) - pulumi.set(__self__, "instance_type", instance_type) + InstancePoolInstancePoolFleetAttributesLaunchTemplateOverride._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability_zone=availability_zone, + instance_type=instance_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability_zone: str, + instance_type: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("availability_zone", availability_zone) + _setter("instance_type", instance_type) @property @pulumi.getter(name="availabilityZone") @@ -1932,9 +2402,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, url: str, basic_auth: Optional['outputs.InstancePoolPreloadedDockerImageBasicAuth'] = None): - pulumi.set(__self__, "url", url) + InstancePoolPreloadedDockerImage._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.InstancePoolPreloadedDockerImageBasicAuth'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -1952,8 +2433,19 @@ class InstancePoolPreloadedDockerImageBasicAuth(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + InstancePoolPreloadedDockerImageBasicAuth._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -1988,10 +2480,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, compute_key: Optional[str] = None, spec: Optional['outputs.JobComputeSpec'] = None): + JobCompute._configure( + lambda key, value: pulumi.set(__self__, key, value), + compute_key=compute_key, + spec=spec, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + compute_key: Optional[str] = None, + spec: Optional['outputs.JobComputeSpec'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if spec is not None: - pulumi.set(__self__, "spec", spec) + _setter("spec", spec) @property @pulumi.getter(name="computeKey") @@ -2008,8 +2511,17 @@ def spec(self) -> Optional['outputs.JobComputeSpec']: class JobComputeSpec(dict): def __init__(__self__, *, kind: Optional[str] = None): + JobComputeSpec._configure( + lambda key, value: pulumi.set(__self__, key, value), + kind=kind, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + kind: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if kind is not None: - pulumi.set(__self__, "kind", kind) + _setter("kind", kind) @property @pulumi.getter @@ -2041,8 +2553,17 @@ def __init__(__self__, *, """ :param str pause_status: Indicate whether this continuous job is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pause_status`. """ + JobContinuous._configure( + lambda key, value: pulumi.set(__self__, key, value), + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="pauseStatus") @@ -2093,17 +2614,36 @@ def __init__(__self__, *, You also need to include a `git_source` block to configure the repository that contains the dbt project. """ - pulumi.set(__self__, "commands", commands) + JobDbtTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: Sequence[str], + catalog: Optional[str] = None, + profiles_directory: Optional[str] = None, + project_directory: Optional[str] = None, + schema: Optional[str] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -2200,18 +2740,37 @@ def __init__(__self__, *, :param Sequence[str] on_starts: (List) list of emails to notify when the run starts. :param Sequence[str] on_successes: (List) list of emails to notify when the run completes successfully. """ + JobEmailNotifications._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + on_duration_warning_threshold_exceededs: Optional[Sequence[str]] = None, + on_failures: Optional[Sequence[str]] = None, + on_starts: Optional[Sequence[str]] = None, + on_successes: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -2295,17 +2854,36 @@ def __init__(__self__, *, :param str provider: case insensitive name of the Git provider. Following values are supported right now (could be a subject for change, consult [Repos API documentation](https://docs.databricks.com/dev-tools/api/latest/repos.html)): `gitHub`, `gitHubEnterprise`, `bitbucketCloud`, `bitbucketServer`, `azureDevOpsServices`, `gitLab`, `gitLabEnterpriseEdition`. :param str tag: name of the Git branch to use. Conflicts with `branch` and `commit`. """ - pulumi.set(__self__, "url", url) + JobGitSource._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + branch=branch, + commit=commit, + job_source=job_source, + provider=provider, + tag=tag, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + branch: Optional[str] = None, + commit: Optional[str] = None, + job_source: Optional['outputs.JobGitSourceJobSource'] = None, + provider: Optional[str] = None, + tag: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if branch is not None: - pulumi.set(__self__, "branch", branch) + _setter("branch", branch) if commit is not None: - pulumi.set(__self__, "commit", commit) + _setter("commit", commit) if job_source is not None: - pulumi.set(__self__, "job_source", job_source) + _setter("job_source", job_source) if provider is not None: - pulumi.set(__self__, "provider", provider) + _setter("provider", provider) if tag is not None: - pulumi.set(__self__, "tag", tag) + _setter("tag", tag) @property @pulumi.getter @@ -2380,10 +2958,23 @@ def __init__(__self__, *, import_from_git_branch: str, job_config_path: str, dirty_state: Optional[str] = None): - pulumi.set(__self__, "import_from_git_branch", import_from_git_branch) - pulumi.set(__self__, "job_config_path", job_config_path) + JobGitSourceJobSource._configure( + lambda key, value: pulumi.set(__self__, key, value), + import_from_git_branch=import_from_git_branch, + job_config_path=job_config_path, + dirty_state=dirty_state, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + import_from_git_branch: str, + job_config_path: str, + dirty_state: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("import_from_git_branch", import_from_git_branch) + _setter("job_config_path", job_config_path) if dirty_state is not None: - pulumi.set(__self__, "dirty_state", dirty_state) + _setter("dirty_state", dirty_state) @property @pulumi.getter(name="importFromGitBranch") @@ -2408,7 +2999,16 @@ def __init__(__self__, *, """ :param Sequence['JobHealthRuleArgs'] rules: list of rules that are represented as objects with the following attributes: """ - pulumi.set(__self__, "rules", rules) + JobHealth._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: Sequence['outputs.JobHealthRule'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -2430,12 +3030,25 @@ def __init__(__self__, *, :param str op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param int value: integer value used to compare to the given metric. """ + JobHealthRule._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[str] = None, + op: Optional[str] = None, + value: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -2490,10 +3103,21 @@ def __init__(__self__, *, :param str job_cluster_key: Identifier that can be referenced in `task` block, so that cluster is shared between tasks :param 'JobJobClusterNewClusterArgs' new_cluster: Same set of parameters as for Cluster resource. """ + JobJobCluster._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_cluster_key=job_cluster_key, + new_cluster=new_cluster, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_cluster_key: Optional[str] = None, + new_cluster: Optional['outputs.JobJobClusterNewCluster'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") @@ -2618,65 +3242,132 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['outputs.JobJobClusterNewClusterWorkloadType'] = None): - pulumi.set(__self__, "spark_version", spark_version) + JobJobClusterNewCluster._configure( + lambda key, value: pulumi.set(__self__, key, value), + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['outputs.JobJobClusterNewClusterAutoscale'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['outputs.JobJobClusterNewClusterAwsAttributes'] = None, + azure_attributes: Optional['outputs.JobJobClusterNewClusterAzureAttributes'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['outputs.JobJobClusterNewClusterClusterLogConf'] = None, + cluster_mount_infos: Optional[Sequence['outputs.JobJobClusterNewClusterClusterMountInfo']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['outputs.JobJobClusterNewClusterDockerImage'] = None, + driver_instance_pool_id: Optional[str] = None, + driver_node_type_id: Optional[str] = None, + enable_elastic_disk: Optional[bool] = None, + enable_local_disk_encryption: Optional[bool] = None, + gcp_attributes: Optional['outputs.JobJobClusterNewClusterGcpAttributes'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['outputs.JobJobClusterNewClusterInitScript']] = None, + instance_pool_id: Optional[str] = None, + node_type_id: Optional[str] = None, + num_workers: Optional[int] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['outputs.JobJobClusterNewClusterWorkloadType'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="sparkVersion") @@ -2853,10 +3544,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + JobJobClusterNewClusterAutoscale._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -2909,22 +3611,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + JobJobClusterNewClusterAwsAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -2992,12 +3717,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + JobJobClusterNewClusterAzureAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -3020,10 +3758,21 @@ class JobJobClusterNewClusterClusterLogConf(dict): def __init__(__self__, *, dbfs: Optional['outputs.JobJobClusterNewClusterClusterLogConfDbfs'] = None, s3: Optional['outputs.JobJobClusterNewClusterClusterLogConfS3'] = None): + JobJobClusterNewClusterClusterLogConf._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.JobJobClusterNewClusterClusterLogConfDbfs'] = None, + s3: Optional['outputs.JobJobClusterNewClusterClusterLogConfS3'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -3040,7 +3789,16 @@ def s3(self) -> Optional['outputs.JobJobClusterNewClusterClusterLogConfS3']: class JobJobClusterNewClusterClusterLogConfDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterClusterLogConfDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -3081,19 +3839,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterClusterLogConfS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -3158,10 +3937,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'outputs.JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfo', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + JobJobClusterNewClusterClusterMountInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'outputs.JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfo', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -3203,9 +3995,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + JobJobClusterNewClusterClusterMountInfoNetworkFilesystemInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -3241,17 +4044,28 @@ def __init__(__self__, *, url: str, basic_auth: Optional['outputs.JobJobClusterNewClusterDockerImageBasicAuth'] = None): """ - :param str url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + :param str url: URL of the job on the given workspace """ - pulumi.set(__self__, "url", url) + JobJobClusterNewClusterDockerImage._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.JobJobClusterNewClusterDockerImageBasicAuth'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter def url(self) -> str: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -3266,8 +4080,19 @@ class JobJobClusterNewClusterDockerImageBasicAuth(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + JobJobClusterNewClusterDockerImageBasicAuth._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -3314,18 +4139,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + JobJobClusterNewClusterGcpAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -3414,20 +4258,41 @@ def __init__(__self__, *, ]) ``` """ + JobJobClusterNewClusterInitScript._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.JobJobClusterNewClusterInitScriptAbfss'] = None, + dbfs: Optional['outputs.JobJobClusterNewClusterInitScriptDbfs'] = None, + file: Optional['outputs.JobJobClusterNewClusterInitScriptFile'] = None, + gcs: Optional['outputs.JobJobClusterNewClusterInitScriptGcs'] = None, + s3: Optional['outputs.JobJobClusterNewClusterInitScriptS3'] = None, + volumes: Optional['outputs.JobJobClusterNewClusterInitScriptVolumes'] = None, + workspace: Optional['outputs.JobJobClusterNewClusterInitScriptWorkspace'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -3518,8 +4383,17 @@ def workspace(self) -> Optional['outputs.JobJobClusterNewClusterInitScriptWorksp class JobJobClusterNewClusterInitScriptAbfss(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobJobClusterNewClusterInitScriptAbfss._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3531,7 +4405,16 @@ def destination(self) -> Optional[str]: class JobJobClusterNewClusterInitScriptDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterInitScriptDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -3543,8 +4426,17 @@ def destination(self) -> str: class JobJobClusterNewClusterInitScriptFile(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobJobClusterNewClusterInitScriptFile._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3556,8 +4448,17 @@ def destination(self) -> Optional[str]: class JobJobClusterNewClusterInitScriptGcs(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobJobClusterNewClusterInitScriptGcs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3598,19 +4499,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + JobJobClusterNewClusterInitScriptS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -3652,8 +4574,17 @@ def region(self) -> Optional[str]: class JobJobClusterNewClusterInitScriptVolumes(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobJobClusterNewClusterInitScriptVolumes._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3665,8 +4596,17 @@ def destination(self) -> Optional[str]: class JobJobClusterNewClusterInitScriptWorkspace(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobJobClusterNewClusterInitScriptWorkspace._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -3678,7 +4618,16 @@ def destination(self) -> Optional[str]: class JobJobClusterNewClusterWorkloadType(dict): def __init__(__self__, *, clients: 'outputs.JobJobClusterNewClusterWorkloadTypeClients'): - pulumi.set(__self__, "clients", clients) + JobJobClusterNewClusterWorkloadType._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'outputs.JobJobClusterNewClusterWorkloadTypeClients', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -3691,10 +4640,21 @@ class JobJobClusterNewClusterWorkloadTypeClients(dict): def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + JobJobClusterNewClusterWorkloadTypeClients._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -3716,18 +4676,37 @@ def __init__(__self__, *, maven: Optional['outputs.JobLibraryMaven'] = None, pypi: Optional['outputs.JobLibraryPypi'] = None, whl: Optional[str] = None): + JobLibrary._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional['outputs.JobLibraryCran'] = None, + egg: Optional[str] = None, + jar: Optional[str] = None, + maven: Optional['outputs.JobLibraryMaven'] = None, + pypi: Optional['outputs.JobLibraryPypi'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -3765,9 +4744,20 @@ class JobLibraryCran(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + JobLibraryCran._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -3786,11 +4776,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + JobLibraryMaven._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -3813,9 +4816,20 @@ class JobLibraryPypi(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + JobLibraryPypi._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -3934,65 +4948,132 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['outputs.JobNewClusterWorkloadType'] = None): - pulumi.set(__self__, "spark_version", spark_version) + JobNewCluster._configure( + lambda key, value: pulumi.set(__self__, key, value), + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['outputs.JobNewClusterAutoscale'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['outputs.JobNewClusterAwsAttributes'] = None, + azure_attributes: Optional['outputs.JobNewClusterAzureAttributes'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['outputs.JobNewClusterClusterLogConf'] = None, + cluster_mount_infos: Optional[Sequence['outputs.JobNewClusterClusterMountInfo']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['outputs.JobNewClusterDockerImage'] = None, + driver_instance_pool_id: Optional[str] = None, + driver_node_type_id: Optional[str] = None, + enable_elastic_disk: Optional[bool] = None, + enable_local_disk_encryption: Optional[bool] = None, + gcp_attributes: Optional['outputs.JobNewClusterGcpAttributes'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['outputs.JobNewClusterInitScript']] = None, + instance_pool_id: Optional[str] = None, + node_type_id: Optional[str] = None, + num_workers: Optional[int] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['outputs.JobNewClusterWorkloadType'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="sparkVersion") @@ -4169,10 +5250,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + JobNewClusterAutoscale._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -4225,22 +5317,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + JobNewClusterAwsAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -4308,12 +5423,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + JobNewClusterAzureAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -4336,10 +5464,21 @@ class JobNewClusterClusterLogConf(dict): def __init__(__self__, *, dbfs: Optional['outputs.JobNewClusterClusterLogConfDbfs'] = None, s3: Optional['outputs.JobNewClusterClusterLogConfS3'] = None): + JobNewClusterClusterLogConf._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.JobNewClusterClusterLogConfDbfs'] = None, + s3: Optional['outputs.JobNewClusterClusterLogConfS3'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -4356,7 +5495,16 @@ def s3(self) -> Optional['outputs.JobNewClusterClusterLogConfS3']: class JobNewClusterClusterLogConfDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + JobNewClusterClusterLogConfDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -4397,19 +5545,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + JobNewClusterClusterLogConfS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -4474,10 +5643,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'outputs.JobNewClusterClusterMountInfoNetworkFilesystemInfo', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + JobNewClusterClusterMountInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'outputs.JobNewClusterClusterMountInfoNetworkFilesystemInfo', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -4519,9 +5701,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + JobNewClusterClusterMountInfoNetworkFilesystemInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -4557,17 +5750,28 @@ def __init__(__self__, *, url: str, basic_auth: Optional['outputs.JobNewClusterDockerImageBasicAuth'] = None): """ - :param str url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + :param str url: URL of the job on the given workspace """ - pulumi.set(__self__, "url", url) + JobNewClusterDockerImage._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.JobNewClusterDockerImageBasicAuth'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter def url(self) -> str: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -4582,8 +5786,19 @@ class JobNewClusterDockerImageBasicAuth(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + JobNewClusterDockerImageBasicAuth._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -4630,18 +5845,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + JobNewClusterGcpAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -4730,20 +5964,41 @@ def __init__(__self__, *, ]) ``` """ + JobNewClusterInitScript._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.JobNewClusterInitScriptAbfss'] = None, + dbfs: Optional['outputs.JobNewClusterInitScriptDbfs'] = None, + file: Optional['outputs.JobNewClusterInitScriptFile'] = None, + gcs: Optional['outputs.JobNewClusterInitScriptGcs'] = None, + s3: Optional['outputs.JobNewClusterInitScriptS3'] = None, + volumes: Optional['outputs.JobNewClusterInitScriptVolumes'] = None, + workspace: Optional['outputs.JobNewClusterInitScriptWorkspace'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -4834,8 +6089,17 @@ def workspace(self) -> Optional['outputs.JobNewClusterInitScriptWorkspace']: class JobNewClusterInitScriptAbfss(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobNewClusterInitScriptAbfss._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -4847,7 +6111,16 @@ def destination(self) -> Optional[str]: class JobNewClusterInitScriptDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + JobNewClusterInitScriptDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -4859,8 +6132,17 @@ def destination(self) -> str: class JobNewClusterInitScriptFile(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobNewClusterInitScriptFile._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -4872,8 +6154,17 @@ def destination(self) -> Optional[str]: class JobNewClusterInitScriptGcs(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobNewClusterInitScriptGcs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -4914,19 +6205,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + JobNewClusterInitScriptS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -4968,8 +6280,17 @@ def region(self) -> Optional[str]: class JobNewClusterInitScriptVolumes(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobNewClusterInitScriptVolumes._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -4981,8 +6302,17 @@ def destination(self) -> Optional[str]: class JobNewClusterInitScriptWorkspace(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobNewClusterInitScriptWorkspace._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -4994,7 +6324,16 @@ def destination(self) -> Optional[str]: class JobNewClusterWorkloadType(dict): def __init__(__self__, *, clients: 'outputs.JobNewClusterWorkloadTypeClients'): - pulumi.set(__self__, "clients", clients) + JobNewClusterWorkloadType._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'outputs.JobNewClusterWorkloadTypeClients', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -5007,10 +6346,21 @@ class JobNewClusterWorkloadTypeClients(dict): def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + JobNewClusterWorkloadTypeClients._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -5053,11 +6403,24 @@ def __init__(__self__, *, :param Mapping[str, Any] base_parameters: (Map) Base parameters to be used for each run of this job. If the run is initiated by a call to run-now with parameters specified, the two parameters maps will be merged. If the same key is specified in base_parameters and in run-now, the value from run-now will be used. If the notebook takes a parameter that is not specified in the job’s base_parameters or the run-now override parameters, the default value from the notebook will be used. Retrieve these parameters in a notebook using `dbutils.widgets.get`. :param str source: Location type of the notebook, can only be `WORKSPACE` or `GIT`. When set to `WORKSPACE`, the notebook will be retrieved from the local Databricks workspace. When set to `GIT`, the notebook will be retrieved from a Git repository defined in `git_source`. If the value is empty, the task will use `GIT` if `git_source` is defined and `WORKSPACE` otherwise. """ - pulumi.set(__self__, "notebook_path", notebook_path) + JobNotebookTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: str, + base_parameters: Optional[Mapping[str, Any]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -5112,10 +6475,21 @@ def __init__(__self__, *, :param bool no_alert_for_canceled_runs: (Bool) don't send alert for cancelled runs. :param bool no_alert_for_skipped_runs: (Bool) don't send alert for skipped runs. """ + JobNotificationSettings._configure( + lambda key, value: pulumi.set(__self__, key, value), + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + no_alert_for_canceled_runs: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="noAlertForCanceledRuns") @@ -5143,10 +6517,21 @@ def __init__(__self__, *, :param str default: Default value of the parameter. :param str name: An optional name for the job. The default value is Untitled. """ + JobParameter._configure( + lambda key, value: pulumi.set(__self__, key, value), + default=default, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default: Optional[str] = None, + name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if default is not None: - pulumi.set(__self__, "default", default) + _setter("default", default) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -5195,9 +6580,20 @@ def __init__(__self__, *, > **Note** The following configuration blocks are only supported inside a `task` block """ - pulumi.set(__self__, "pipeline_id", pipeline_id) + JobPipelineTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: str, + full_refresh: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -5252,14 +6648,29 @@ def __init__(__self__, *, :param str package_name: Name of Python package :param Sequence[str] parameters: Parameters for the task """ + JobPythonWheelTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[str] = None, + named_parameters: Optional[Mapping[str, Any]] = None, + package_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -5298,6 +6709,11 @@ def parameters(self) -> Optional[Sequence[str]]: class JobQueue(dict): def __init__(__self__): pass + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + opts: Optional[pulumi.ResourceOptions]=None): + pass @pulumi.output_type @@ -5339,10 +6755,21 @@ def __init__(__self__, *, ``` :param str user_name: The email of an active workspace user. Non-admin users can only set this field to their own email. """ + JobRunAs._configure( + lambda key, value: pulumi.set(__self__, key, value), + service_principal_name=service_principal_name, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + service_principal_name: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if service_principal_name is not None: - pulumi.set(__self__, "service_principal_name", service_principal_name) + _setter("service_principal_name", service_principal_name) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="servicePrincipalName") @@ -5400,9 +6827,20 @@ def __init__(__self__, *, :param int job_id: (String) ID of the job :param Mapping[str, Any] job_parameters: (Map) Job parameters for the task """ - pulumi.set(__self__, "job_id", job_id) + JobRunJobTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: int, + job_parameters: Optional[Mapping[str, Any]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -5453,10 +6891,23 @@ def __init__(__self__, *, :param str timezone_id: A Java timezone ID. The schedule for a job will be resolved with respect to this timezone. See Java TimeZone for details. This field is required. :param str pause_status: Indicate whether this schedule is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted and a schedule is provided, the server will default to using `UNPAUSED` as a value for `pause_status`. """ - pulumi.set(__self__, "quartz_cron_expression", quartz_cron_expression) - pulumi.set(__self__, "timezone_id", timezone_id) + JobSchedule._configure( + lambda key, value: pulumi.set(__self__, key, value), + quartz_cron_expression=quartz_cron_expression, + timezone_id=timezone_id, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + quartz_cron_expression: str, + timezone_id: str, + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("quartz_cron_expression", quartz_cron_expression) + _setter("timezone_id", timezone_id) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="quartzCronExpression") @@ -5512,12 +6963,25 @@ def __init__(__self__, *, :param str main_class_name: The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. The code should use `SparkContext.getOrCreate` to obtain a Spark context; otherwise, runs of the job will fail. :param Sequence[str] parameters: (List) Parameters passed to the main method. """ + JobSparkJarTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[str] = None, + main_class_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -5569,11 +7033,24 @@ def __init__(__self__, *, :param Sequence[str] parameters: (List) Command line parameters passed to the Python file. :param str source: Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `git_source`. """ - pulumi.set(__self__, "python_file", python_file) + JobSparkPythonTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: str, + parameters: Optional[Sequence[str]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -5607,8 +7084,17 @@ def __init__(__self__, *, """ :param Sequence[str] parameters: (List) Command-line parameters passed to spark submit. """ + JobSparkSubmitTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -5725,58 +7211,117 @@ def __init__(__self__, *, * `*_task` - (Required) one of the specific task blocks described below: :param int timeout_seconds: (Integer) An optional timeout applied to each run of this job. The default behavior is to have no timeout. """ + JobTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + compute_key=compute_key, + condition_task=condition_task, + dbt_task=dbt_task, + depends_ons=depends_ons, + description=description, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + health=health, + job_cluster_key=job_cluster_key, + libraries=libraries, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + retry_on_timeout=retry_on_timeout, + run_if=run_if, + run_job_task=run_job_task, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + sql_task=sql_task, + task_key=task_key, + timeout_seconds=timeout_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + compute_key: Optional[str] = None, + condition_task: Optional['outputs.JobTaskConditionTask'] = None, + dbt_task: Optional['outputs.JobTaskDbtTask'] = None, + depends_ons: Optional[Sequence['outputs.JobTaskDependsOn']] = None, + description: Optional[str] = None, + email_notifications: Optional['outputs.JobTaskEmailNotifications'] = None, + existing_cluster_id: Optional[str] = None, + health: Optional['outputs.JobTaskHealth'] = None, + job_cluster_key: Optional[str] = None, + libraries: Optional[Sequence['outputs.JobTaskLibrary']] = None, + max_retries: Optional[int] = None, + min_retry_interval_millis: Optional[int] = None, + new_cluster: Optional['outputs.JobTaskNewCluster'] = None, + notebook_task: Optional['outputs.JobTaskNotebookTask'] = None, + notification_settings: Optional['outputs.JobTaskNotificationSettings'] = None, + pipeline_task: Optional['outputs.JobTaskPipelineTask'] = None, + python_wheel_task: Optional['outputs.JobTaskPythonWheelTask'] = None, + retry_on_timeout: Optional[bool] = None, + run_if: Optional[str] = None, + run_job_task: Optional['outputs.JobTaskRunJobTask'] = None, + spark_jar_task: Optional['outputs.JobTaskSparkJarTask'] = None, + spark_python_task: Optional['outputs.JobTaskSparkPythonTask'] = None, + spark_submit_task: Optional['outputs.JobTaskSparkSubmitTask'] = None, + sql_task: Optional['outputs.JobTaskSqlTask'] = None, + task_key: Optional[str] = None, + timeout_seconds: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if condition_task is not None: - pulumi.set(__self__, "condition_task", condition_task) + _setter("condition_task", condition_task) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if depends_ons is not None: - pulumi.set(__self__, "depends_ons", depends_ons) + _setter("depends_ons", depends_ons) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if retry_on_timeout is not None: - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + _setter("retry_on_timeout", retry_on_timeout) if run_if is not None: - pulumi.set(__self__, "run_if", run_if) + _setter("run_if", run_if) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if sql_task is not None: - pulumi.set(__self__, "sql_task", sql_task) + _setter("sql_task", sql_task) if task_key is not None: - pulumi.set(__self__, "task_key", task_key) + _setter("task_key", task_key) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) @property @pulumi.getter(name="computeKey") @@ -5958,12 +7503,25 @@ def __init__(__self__, *, """ :param str op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. """ + JobTaskConditionTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + left=left, + op=op, + right=right, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + left: Optional[str] = None, + op: Optional[str] = None, + right: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if left is not None: - pulumi.set(__self__, "left", left) + _setter("left", left) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if right is not None: - pulumi.set(__self__, "right", right) + _setter("right", right) @property @pulumi.getter @@ -6024,17 +7582,36 @@ def __init__(__self__, *, You also need to include a `git_source` block to configure the repository that contains the dbt project. """ - pulumi.set(__self__, "commands", commands) + JobTaskDbtTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: Sequence[str], + catalog: Optional[str] = None, + profiles_directory: Optional[str] = None, + project_directory: Optional[str] = None, + schema: Optional[str] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -6112,9 +7689,20 @@ def __init__(__self__, *, """ :param str task_key: The name of the task this task depends on. """ - pulumi.set(__self__, "task_key", task_key) + JobTaskDependsOn._configure( + lambda key, value: pulumi.set(__self__, key, value), + task_key=task_key, + outcome=outcome, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + task_key: str, + outcome: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("task_key", task_key) if outcome is not None: - pulumi.set(__self__, "outcome", outcome) + _setter("outcome", outcome) @property @pulumi.getter(name="taskKey") @@ -6174,18 +7762,37 @@ def __init__(__self__, *, :param Sequence[str] on_starts: (List) list of emails to notify when the run starts. :param Sequence[str] on_successes: (List) list of emails to notify when the run completes successfully. """ + JobTaskEmailNotifications._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + on_duration_warning_threshold_exceededs: Optional[Sequence[str]] = None, + on_failures: Optional[Sequence[str]] = None, + on_starts: Optional[Sequence[str]] = None, + on_successes: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -6243,7 +7850,16 @@ def __init__(__self__, *, """ :param Sequence['JobTaskHealthRuleArgs'] rules: list of rules that are represented as objects with the following attributes: """ - pulumi.set(__self__, "rules", rules) + JobTaskHealth._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: Sequence['outputs.JobTaskHealthRule'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -6265,12 +7881,25 @@ def __init__(__self__, *, :param str op: string specifying the operation used to evaluate the given metric. The only supported operation is `GREATER_THAN`. :param int value: integer value used to compare to the given metric. """ + JobTaskHealthRule._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[str] = None, + op: Optional[str] = None, + value: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -6306,18 +7935,37 @@ def __init__(__self__, *, maven: Optional['outputs.JobTaskLibraryMaven'] = None, pypi: Optional['outputs.JobTaskLibraryPypi'] = None, whl: Optional[str] = None): + JobTaskLibrary._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional['outputs.JobTaskLibraryCran'] = None, + egg: Optional[str] = None, + jar: Optional[str] = None, + maven: Optional['outputs.JobTaskLibraryMaven'] = None, + pypi: Optional['outputs.JobTaskLibraryPypi'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -6355,9 +8003,20 @@ class JobTaskLibraryCran(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + JobTaskLibraryCran._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -6376,11 +8035,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + JobTaskLibraryMaven._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -6403,9 +8075,20 @@ class JobTaskLibraryPypi(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + JobTaskLibraryPypi._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -6524,65 +8207,132 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['outputs.JobTaskNewClusterWorkloadType'] = None): - pulumi.set(__self__, "spark_version", spark_version) + JobTaskNewCluster._configure( + lambda key, value: pulumi.set(__self__, key, value), + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['outputs.JobTaskNewClusterAutoscale'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['outputs.JobTaskNewClusterAwsAttributes'] = None, + azure_attributes: Optional['outputs.JobTaskNewClusterAzureAttributes'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['outputs.JobTaskNewClusterClusterLogConf'] = None, + cluster_mount_infos: Optional[Sequence['outputs.JobTaskNewClusterClusterMountInfo']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['outputs.JobTaskNewClusterDockerImage'] = None, + driver_instance_pool_id: Optional[str] = None, + driver_node_type_id: Optional[str] = None, + enable_elastic_disk: Optional[bool] = None, + enable_local_disk_encryption: Optional[bool] = None, + gcp_attributes: Optional['outputs.JobTaskNewClusterGcpAttributes'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['outputs.JobTaskNewClusterInitScript']] = None, + instance_pool_id: Optional[str] = None, + node_type_id: Optional[str] = None, + num_workers: Optional[int] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['outputs.JobTaskNewClusterWorkloadType'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="sparkVersion") @@ -6759,10 +8509,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + JobTaskNewClusterAutoscale._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -6815,22 +8576,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + JobTaskNewClusterAwsAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -6898,12 +8682,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + JobTaskNewClusterAzureAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -6926,10 +8723,21 @@ class JobTaskNewClusterClusterLogConf(dict): def __init__(__self__, *, dbfs: Optional['outputs.JobTaskNewClusterClusterLogConfDbfs'] = None, s3: Optional['outputs.JobTaskNewClusterClusterLogConfS3'] = None): + JobTaskNewClusterClusterLogConf._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.JobTaskNewClusterClusterLogConfDbfs'] = None, + s3: Optional['outputs.JobTaskNewClusterClusterLogConfS3'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -6946,7 +8754,16 @@ def s3(self) -> Optional['outputs.JobTaskNewClusterClusterLogConfS3']: class JobTaskNewClusterClusterLogConfDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterClusterLogConfDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -6987,19 +8804,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterClusterLogConfS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -7064,10 +8902,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'outputs.JobTaskNewClusterClusterMountInfoNetworkFilesystemInfo', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + JobTaskNewClusterClusterMountInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'outputs.JobTaskNewClusterClusterMountInfoNetworkFilesystemInfo', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -7109,9 +8960,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + JobTaskNewClusterClusterMountInfoNetworkFilesystemInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -7147,17 +9009,28 @@ def __init__(__self__, *, url: str, basic_auth: Optional['outputs.JobTaskNewClusterDockerImageBasicAuth'] = None): """ - :param str url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + :param str url: URL of the job on the given workspace """ - pulumi.set(__self__, "url", url) + JobTaskNewClusterDockerImage._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.JobTaskNewClusterDockerImageBasicAuth'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter def url(self) -> str: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -7172,8 +9045,19 @@ class JobTaskNewClusterDockerImageBasicAuth(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + JobTaskNewClusterDockerImageBasicAuth._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -7220,18 +9104,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + JobTaskNewClusterGcpAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -7320,20 +9223,41 @@ def __init__(__self__, *, ]) ``` """ + JobTaskNewClusterInitScript._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.JobTaskNewClusterInitScriptAbfss'] = None, + dbfs: Optional['outputs.JobTaskNewClusterInitScriptDbfs'] = None, + file: Optional['outputs.JobTaskNewClusterInitScriptFile'] = None, + gcs: Optional['outputs.JobTaskNewClusterInitScriptGcs'] = None, + s3: Optional['outputs.JobTaskNewClusterInitScriptS3'] = None, + volumes: Optional['outputs.JobTaskNewClusterInitScriptVolumes'] = None, + workspace: Optional['outputs.JobTaskNewClusterInitScriptWorkspace'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -7424,8 +9348,17 @@ def workspace(self) -> Optional['outputs.JobTaskNewClusterInitScriptWorkspace']: class JobTaskNewClusterInitScriptAbfss(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobTaskNewClusterInitScriptAbfss._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7437,7 +9370,16 @@ def destination(self) -> Optional[str]: class JobTaskNewClusterInitScriptDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterInitScriptDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -7449,8 +9391,17 @@ def destination(self) -> str: class JobTaskNewClusterInitScriptFile(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobTaskNewClusterInitScriptFile._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7462,8 +9413,17 @@ def destination(self) -> Optional[str]: class JobTaskNewClusterInitScriptGcs(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobTaskNewClusterInitScriptGcs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7504,19 +9464,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + JobTaskNewClusterInitScriptS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -7558,8 +9539,17 @@ def region(self) -> Optional[str]: class JobTaskNewClusterInitScriptVolumes(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobTaskNewClusterInitScriptVolumes._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7571,8 +9561,17 @@ def destination(self) -> Optional[str]: class JobTaskNewClusterInitScriptWorkspace(dict): def __init__(__self__, *, destination: Optional[str] = None): + JobTaskNewClusterInitScriptWorkspace._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -7584,7 +9583,16 @@ def destination(self) -> Optional[str]: class JobTaskNewClusterWorkloadType(dict): def __init__(__self__, *, clients: 'outputs.JobTaskNewClusterWorkloadTypeClients'): - pulumi.set(__self__, "clients", clients) + JobTaskNewClusterWorkloadType._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'outputs.JobTaskNewClusterWorkloadTypeClients', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -7597,10 +9605,21 @@ class JobTaskNewClusterWorkloadTypeClients(dict): def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + JobTaskNewClusterWorkloadTypeClients._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -7643,11 +9662,24 @@ def __init__(__self__, *, :param Mapping[str, Any] base_parameters: (Map) Base parameters to be used for each run of this job. If the run is initiated by a call to run-now with parameters specified, the two parameters maps will be merged. If the same key is specified in base_parameters and in run-now, the value from run-now will be used. If the notebook takes a parameter that is not specified in the job’s base_parameters or the run-now override parameters, the default value from the notebook will be used. Retrieve these parameters in a notebook using `dbutils.widgets.get`. :param str source: Location type of the notebook, can only be `WORKSPACE` or `GIT`. When set to `WORKSPACE`, the notebook will be retrieved from the local Databricks workspace. When set to `GIT`, the notebook will be retrieved from a Git repository defined in `git_source`. If the value is empty, the task will use `GIT` if `git_source` is defined and `WORKSPACE` otherwise. """ - pulumi.set(__self__, "notebook_path", notebook_path) + JobTaskNotebookTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: str, + base_parameters: Optional[Mapping[str, Any]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -7706,12 +9738,25 @@ def __init__(__self__, *, :param bool no_alert_for_canceled_runs: (Bool) don't send alert for cancelled runs. :param bool no_alert_for_skipped_runs: (Bool) don't send alert for skipped runs. """ + JobTaskNotificationSettings._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_canceled_runs: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="alertOnLastAttempt") @@ -7768,9 +9813,20 @@ def __init__(__self__, *, > **Note** The following configuration blocks are only supported inside a `task` block """ - pulumi.set(__self__, "pipeline_id", pipeline_id) + JobTaskPipelineTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: str, + full_refresh: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -7825,14 +9881,29 @@ def __init__(__self__, *, :param str package_name: Name of Python package :param Sequence[str] parameters: Parameters for the task """ + JobTaskPythonWheelTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[str] = None, + named_parameters: Optional[Mapping[str, Any]] = None, + package_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -7895,9 +9966,20 @@ def __init__(__self__, *, :param int job_id: (String) ID of the job :param Mapping[str, Any] job_parameters: (Map) Job parameters for the task """ - pulumi.set(__self__, "job_id", job_id) + JobTaskRunJobTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: int, + job_parameters: Optional[Mapping[str, Any]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -7945,12 +10027,25 @@ def __init__(__self__, *, :param str main_class_name: The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. The code should use `SparkContext.getOrCreate` to obtain a Spark context; otherwise, runs of the job will fail. :param Sequence[str] parameters: (List) Parameters passed to the main method. """ + JobTaskSparkJarTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[str] = None, + main_class_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -8002,11 +10097,24 @@ def __init__(__self__, *, :param Sequence[str] parameters: (List) Command line parameters passed to the Python file. :param str source: Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `git_source`. """ - pulumi.set(__self__, "python_file", python_file) + JobTaskSparkPythonTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: str, + parameters: Optional[Sequence[str]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -8040,8 +10148,17 @@ def __init__(__self__, *, """ :param Sequence[str] parameters: (List) Command-line parameters passed to spark submit. """ + JobTaskSparkSubmitTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -8129,18 +10246,37 @@ def __init__(__self__, *, :param 'JobTaskSqlTaskQueryArgs' query: block consisting of single string field: `query_id` - identifier of the Databricks SQL Query (databricks_sql_query). :param str warehouse_id: ID of the (the databricks_sql_endpoint) that will be used to execute the task. Only Serverless & Pro warehouses are supported right now. """ + JobTaskSqlTask._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert=alert, + dashboard=dashboard, + file=file, + parameters=parameters, + query=query, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert: Optional['outputs.JobTaskSqlTaskAlert'] = None, + dashboard: Optional['outputs.JobTaskSqlTaskDashboard'] = None, + file: Optional['outputs.JobTaskSqlTaskFile'] = None, + parameters: Optional[Mapping[str, Any]] = None, + query: Optional['outputs.JobTaskSqlTaskQuery'] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert is not None: - pulumi.set(__self__, "alert", alert) + _setter("alert", alert) if dashboard is not None: - pulumi.set(__self__, "dashboard", dashboard) + _setter("dashboard", dashboard) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if query is not None: - pulumi.set(__self__, "query", query) + _setter("query", query) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -8264,10 +10400,23 @@ def __init__(__self__, *, :param Sequence['JobTaskSqlTaskAlertSubscriptionArgs'] subscriptions: a list of subscription blocks consisting out of one of the required fields: `user_name` for user emails or `destination_id` - for Alert destination's identifier. :param bool pause_subscriptions: flag that specifies if subscriptions are paused or not. """ - pulumi.set(__self__, "alert_id", alert_id) - pulumi.set(__self__, "subscriptions", subscriptions) + JobTaskSqlTaskAlert._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_id=alert_id, + subscriptions=subscriptions, + pause_subscriptions=pause_subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_id: str, + subscriptions: Sequence['outputs.JobTaskSqlTaskAlertSubscription'], + pause_subscriptions: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("alert_id", alert_id) + _setter("subscriptions", subscriptions) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) @property @pulumi.getter(name="alertId") @@ -8321,10 +10470,21 @@ def __init__(__self__, *, """ :param str user_name: The email of an active workspace user. Non-admin users can only set this field to their own email. """ + JobTaskSqlTaskAlertSubscription._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -8374,13 +10534,28 @@ def __init__(__self__, *, :param bool pause_subscriptions: flag that specifies if subscriptions are paused or not. :param Sequence['JobTaskSqlTaskDashboardSubscriptionArgs'] subscriptions: a list of subscription blocks consisting out of one of the required fields: `user_name` for user emails or `destination_id` - for Alert destination's identifier. """ - pulumi.set(__self__, "dashboard_id", dashboard_id) + JobTaskSqlTaskDashboard._configure( + lambda key, value: pulumi.set(__self__, key, value), + dashboard_id=dashboard_id, + custom_subject=custom_subject, + pause_subscriptions=pause_subscriptions, + subscriptions=subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dashboard_id: str, + custom_subject: Optional[str] = None, + pause_subscriptions: Optional[bool] = None, + subscriptions: Optional[Sequence['outputs.JobTaskSqlTaskDashboardSubscription']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dashboard_id", dashboard_id) if custom_subject is not None: - pulumi.set(__self__, "custom_subject", custom_subject) + _setter("custom_subject", custom_subject) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) if subscriptions is not None: - pulumi.set(__self__, "subscriptions", subscriptions) + _setter("subscriptions", subscriptions) @property @pulumi.getter(name="dashboardId") @@ -8442,10 +10617,21 @@ def __init__(__self__, *, """ :param str user_name: The email of an active workspace user. Non-admin users can only set this field to their own email. """ + JobTaskSqlTaskDashboardSubscription._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -8465,7 +10651,16 @@ def user_name(self) -> Optional[str]: class JobTaskSqlTaskFile(dict): def __init__(__self__, *, path: str): - pulumi.set(__self__, "path", path) + JobTaskSqlTaskFile._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -8494,7 +10689,16 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, query_id: str): - pulumi.set(__self__, "query_id", query_id) + JobTaskSqlTaskQuery._configure( + lambda key, value: pulumi.set(__self__, key, value), + query_id=query_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + query_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("query_id", query_id) @property @pulumi.getter(name="queryId") @@ -8530,9 +10734,20 @@ def __init__(__self__, *, :param 'JobTriggerFileArrivalArgs' file_arrival: configuration block to define a trigger for [File Arrival events](https://learn.microsoft.com/en-us/azure/databricks/workflows/jobs/file-arrival-triggers) consisting of following attributes: :param str pause_status: Indicate whether this trigger is paused or not. Either `PAUSED` or `UNPAUSED`. When the `pause_status` field is omitted in the block, the server will default to using `UNPAUSED` as a value for `pause_status`. """ - pulumi.set(__self__, "file_arrival", file_arrival) + JobTrigger._configure( + lambda key, value: pulumi.set(__self__, key, value), + file_arrival=file_arrival, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + file_arrival: 'outputs.JobTriggerFileArrival', + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("file_arrival", file_arrival) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="fileArrival") @@ -8577,21 +10792,34 @@ def __init__(__self__, *, min_time_between_triggers_seconds: Optional[int] = None, wait_after_last_change_seconds: Optional[int] = None): """ - :param str url: string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + :param str url: URL of the job on the given workspace :param int min_time_between_triggers_seconds: If set, the trigger starts a run only after the specified amount of time passed since the last time the trigger fired. The minimum allowed value is 60 seconds. :param int wait_after_last_change_seconds: If set, the trigger starts a run only after no file activity has occurred for the specified amount of time. This makes it possible to wait for a batch of incoming files to arrive before triggering a run. The minimum allowed value is 60 seconds. """ - pulumi.set(__self__, "url", url) + JobTriggerFileArrival._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + min_time_between_triggers_seconds=min_time_between_triggers_seconds, + wait_after_last_change_seconds=wait_after_last_change_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + min_time_between_triggers_seconds: Optional[int] = None, + wait_after_last_change_seconds: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if min_time_between_triggers_seconds is not None: - pulumi.set(__self__, "min_time_between_triggers_seconds", min_time_between_triggers_seconds) + _setter("min_time_between_triggers_seconds", min_time_between_triggers_seconds) if wait_after_last_change_seconds is not None: - pulumi.set(__self__, "wait_after_last_change_seconds", wait_after_last_change_seconds) + _setter("wait_after_last_change_seconds", wait_after_last_change_seconds) @property @pulumi.getter def url(self) -> str: """ - string with URL under the Unity Catalog external location that will be monitored for new files. Please note that have a trailing slash character (`/`). + URL of the job on the given workspace """ return pulumi.get(self, "url") @@ -8656,14 +10884,29 @@ def __init__(__self__, *, :param Sequence['JobWebhookNotificationsOnStartArgs'] on_starts: (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified. :param Sequence['JobWebhookNotificationsOnSuccessArgs'] on_successes: (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified. """ + JobWebhookNotifications._configure( + lambda key, value: pulumi.set(__self__, key, value), + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + on_duration_warning_threshold_exceededs: Optional[Sequence['outputs.JobWebhookNotificationsOnDurationWarningThresholdExceeded']] = None, + on_failures: Optional[Sequence['outputs.JobWebhookNotificationsOnFailure']] = None, + on_starts: Optional[Sequence['outputs.JobWebhookNotificationsOnStart']] = None, + on_successes: Optional[Sequence['outputs.JobWebhookNotificationsOnSuccess']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="onDurationWarningThresholdExceededs") @@ -8711,19 +10954,24 @@ class JobWebhookNotificationsOnDurationWarningThresholdExceeded(dict): def __init__(__self__, *, id: str): """ - :param str id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param str id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnDurationWarningThresholdExceeded._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> str: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -8733,19 +10981,24 @@ class JobWebhookNotificationsOnFailure(dict): def __init__(__self__, *, id: str): """ - :param str id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param str id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnFailure._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> str: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -8755,19 +11008,24 @@ class JobWebhookNotificationsOnStart(dict): def __init__(__self__, *, id: str): """ - :param str id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param str id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnStart._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> str: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -8777,19 +11035,24 @@ class JobWebhookNotificationsOnSuccess(dict): def __init__(__self__, *, id: str): """ - :param str id: ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + :param str id: ID of the job """ - pulumi.set(__self__, "id", id) + JobWebhookNotificationsOnSuccess._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter def id(self) -> str: """ - ID of the system notification that is notified when an event defined in `webhook_notifications` is triggered. - - > **Note** The following configuration blocks can be standalone or nested inside a `task` block + ID of the job """ return pulumi.get(self, "id") @@ -8799,9 +11062,20 @@ class LibraryCran(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + LibraryCran._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -8820,11 +11094,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + LibraryMaven._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -8847,9 +11134,20 @@ class LibraryPypi(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + LibraryPypi._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -8888,7 +11186,16 @@ def __init__(__self__, *, `azure_managed_identity` optional configuration block for using managed identity as credential details for Azure (Recommended): """ - pulumi.set(__self__, "role_arn", role_arn) + MetastoreDataAccessAwsIamRole._configure( + lambda key, value: pulumi.set(__self__, key, value), + role_arn=role_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role_arn: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role_arn", role_arn) @property @pulumi.getter(name="roleArn") @@ -8934,11 +11241,24 @@ def __init__(__self__, *, `databricks_gcp_service_account` optional configuration block for creating a Databricks-managed GCP Service Account: """ - pulumi.set(__self__, "access_connector_id", access_connector_id) + MetastoreDataAccessAzureManagedIdentity._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_connector_id=access_connector_id, + credential_id=credential_id, + managed_identity_id=managed_identity_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_connector_id: str, + credential_id: Optional[str] = None, + managed_identity_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("access_connector_id", access_connector_id) if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if managed_identity_id is not None: - pulumi.set(__self__, "managed_identity_id", managed_identity_id) + _setter("managed_identity_id", managed_identity_id) @property @pulumi.getter(name="accessConnectorId") @@ -8996,9 +11316,22 @@ def __init__(__self__, *, :param str client_secret: The client secret generated for the above app ID in AAD. **This field is redacted on output** :param str directory_id: The directory ID corresponding to the Azure Active Directory (AAD) tenant of the application """ - pulumi.set(__self__, "application_id", application_id) - pulumi.set(__self__, "client_secret", client_secret) - pulumi.set(__self__, "directory_id", directory_id) + MetastoreDataAccessAzureServicePrincipal._configure( + lambda key, value: pulumi.set(__self__, key, value), + application_id=application_id, + client_secret=client_secret, + directory_id=directory_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + application_id: str, + client_secret: str, + directory_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("application_id", application_id) + _setter("client_secret", client_secret) + _setter("directory_id", directory_id) @property @pulumi.getter(name="applicationId") @@ -9052,10 +11385,21 @@ def __init__(__self__, *, `azure_service_principal` optional configuration block for credential details for Azure (Legacy): """ + MetastoreDataAccessDatabricksGcpServiceAccount._configure( + lambda key, value: pulumi.set(__self__, key, value), + credential_id=credential_id, + email=email, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + credential_id: Optional[str] = None, + email: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if email is not None: - pulumi.set(__self__, "email", email) + _setter("email", email) @property @pulumi.getter(name="credentialId") @@ -9103,9 +11447,22 @@ def __init__(__self__, *, `azure_service_principal` optional configuration block for credential details for Azure (Legacy): """ - pulumi.set(__self__, "email", email) - pulumi.set(__self__, "private_key", private_key) - pulumi.set(__self__, "private_key_id", private_key_id) + MetastoreDataAccessGcpServiceAccountKey._configure( + lambda key, value: pulumi.set(__self__, key, value), + email=email, + private_key=private_key, + private_key_id=private_key_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + email: str, + private_key: str, + private_key_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("email", email) + _setter("private_key", private_key) + _setter("private_key_id", private_key_id) @property @pulumi.getter @@ -9133,10 +11490,21 @@ class MlflowModelTag(dict): def __init__(__self__, *, key: Optional[str] = None, value: Optional[str] = None): + MlflowModelTag._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: Optional[str] = None, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if key is not None: - pulumi.set(__self__, "key", key) + _setter("key", key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -9179,13 +11547,28 @@ def __init__(__self__, *, :param bool enable_ssl_verification: Enable/disable SSL certificate validation. Default is `true`. For self-signed certificates, this field must be `false` AND the destination server must disable certificate validation as well. For security purposes, it is encouraged to perform secret validation with the HMAC-encoded portion of the payload and acknowledge the risk associated with disabling hostname validation whereby it becomes more likely that requests can be maliciously routed to an unintended host. :param str secret: Shared secret required for HMAC encoding payload. The HMAC-encoded payload will be sent in the header as `X-Databricks-Signature: encoded_payload`. """ - pulumi.set(__self__, "url", url) + MlflowWebhookHttpUrlSpec._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + authorization=authorization, + enable_ssl_verification=enable_ssl_verification, + secret=secret, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + authorization: Optional[str] = None, + enable_ssl_verification: Optional[bool] = None, + secret: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if authorization is not None: - pulumi.set(__self__, "authorization", authorization) + _setter("authorization", authorization) if enable_ssl_verification is not None: - pulumi.set(__self__, "enable_ssl_verification", enable_ssl_verification) + _setter("enable_ssl_verification", enable_ssl_verification) if secret is not None: - pulumi.set(__self__, "secret", secret) + _setter("secret", secret) @property @pulumi.getter @@ -9252,10 +11635,23 @@ def __init__(__self__, *, :param str job_id: ID of the Databricks job that the webhook runs. :param str workspace_url: URL of the workspace containing the job that this webhook runs. If not specified, the job’s workspace URL is assumed to be the same as the workspace where the webhook is created. """ - pulumi.set(__self__, "access_token", access_token) - pulumi.set(__self__, "job_id", job_id) + MlflowWebhookJobSpec._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_token=access_token, + job_id=job_id, + workspace_url=workspace_url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_token: str, + job_id: str, + workspace_url: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("access_token", access_token) + _setter("job_id", job_id) if workspace_url is not None: - pulumi.set(__self__, "workspace_url", workspace_url) + _setter("workspace_url", workspace_url) @property @pulumi.getter(name="accessToken") @@ -9310,9 +11706,20 @@ def __init__(__self__, *, :param Sequence['ModelServingConfigServedModelArgs'] served_models: Each block represents a served model for the endpoint to serve. A model serving endpoint can have up to 10 served models. :param 'ModelServingConfigTrafficConfigArgs' traffic_config: A single block represents the traffic split configuration amongst the served models. """ - pulumi.set(__self__, "served_models", served_models) + ModelServingConfig._configure( + lambda key, value: pulumi.set(__self__, key, value), + served_models=served_models, + traffic_config=traffic_config, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + served_models: Sequence['outputs.ModelServingConfigServedModel'], + traffic_config: Optional['outputs.ModelServingConfigTrafficConfig'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("served_models", served_models) if traffic_config is not None: - pulumi.set(__self__, "traffic_config", traffic_config) + _setter("traffic_config", traffic_config) @property @pulumi.getter(name="servedModels") @@ -9377,17 +11784,38 @@ def __init__(__self__, *, :param str name: The name of a served model. It must be unique across an endpoint. If not specified, this field will default to `modelname-modelversion`. A served model name can consist of alphanumeric characters, dashes, and underscores. :param bool scale_to_zero_enabled: Whether the compute resources for the served model should scale down to zero. If scale-to-zero is enabled, the lower bound of the provisioned concurrency for each workload size will be 0. The default value is `true`. """ - pulumi.set(__self__, "model_name", model_name) - pulumi.set(__self__, "model_version", model_version) - pulumi.set(__self__, "workload_size", workload_size) + ModelServingConfigServedModel._configure( + lambda key, value: pulumi.set(__self__, key, value), + model_name=model_name, + model_version=model_version, + workload_size=workload_size, + environment_vars=environment_vars, + instance_profile_arn=instance_profile_arn, + name=name, + scale_to_zero_enabled=scale_to_zero_enabled, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + model_name: str, + model_version: str, + workload_size: str, + environment_vars: Optional[Mapping[str, Any]] = None, + instance_profile_arn: Optional[str] = None, + name: Optional[str] = None, + scale_to_zero_enabled: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("model_name", model_name) + _setter("model_version", model_version) + _setter("workload_size", workload_size) if environment_vars is not None: - pulumi.set(__self__, "environment_vars", environment_vars) + _setter("environment_vars", environment_vars) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if scale_to_zero_enabled is not None: - pulumi.set(__self__, "scale_to_zero_enabled", scale_to_zero_enabled) + _setter("scale_to_zero_enabled", scale_to_zero_enabled) @property @pulumi.getter(name="modelName") @@ -9453,8 +11881,17 @@ def __init__(__self__, *, """ :param Sequence['ModelServingConfigTrafficConfigRouteArgs'] routes: Each block represents a route that defines traffic to each served model. Each `served_models` block needs to have a corresponding `routes` block """ + ModelServingConfigTrafficConfig._configure( + lambda key, value: pulumi.set(__self__, key, value), + routes=routes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + routes: Optional[Sequence['outputs.ModelServingConfigTrafficConfigRoute']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if routes is not None: - pulumi.set(__self__, "routes", routes) + _setter("routes", routes) @property @pulumi.getter @@ -9493,8 +11930,19 @@ def __init__(__self__, *, :param str served_model_name: The name of the served model this route configures traffic for. This needs to match the name of a `served_models` block :param int traffic_percentage: The percentage of endpoint traffic to send to this route. It must be an integer between 0 and 100 inclusive. """ - pulumi.set(__self__, "served_model_name", served_model_name) - pulumi.set(__self__, "traffic_percentage", traffic_percentage) + ModelServingConfigTrafficConfigRoute._configure( + lambda key, value: pulumi.set(__self__, key, value), + served_model_name=served_model_name, + traffic_percentage=traffic_percentage, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + served_model_name: str, + traffic_percentage: int, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("served_model_name", served_model_name) + _setter("traffic_percentage", traffic_percentage) @property @pulumi.getter(name="servedModelName") @@ -9518,9 +11966,20 @@ class ModelServingTag(dict): def __init__(__self__, *, key: str, value: Optional[str] = None): - pulumi.set(__self__, "key", key) + ModelServingTag._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: str, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key", key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -9573,18 +12032,41 @@ def __init__(__self__, *, directory: Optional[str] = None, storage_account_name: Optional[str] = None, tenant_id: Optional[str] = None): - pulumi.set(__self__, "client_id", client_id) - pulumi.set(__self__, "client_secret_key", client_secret_key) - pulumi.set(__self__, "client_secret_scope", client_secret_scope) - pulumi.set(__self__, "initialize_file_system", initialize_file_system) + MountAbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + client_id=client_id, + client_secret_key=client_secret_key, + client_secret_scope=client_secret_scope, + initialize_file_system=initialize_file_system, + container_name=container_name, + directory=directory, + storage_account_name=storage_account_name, + tenant_id=tenant_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + client_id: str, + client_secret_key: str, + client_secret_scope: str, + initialize_file_system: bool, + container_name: Optional[str] = None, + directory: Optional[str] = None, + storage_account_name: Optional[str] = None, + tenant_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("client_id", client_id) + _setter("client_secret_key", client_secret_key) + _setter("client_secret_scope", client_secret_scope) + _setter("initialize_file_system", initialize_file_system) if container_name is not None: - pulumi.set(__self__, "container_name", container_name) + _setter("container_name", container_name) if directory is not None: - pulumi.set(__self__, "directory", directory) + _setter("directory", directory) if storage_account_name is not None: - pulumi.set(__self__, "storage_account_name", storage_account_name) + _setter("storage_account_name", storage_account_name) if tenant_id is not None: - pulumi.set(__self__, "tenant_id", tenant_id) + _setter("tenant_id", tenant_id) @property @pulumi.getter(name="clientId") @@ -9664,17 +12146,38 @@ def __init__(__self__, *, spark_conf_prefix: Optional[str] = None, storage_resource_name: Optional[str] = None, tenant_id: Optional[str] = None): - pulumi.set(__self__, "client_id", client_id) - pulumi.set(__self__, "client_secret_key", client_secret_key) - pulumi.set(__self__, "client_secret_scope", client_secret_scope) + MountAdl._configure( + lambda key, value: pulumi.set(__self__, key, value), + client_id=client_id, + client_secret_key=client_secret_key, + client_secret_scope=client_secret_scope, + directory=directory, + spark_conf_prefix=spark_conf_prefix, + storage_resource_name=storage_resource_name, + tenant_id=tenant_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + client_id: str, + client_secret_key: str, + client_secret_scope: str, + directory: Optional[str] = None, + spark_conf_prefix: Optional[str] = None, + storage_resource_name: Optional[str] = None, + tenant_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("client_id", client_id) + _setter("client_secret_key", client_secret_key) + _setter("client_secret_scope", client_secret_scope) if directory is not None: - pulumi.set(__self__, "directory", directory) + _setter("directory", directory) if spark_conf_prefix is not None: - pulumi.set(__self__, "spark_conf_prefix", spark_conf_prefix) + _setter("spark_conf_prefix", spark_conf_prefix) if storage_resource_name is not None: - pulumi.set(__self__, "storage_resource_name", storage_resource_name) + _setter("storage_resource_name", storage_resource_name) if tenant_id is not None: - pulumi.set(__self__, "tenant_id", tenant_id) + _setter("tenant_id", tenant_id) @property @pulumi.getter(name="clientId") @@ -9736,9 +12239,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, bucket_name: str, service_account: Optional[str] = None): - pulumi.set(__self__, "bucket_name", bucket_name) + MountGs._configure( + lambda key, value: pulumi.set(__self__, key, value), + bucket_name=bucket_name, + service_account=service_account, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + bucket_name: str, + service_account: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("bucket_name", bucket_name) if service_account is not None: - pulumi.set(__self__, "service_account", service_account) + _setter("service_account", service_account) @property @pulumi.getter(name="bucketName") @@ -9775,9 +12289,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, bucket_name: str, instance_profile: Optional[str] = None): - pulumi.set(__self__, "bucket_name", bucket_name) + MountS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + bucket_name=bucket_name, + instance_profile=instance_profile, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + bucket_name: str, + instance_profile: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("bucket_name", bucket_name) if instance_profile is not None: - pulumi.set(__self__, "instance_profile", instance_profile) + _setter("instance_profile", instance_profile) @property @pulumi.getter(name="bucketName") @@ -9824,15 +12349,34 @@ def __init__(__self__, *, container_name: Optional[str] = None, directory: Optional[str] = None, storage_account_name: Optional[str] = None): - pulumi.set(__self__, "auth_type", auth_type) - pulumi.set(__self__, "token_secret_key", token_secret_key) - pulumi.set(__self__, "token_secret_scope", token_secret_scope) + MountWasb._configure( + lambda key, value: pulumi.set(__self__, key, value), + auth_type=auth_type, + token_secret_key=token_secret_key, + token_secret_scope=token_secret_scope, + container_name=container_name, + directory=directory, + storage_account_name=storage_account_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + auth_type: str, + token_secret_key: str, + token_secret_scope: str, + container_name: Optional[str] = None, + directory: Optional[str] = None, + storage_account_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("auth_type", auth_type) + _setter("token_secret_key", token_secret_key) + _setter("token_secret_scope", token_secret_scope) if container_name is not None: - pulumi.set(__self__, "container_name", container_name) + _setter("container_name", container_name) if directory is not None: - pulumi.set(__self__, "directory", directory) + _setter("directory", directory) if storage_account_name is not None: - pulumi.set(__self__, "storage_account_name", storage_account_name) + _setter("storage_account_name", storage_account_name) @property @pulumi.getter(name="authType") @@ -9897,10 +12441,23 @@ def __init__(__self__, *, :param str key_arn: The AWS KMS key's Amazon Resource Name (ARN). :param str key_region: (Computed) The AWS region in which KMS key is deployed to. This is not required. """ - pulumi.set(__self__, "key_alias", key_alias) - pulumi.set(__self__, "key_arn", key_arn) + MwsCustomerManagedKeysAwsKeyInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + key_alias=key_alias, + key_arn=key_arn, + key_region=key_region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key_alias: str, + key_arn: str, + key_region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key_alias", key_alias) + _setter("key_arn", key_arn) if key_region is not None: - pulumi.set(__self__, "key_region", key_region) + _setter("key_region", key_region) @property @pulumi.getter(name="keyAlias") @@ -9951,7 +12508,16 @@ def __init__(__self__, *, """ :param str kms_key_id: The GCP KMS key's resource name. """ - pulumi.set(__self__, "kms_key_id", kms_key_id) + MwsCustomerManagedKeysGcpKeyInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + kms_key_id=kms_key_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + kms_key_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("kms_key_id", kms_key_id) @property @pulumi.getter(name="kmsKeyId") @@ -9986,10 +12552,21 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, error_message: Optional[str] = None, error_type: Optional[str] = None): + MwsNetworksErrorMessage._configure( + lambda key, value: pulumi.set(__self__, key, value), + error_message=error_message, + error_type=error_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + error_message: Optional[str] = None, + error_type: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if error_message is not None: - pulumi.set(__self__, "error_message", error_message) + _setter("error_message", error_message) if error_type is not None: - pulumi.set(__self__, "error_type", error_type) + _setter("error_type", error_type) @property @pulumi.getter(name="errorMessage") @@ -10046,12 +12623,31 @@ def __init__(__self__, *, :param str subnet_region: The Google Cloud region of the workspace data plane. For example, `us-east4`. :param str vpc_id: The ID of the VPC associated with this network. VPC IDs can be used in multiple network configurations. """ - pulumi.set(__self__, "network_project_id", network_project_id) - pulumi.set(__self__, "pod_ip_range_name", pod_ip_range_name) - pulumi.set(__self__, "service_ip_range_name", service_ip_range_name) - pulumi.set(__self__, "subnet_id", subnet_id) - pulumi.set(__self__, "subnet_region", subnet_region) - pulumi.set(__self__, "vpc_id", vpc_id) + MwsNetworksGcpNetworkInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + network_project_id=network_project_id, + pod_ip_range_name=pod_ip_range_name, + service_ip_range_name=service_ip_range_name, + subnet_id=subnet_id, + subnet_region=subnet_region, + vpc_id=vpc_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + network_project_id: str, + pod_ip_range_name: str, + service_ip_range_name: str, + subnet_id: str, + subnet_region: str, + vpc_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("network_project_id", network_project_id) + _setter("pod_ip_range_name", pod_ip_range_name) + _setter("service_ip_range_name", service_ip_range_name) + _setter("subnet_id", subnet_id) + _setter("subnet_region", subnet_region) + _setter("vpc_id", vpc_id) @property @pulumi.getter(name="networkProjectId") @@ -10126,8 +12722,19 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, dataplane_relays: Sequence[str], rest_apis: Sequence[str]): - pulumi.set(__self__, "dataplane_relays", dataplane_relays) - pulumi.set(__self__, "rest_apis", rest_apis) + MwsNetworksVpcEndpoints._configure( + lambda key, value: pulumi.set(__self__, key, value), + dataplane_relays=dataplane_relays, + rest_apis=rest_apis, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dataplane_relays: Sequence[str], + rest_apis: Sequence[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dataplane_relays", dataplane_relays) + _setter("rest_apis", rest_apis) @property @pulumi.getter(name="dataplaneRelays") @@ -10180,13 +12787,30 @@ def __init__(__self__, *, :param str psc_connection_id: The unique ID of this PSC connection. :param str service_attachment_id: The service attachment this PSC connection connects to. """ - pulumi.set(__self__, "endpoint_region", endpoint_region) - pulumi.set(__self__, "project_id", project_id) - pulumi.set(__self__, "psc_endpoint_name", psc_endpoint_name) + MwsVpcEndpointGcpVpcEndpointInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + endpoint_region=endpoint_region, + project_id=project_id, + psc_endpoint_name=psc_endpoint_name, + psc_connection_id=psc_connection_id, + service_attachment_id=service_attachment_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + endpoint_region: str, + project_id: str, + psc_endpoint_name: str, + psc_connection_id: Optional[str] = None, + service_attachment_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("endpoint_region", endpoint_region) + _setter("project_id", project_id) + _setter("psc_endpoint_name", psc_endpoint_name) if psc_connection_id is not None: - pulumi.set(__self__, "psc_connection_id", psc_connection_id) + _setter("psc_connection_id", psc_connection_id) if service_attachment_id is not None: - pulumi.set(__self__, "service_attachment_id", service_attachment_id) + _setter("service_attachment_id", service_attachment_id) @property @pulumi.getter(name="endpointRegion") @@ -10236,7 +12860,16 @@ def __init__(__self__, *, """ :param 'MwsWorkspacesCloudResourceContainerGcpArgs' gcp: A block that consists of the following field: """ - pulumi.set(__self__, "gcp", gcp) + MwsWorkspacesCloudResourceContainer._configure( + lambda key, value: pulumi.set(__self__, key, value), + gcp=gcp, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gcp: 'outputs.MwsWorkspacesCloudResourceContainerGcp', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("gcp", gcp) @property @pulumi.getter @@ -10271,7 +12904,16 @@ def __init__(__self__, *, """ :param str project_id: The Google Cloud project ID, which the workspace uses to instantiate cloud resources for your workspace. """ - pulumi.set(__self__, "project_id", project_id) + MwsWorkspacesCloudResourceContainerGcp._configure( + lambda key, value: pulumi.set(__self__, key, value), + project_id=project_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + project_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("project_id", project_id) @property @pulumi.getter(name="projectId") @@ -10309,9 +12951,22 @@ def __init__(__self__, *, authoritative_user_email: str, authoritative_user_full_name: str, customer_name: str): - pulumi.set(__self__, "authoritative_user_email", authoritative_user_email) - pulumi.set(__self__, "authoritative_user_full_name", authoritative_user_full_name) - pulumi.set(__self__, "customer_name", customer_name) + MwsWorkspacesExternalCustomerInfo._configure( + lambda key, value: pulumi.set(__self__, key, value), + authoritative_user_email=authoritative_user_email, + authoritative_user_full_name=authoritative_user_full_name, + customer_name=customer_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + authoritative_user_email: str, + authoritative_user_full_name: str, + customer_name: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("authoritative_user_email", authoritative_user_email) + _setter("authoritative_user_full_name", authoritative_user_full_name) + _setter("customer_name", customer_name) @property @pulumi.getter(name="authoritativeUserEmail") @@ -10356,9 +13011,22 @@ def __init__(__self__, *, gke_cluster_pod_ip_range: str, gke_cluster_service_ip_range: str, subnet_cidr: str): - pulumi.set(__self__, "gke_cluster_pod_ip_range", gke_cluster_pod_ip_range) - pulumi.set(__self__, "gke_cluster_service_ip_range", gke_cluster_service_ip_range) - pulumi.set(__self__, "subnet_cidr", subnet_cidr) + MwsWorkspacesGcpManagedNetworkConfig._configure( + lambda key, value: pulumi.set(__self__, key, value), + gke_cluster_pod_ip_range=gke_cluster_pod_ip_range, + gke_cluster_service_ip_range=gke_cluster_service_ip_range, + subnet_cidr=subnet_cidr, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gke_cluster_pod_ip_range: str, + gke_cluster_service_ip_range: str, + subnet_cidr: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("gke_cluster_pod_ip_range", gke_cluster_pod_ip_range) + _setter("gke_cluster_service_ip_range", gke_cluster_service_ip_range) + _setter("subnet_cidr", subnet_cidr) @property @pulumi.getter(name="gkeClusterPodIpRange") @@ -10404,8 +13072,19 @@ def __init__(__self__, *, :param str connectivity_type: Specifies the network connectivity types for the GKE nodes and the GKE master network. Possible values are: `PRIVATE_NODE_PUBLIC_MASTER`, `PUBLIC_NODE_PUBLIC_MASTER`. :param str master_ip_range: The IP range from which to allocate GKE cluster master resources. This field will be ignored if GKE private cluster is not enabled. It must be exactly as big as `/28`. """ - pulumi.set(__self__, "connectivity_type", connectivity_type) - pulumi.set(__self__, "master_ip_range", master_ip_range) + MwsWorkspacesGkeConfig._configure( + lambda key, value: pulumi.set(__self__, key, value), + connectivity_type=connectivity_type, + master_ip_range=master_ip_range, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + connectivity_type: str, + master_ip_range: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("connectivity_type", connectivity_type) + _setter("master_ip_range", master_ip_range) @property @pulumi.getter(name="connectivityType") @@ -10455,14 +13134,29 @@ def __init__(__self__, *, """ :param int lifetime_seconds: Token expiry lifetime. By default its 2592000 (30 days). """ + MwsWorkspacesToken._configure( + lambda key, value: pulumi.set(__self__, key, value), + comment=comment, + lifetime_seconds=lifetime_seconds, + token_id=token_id, + token_value=token_value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + comment: Optional[str] = None, + lifetime_seconds: Optional[int] = None, + token_id: Optional[str] = None, + token_value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if lifetime_seconds is not None: - pulumi.set(__self__, "lifetime_seconds", lifetime_seconds) + _setter("lifetime_seconds", lifetime_seconds) if token_id is not None: - pulumi.set(__self__, "token_id", token_id) + _setter("token_id", token_id) if token_value is not None: - pulumi.set(__self__, "token_value", token_value) + _setter("token_value", token_value) @property @pulumi.getter @@ -10526,13 +13220,28 @@ def __init__(__self__, *, :param str service_principal_name: Application ID of the service_principal. :param str user_name: name of the user. """ - pulumi.set(__self__, "permission_level", permission_level) + PermissionsAccessControl._configure( + lambda key, value: pulumi.set(__self__, key, value), + permission_level=permission_level, + group_name=group_name, + service_principal_name=service_principal_name, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permission_level: str, + group_name: Optional[str] = None, + service_principal_name: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("permission_level", permission_level) if group_name is not None: - pulumi.set(__self__, "group_name", group_name) + _setter("group_name", group_name) if service_principal_name is not None: - pulumi.set(__self__, "service_principal_name", service_principal_name) + _setter("service_principal_name", service_principal_name) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="permissionLevel") @@ -10640,44 +13349,89 @@ def __init__(__self__, *, spark_conf: Optional[Mapping[str, Any]] = None, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None): + PipelineCluster._configure( + lambda key, value: pulumi.set(__self__, key, value), + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_log_conf=cluster_log_conf, + custom_tags=custom_tags, + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_local_disk_encryption=enable_local_disk_encryption, + gcp_attributes=gcp_attributes, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + label=label, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['outputs.PipelineClusterAutoscale'] = None, + aws_attributes: Optional['outputs.PipelineClusterAwsAttributes'] = None, + azure_attributes: Optional['outputs.PipelineClusterAzureAttributes'] = None, + cluster_log_conf: Optional['outputs.PipelineClusterClusterLogConf'] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + driver_instance_pool_id: Optional[str] = None, + driver_node_type_id: Optional[str] = None, + enable_local_disk_encryption: Optional[bool] = None, + gcp_attributes: Optional['outputs.PipelineClusterGcpAttributes'] = None, + init_scripts: Optional[Sequence['outputs.PipelineClusterInitScript']] = None, + instance_pool_id: Optional[str] = None, + label: Optional[str] = None, + node_type_id: Optional[str] = None, + num_workers: Optional[int] = None, + policy_id: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if driver_instance_pool_id is not None: - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_instance_pool_id", driver_instance_pool_id) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if label is not None: - pulumi.set(__self__, "label", label) + _setter("label", label) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) @property @pulumi.getter(name="applyPolicyDefaultValues") @@ -10800,12 +13554,25 @@ def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None, mode: Optional[str] = None): + PipelineClusterAutoscale._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + mode=mode, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + mode: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) if mode is not None: - pulumi.set(__self__, "mode", mode) + _setter("mode", mode) @property @pulumi.getter(name="maxWorkers") @@ -10863,22 +13630,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + PipelineClusterAwsAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -10946,12 +13736,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + PipelineClusterAzureAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -10974,10 +13777,21 @@ class PipelineClusterClusterLogConf(dict): def __init__(__self__, *, dbfs: Optional['outputs.PipelineClusterClusterLogConfDbfs'] = None, s3: Optional['outputs.PipelineClusterClusterLogConfS3'] = None): + PipelineClusterClusterLogConf._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.PipelineClusterClusterLogConfDbfs'] = None, + s3: Optional['outputs.PipelineClusterClusterLogConfS3'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -10994,7 +13808,16 @@ def s3(self) -> Optional['outputs.PipelineClusterClusterLogConfS3']: class PipelineClusterClusterLogConfDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + PipelineClusterClusterLogConfDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -11035,19 +13858,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + PipelineClusterClusterLogConfS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -11113,14 +13957,29 @@ def __init__(__self__, *, google_service_account: Optional[str] = None, local_ssd_count: Optional[int] = None, zone_id: Optional[str] = None): + PipelineClusterGcpAttributes._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -11153,20 +14012,41 @@ def __init__(__self__, *, s3: Optional['outputs.PipelineClusterInitScriptS3'] = None, volumes: Optional['outputs.PipelineClusterInitScriptVolumes'] = None, workspace: Optional['outputs.PipelineClusterInitScriptWorkspace'] = None): + PipelineClusterInitScript._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.PipelineClusterInitScriptAbfss'] = None, + dbfs: Optional['outputs.PipelineClusterInitScriptDbfs'] = None, + file: Optional['outputs.PipelineClusterInitScriptFile'] = None, + gcs: Optional['outputs.PipelineClusterInitScriptGcs'] = None, + s3: Optional['outputs.PipelineClusterInitScriptS3'] = None, + volumes: Optional['outputs.PipelineClusterInitScriptVolumes'] = None, + workspace: Optional['outputs.PipelineClusterInitScriptWorkspace'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -11211,8 +14091,17 @@ def workspace(self) -> Optional['outputs.PipelineClusterInitScriptWorkspace']: class PipelineClusterInitScriptAbfss(dict): def __init__(__self__, *, destination: Optional[str] = None): + PipelineClusterInitScriptAbfss._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11224,7 +14113,16 @@ def destination(self) -> Optional[str]: class PipelineClusterInitScriptDbfs(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + PipelineClusterInitScriptDbfs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -11236,8 +14134,17 @@ def destination(self) -> str: class PipelineClusterInitScriptFile(dict): def __init__(__self__, *, destination: Optional[str] = None): + PipelineClusterInitScriptFile._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11249,8 +14156,17 @@ def destination(self) -> Optional[str]: class PipelineClusterInitScriptGcs(dict): def __init__(__self__, *, destination: Optional[str] = None): + PipelineClusterInitScriptGcs._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11291,19 +14207,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + PipelineClusterInitScriptS3._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -11345,8 +14282,17 @@ def region(self) -> Optional[str]: class PipelineClusterInitScriptVolumes(dict): def __init__(__self__, *, destination: Optional[str] = None): + PipelineClusterInitScriptVolumes._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11358,8 +14304,17 @@ def destination(self) -> Optional[str]: class PipelineClusterInitScriptWorkspace(dict): def __init__(__self__, *, destination: Optional[str] = None): + PipelineClusterInitScriptWorkspace._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -11372,10 +14327,21 @@ class PipelineFilters(dict): def __init__(__self__, *, excludes: Optional[Sequence[str]] = None, includes: Optional[Sequence[str]] = None): + PipelineFilters._configure( + lambda key, value: pulumi.set(__self__, key, value), + excludes=excludes, + includes=includes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + excludes: Optional[Sequence[str]] = None, + includes: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if excludes is not None: - pulumi.set(__self__, "excludes", excludes) + _setter("excludes", excludes) if includes is not None: - pulumi.set(__self__, "includes", includes) + _setter("includes", includes) @property @pulumi.getter @@ -11396,16 +14362,33 @@ def __init__(__self__, *, maven: Optional['outputs.PipelineLibraryMaven'] = None, notebook: Optional['outputs.PipelineLibraryNotebook'] = None, whl: Optional[str] = None): + PipelineLibrary._configure( + lambda key, value: pulumi.set(__self__, key, value), + file=file, + jar=jar, + maven=maven, + notebook=notebook, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + file: Optional['outputs.PipelineLibraryFile'] = None, + jar: Optional[str] = None, + maven: Optional['outputs.PipelineLibraryMaven'] = None, + notebook: Optional['outputs.PipelineLibraryNotebook'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if notebook is not None: - pulumi.set(__self__, "notebook", notebook) + _setter("notebook", notebook) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -11437,7 +14420,16 @@ def whl(self) -> Optional[str]: class PipelineLibraryFile(dict): def __init__(__self__, *, path: str): - pulumi.set(__self__, "path", path) + PipelineLibraryFile._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -11451,11 +14443,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + PipelineLibraryMaven._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -11477,7 +14482,16 @@ def repo(self) -> Optional[str]: class PipelineLibraryNotebook(dict): def __init__(__self__, *, path: str): - pulumi.set(__self__, "path", path) + PipelineLibraryNotebook._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -11515,8 +14529,19 @@ def __init__(__self__, *, * `on-flow-failure` - a single data flow fails. :param Sequence[str] email_recipients: non-empty list of emails to notify. """ - pulumi.set(__self__, "alerts", alerts) - pulumi.set(__self__, "email_recipients", email_recipients) + PipelineNotification._configure( + lambda key, value: pulumi.set(__self__, key, value), + alerts=alerts, + email_recipients=email_recipients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alerts: Sequence[str], + email_recipients: Sequence[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("alerts", alerts) + _setter("email_recipients", email_recipients) @property @pulumi.getter @@ -11563,7 +14588,16 @@ def __init__(__self__, *, """ :param Sequence[str] allowed_ip_addresses: Allowed IP Addresses in CIDR notation. Limit of 100. """ - pulumi.set(__self__, "allowed_ip_addresses", allowed_ip_addresses) + RecipientIpAccessList._configure( + lambda key, value: pulumi.set(__self__, key, value), + allowed_ip_addresses=allowed_ip_addresses, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allowed_ip_addresses: Sequence[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allowed_ip_addresses", allowed_ip_addresses) @property @pulumi.getter(name="allowedIpAddresses") @@ -11620,20 +14654,41 @@ def __init__(__self__, *, :param int updated_at: Time at which this recipient Token was updated, in epoch milliseconds. :param str updated_by: Username of recipient Token updater. """ + RecipientToken._configure( + lambda key, value: pulumi.set(__self__, key, value), + activation_url=activation_url, + created_at=created_at, + created_by=created_by, + expiration_time=expiration_time, + id=id, + updated_at=updated_at, + updated_by=updated_by, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + activation_url: Optional[str] = None, + created_at: Optional[int] = None, + created_by: Optional[str] = None, + expiration_time: Optional[int] = None, + id: Optional[str] = None, + updated_at: Optional[int] = None, + updated_by: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if activation_url is not None: - pulumi.set(__self__, "activation_url", activation_url) + _setter("activation_url", activation_url) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if expiration_time is not None: - pulumi.set(__self__, "expiration_time", expiration_time) + _setter("expiration_time", expiration_time) if id is not None: - pulumi.set(__self__, "id", id) + _setter("id", id) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) if updated_by is not None: - pulumi.set(__self__, "updated_by", updated_by) + _setter("updated_by", updated_by) @property @pulumi.getter(name="activationUrl") @@ -11696,7 +14751,16 @@ def updated_by(self) -> Optional[str]: class RepoSparseCheckout(dict): def __init__(__self__, *, patterns: Sequence[str]): - pulumi.set(__self__, "patterns", patterns) + RepoSparseCheckout._configure( + lambda key, value: pulumi.set(__self__, key, value), + patterns=patterns, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + patterns: Sequence[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("patterns", patterns) @property @pulumi.getter @@ -11728,8 +14792,19 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, dns_name: str, resource_id: str): - pulumi.set(__self__, "dns_name", dns_name) - pulumi.set(__self__, "resource_id", resource_id) + SecretScopeKeyvaultMetadata._configure( + lambda key, value: pulumi.set(__self__, key, value), + dns_name=dns_name, + resource_id=resource_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dns_name: str, + resource_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dns_name", dns_name) + _setter("resource_id", resource_id) @property @pulumi.getter(name="dnsName") @@ -11797,26 +14872,55 @@ def __init__(__self__, *, :param int start_version: The start version associated with the object for cdf. This allows data providers to control the lowest object version that is accessible by clients. :param str status: Status of the object, one of: `ACTIVE`, `PERMISSION_DENIED`. """ - pulumi.set(__self__, "data_object_type", data_object_type) - pulumi.set(__self__, "name", name) + ShareObject._configure( + lambda key, value: pulumi.set(__self__, key, value), + data_object_type=data_object_type, + name=name, + added_at=added_at, + added_by=added_by, + cdf_enabled=cdf_enabled, + comment=comment, + history_data_sharing_status=history_data_sharing_status, + partitions=partitions, + shared_as=shared_as, + start_version=start_version, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + data_object_type: str, + name: str, + added_at: Optional[int] = None, + added_by: Optional[str] = None, + cdf_enabled: Optional[bool] = None, + comment: Optional[str] = None, + history_data_sharing_status: Optional[str] = None, + partitions: Optional[Sequence['outputs.ShareObjectPartition']] = None, + shared_as: Optional[str] = None, + start_version: Optional[int] = None, + status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("data_object_type", data_object_type) + _setter("name", name) if added_at is not None: - pulumi.set(__self__, "added_at", added_at) + _setter("added_at", added_at) if added_by is not None: - pulumi.set(__self__, "added_by", added_by) + _setter("added_by", added_by) if cdf_enabled is not None: - pulumi.set(__self__, "cdf_enabled", cdf_enabled) + _setter("cdf_enabled", cdf_enabled) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if history_data_sharing_status is not None: - pulumi.set(__self__, "history_data_sharing_status", history_data_sharing_status) + _setter("history_data_sharing_status", history_data_sharing_status) if partitions is not None: - pulumi.set(__self__, "partitions", partitions) + _setter("partitions", partitions) if shared_as is not None: - pulumi.set(__self__, "shared_as", shared_as) + _setter("shared_as", shared_as) if start_version is not None: - pulumi.set(__self__, "start_version", start_version) + _setter("start_version", start_version) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter(name="dataObjectType") @@ -11904,11 +15008,26 @@ def status(self) -> Optional[str]: class ShareObjectPartition(dict): def __init__(__self__, *, values: Sequence['outputs.ShareObjectPartitionValue']): - pulumi.set(__self__, "values", values) + """ + :param Sequence['ShareObjectPartitionValueArgs'] values: The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + """ + ShareObjectPartition._configure( + lambda key, value: pulumi.set(__self__, key, value), + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + values: Sequence['outputs.ShareObjectPartitionValue'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("values", values) @property @pulumi.getter def values(self) -> Sequence['outputs.ShareObjectPartitionValue']: + """ + The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. + """ return pulumi.get(self, "values") @@ -11942,12 +15061,27 @@ def __init__(__self__, *, :param str recipient_property_key: The key of a Delta Sharing recipient's property. For example `databricks-account-id`. When this field is set, field `value` can not be set. :param str value: The value of the partition column. When this value is not set, it means null value. When this field is set, field `recipient_property_key` can not be set. """ - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "op", op) + ShareObjectPartitionValue._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + op=op, + recipient_property_key=recipient_property_key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: str, + op: str, + recipient_property_key: Optional[str] = None, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("op", op) if recipient_property_key is not None: - pulumi.set(__self__, "recipient_property_key", recipient_property_key) + _setter("recipient_property_key", recipient_property_key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12018,15 +15152,34 @@ def __init__(__self__, *, :param str custom_subject: Custom subject of alert notification, if it exists. This includes email subject, Slack notification header, etc. See [Alerts API reference](https://docs.databricks.com/sql/user/alerts/index.html) for custom templating instructions. :param bool muted: Whether or not the alert is muted. If an alert is muted, it will not notify users and alert destinations when triggered. """ - pulumi.set(__self__, "column", column) - pulumi.set(__self__, "op", op) - pulumi.set(__self__, "value", value) + SqlAlertOptions._configure( + lambda key, value: pulumi.set(__self__, key, value), + column=column, + op=op, + value=value, + custom_body=custom_body, + custom_subject=custom_subject, + muted=muted, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + column: str, + op: str, + value: str, + custom_body: Optional[str] = None, + custom_subject: Optional[str] = None, + muted: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("column", column) + _setter("op", op) + _setter("value", value) if custom_body is not None: - pulumi.set(__self__, "custom_body", custom_body) + _setter("custom_body", custom_body) if custom_subject is not None: - pulumi.set(__self__, "custom_subject", custom_subject) + _setter("custom_subject", custom_subject) if muted is not None: - pulumi.set(__self__, "muted", muted) + _setter("muted", muted) @property @pulumi.getter @@ -12084,8 +15237,17 @@ def __init__(__self__, *, """ :param str name: Name of the Databricks SQL release channel. Possible values are: `CHANNEL_NAME_PREVIEW` and `CHANNEL_NAME_CURRENT`. Default is `CHANNEL_NAME_CURRENT`. """ + SqlEndpointChannel._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -12104,13 +15266,30 @@ def __init__(__self__, *, protocol: str, host: Optional[str] = None, hostname: Optional[str] = None): - pulumi.set(__self__, "path", path) - pulumi.set(__self__, "port", port) - pulumi.set(__self__, "protocol", protocol) + SqlEndpointOdbcParams._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + port=port, + protocol=protocol, + host=host, + hostname=hostname, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + port: int, + protocol: str, + host: Optional[str] = None, + hostname: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) + _setter("port", port) + _setter("protocol", protocol) if host is not None: - pulumi.set(__self__, "host", host) + _setter("host", host) if hostname is not None: - pulumi.set(__self__, "hostname", hostname) + _setter("hostname", hostname) @property @pulumi.getter @@ -12159,7 +15338,16 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, custom_tags: Sequence['outputs.SqlEndpointTagsCustomTag']): - pulumi.set(__self__, "custom_tags", custom_tags) + SqlEndpointTags._configure( + lambda key, value: pulumi.set(__self__, key, value), + custom_tags=custom_tags, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + custom_tags: Sequence['outputs.SqlEndpointTagsCustomTag'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("custom_tags", custom_tags) @property @pulumi.getter(name="customTags") @@ -12172,8 +15360,19 @@ class SqlEndpointTagsCustomTag(dict): def __init__(__self__, *, key: str, value: str): - pulumi.set(__self__, "key", key) - pulumi.set(__self__, "value", value) + SqlEndpointTagsCustomTag._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: str, + value: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key", key) + _setter("value", value) @property @pulumi.getter @@ -12197,8 +15396,19 @@ def __init__(__self__, *, [Available](https://docs.databricks.com/security/access-control/table-acls/object-privileges.html) privilege names are: """ - pulumi.set(__self__, "principal", principal) - pulumi.set(__self__, "privileges", privileges) + SqlPermissionsPrivilegeAssignment._configure( + lambda key, value: pulumi.set(__self__, key, value), + principal=principal, + privileges=privileges, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + principal: str, + privileges: Sequence[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("principal", principal) + _setter("privileges", privileges) @property @pulumi.getter @@ -12255,29 +15465,60 @@ def __init__(__self__, *, query: Optional['outputs.SqlQueryParameterQuery'] = None, text: Optional['outputs.SqlQueryParameterText'] = None, title: Optional[str] = None): - pulumi.set(__self__, "name", name) + SqlQueryParameter._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + date=date, + date_range=date_range, + datetime=datetime, + datetime_range=datetime_range, + datetimesec=datetimesec, + datetimesec_range=datetimesec_range, + enum=enum, + number=number, + query=query, + text=text, + title=title, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: str, + date: Optional['outputs.SqlQueryParameterDate'] = None, + date_range: Optional['outputs.SqlQueryParameterDateRange'] = None, + datetime: Optional['outputs.SqlQueryParameterDatetime'] = None, + datetime_range: Optional['outputs.SqlQueryParameterDatetimeRange'] = None, + datetimesec: Optional['outputs.SqlQueryParameterDatetimesec'] = None, + datetimesec_range: Optional['outputs.SqlQueryParameterDatetimesecRange'] = None, + enum: Optional['outputs.SqlQueryParameterEnum'] = None, + number: Optional['outputs.SqlQueryParameterNumber'] = None, + query: Optional['outputs.SqlQueryParameterQuery'] = None, + text: Optional['outputs.SqlQueryParameterText'] = None, + title: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) if date is not None: - pulumi.set(__self__, "date", date) + _setter("date", date) if date_range is not None: - pulumi.set(__self__, "date_range", date_range) + _setter("date_range", date_range) if datetime is not None: - pulumi.set(__self__, "datetime", datetime) + _setter("datetime", datetime) if datetime_range is not None: - pulumi.set(__self__, "datetime_range", datetime_range) + _setter("datetime_range", datetime_range) if datetimesec is not None: - pulumi.set(__self__, "datetimesec", datetimesec) + _setter("datetimesec", datetimesec) if datetimesec_range is not None: - pulumi.set(__self__, "datetimesec_range", datetimesec_range) + _setter("datetimesec_range", datetimesec_range) if enum is not None: - pulumi.set(__self__, "enum", enum) + _setter("enum", enum) if number is not None: - pulumi.set(__self__, "number", number) + _setter("number", number) if query is not None: - pulumi.set(__self__, "query", query) + _setter("query", query) if text is not None: - pulumi.set(__self__, "text", text) + _setter("text", text) if title is not None: - pulumi.set(__self__, "title", title) + _setter("title", title) @property @pulumi.getter @@ -12344,7 +15585,16 @@ def title(self) -> Optional[str]: class SqlQueryParameterDate(dict): def __init__(__self__, *, value: str): - pulumi.set(__self__, "value", value) + SqlQueryParameterDate._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12357,10 +15607,21 @@ class SqlQueryParameterDateRange(dict): def __init__(__self__, *, range: Optional['outputs.SqlQueryParameterDateRangeRange'] = None, value: Optional[str] = None): + SqlQueryParameterDateRange._configure( + lambda key, value: pulumi.set(__self__, key, value), + range=range, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + range: Optional['outputs.SqlQueryParameterDateRangeRange'] = None, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if range is not None: - pulumi.set(__self__, "range", range) + _setter("range", range) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12378,8 +15639,19 @@ class SqlQueryParameterDateRangeRange(dict): def __init__(__self__, *, end: str, start: str): - pulumi.set(__self__, "end", end) - pulumi.set(__self__, "start", start) + SqlQueryParameterDateRangeRange._configure( + lambda key, value: pulumi.set(__self__, key, value), + end=end, + start=start, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + end: str, + start: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("end", end) + _setter("start", start) @property @pulumi.getter @@ -12396,7 +15668,16 @@ def start(self) -> str: class SqlQueryParameterDatetime(dict): def __init__(__self__, *, value: str): - pulumi.set(__self__, "value", value) + SqlQueryParameterDatetime._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12409,10 +15690,21 @@ class SqlQueryParameterDatetimeRange(dict): def __init__(__self__, *, range: Optional['outputs.SqlQueryParameterDatetimeRangeRange'] = None, value: Optional[str] = None): + SqlQueryParameterDatetimeRange._configure( + lambda key, value: pulumi.set(__self__, key, value), + range=range, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + range: Optional['outputs.SqlQueryParameterDatetimeRangeRange'] = None, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if range is not None: - pulumi.set(__self__, "range", range) + _setter("range", range) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12430,8 +15722,19 @@ class SqlQueryParameterDatetimeRangeRange(dict): def __init__(__self__, *, end: str, start: str): - pulumi.set(__self__, "end", end) - pulumi.set(__self__, "start", start) + SqlQueryParameterDatetimeRangeRange._configure( + lambda key, value: pulumi.set(__self__, key, value), + end=end, + start=start, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + end: str, + start: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("end", end) + _setter("start", start) @property @pulumi.getter @@ -12448,7 +15751,16 @@ def start(self) -> str: class SqlQueryParameterDatetimesec(dict): def __init__(__self__, *, value: str): - pulumi.set(__self__, "value", value) + SqlQueryParameterDatetimesec._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12461,10 +15773,21 @@ class SqlQueryParameterDatetimesecRange(dict): def __init__(__self__, *, range: Optional['outputs.SqlQueryParameterDatetimesecRangeRange'] = None, value: Optional[str] = None): + SqlQueryParameterDatetimesecRange._configure( + lambda key, value: pulumi.set(__self__, key, value), + range=range, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + range: Optional['outputs.SqlQueryParameterDatetimesecRangeRange'] = None, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if range is not None: - pulumi.set(__self__, "range", range) + _setter("range", range) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -12482,8 +15805,19 @@ class SqlQueryParameterDatetimesecRangeRange(dict): def __init__(__self__, *, end: str, start: str): - pulumi.set(__self__, "end", end) - pulumi.set(__self__, "start", start) + SqlQueryParameterDatetimesecRangeRange._configure( + lambda key, value: pulumi.set(__self__, key, value), + end=end, + start=start, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + end: str, + start: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("end", end) + _setter("start", start) @property @pulumi.getter @@ -12503,13 +15837,28 @@ def __init__(__self__, *, multiple: Optional['outputs.SqlQueryParameterEnumMultiple'] = None, value: Optional[str] = None, values: Optional[Sequence[str]] = None): - pulumi.set(__self__, "options", options) + SqlQueryParameterEnum._configure( + lambda key, value: pulumi.set(__self__, key, value), + options=options, + multiple=multiple, + value=value, + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + options: Sequence[str], + multiple: Optional['outputs.SqlQueryParameterEnumMultiple'] = None, + value: Optional[str] = None, + values: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("options", options) if multiple is not None: - pulumi.set(__self__, "multiple", multiple) + _setter("multiple", multiple) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) if values is not None: - pulumi.set(__self__, "values", values) + _setter("values", values) @property @pulumi.getter @@ -12538,9 +15887,22 @@ def __init__(__self__, *, prefix: str, separator: str, suffix: str): - pulumi.set(__self__, "prefix", prefix) - pulumi.set(__self__, "separator", separator) - pulumi.set(__self__, "suffix", suffix) + SqlQueryParameterEnumMultiple._configure( + lambda key, value: pulumi.set(__self__, key, value), + prefix=prefix, + separator=separator, + suffix=suffix, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + prefix: str, + separator: str, + suffix: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("prefix", prefix) + _setter("separator", separator) + _setter("suffix", suffix) @property @pulumi.getter @@ -12562,7 +15924,16 @@ def suffix(self) -> str: class SqlQueryParameterNumber(dict): def __init__(__self__, *, value: float): - pulumi.set(__self__, "value", value) + SqlQueryParameterNumber._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: float, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12594,13 +15965,28 @@ def __init__(__self__, *, multiple: Optional['outputs.SqlQueryParameterQueryMultiple'] = None, value: Optional[str] = None, values: Optional[Sequence[str]] = None): - pulumi.set(__self__, "query_id", query_id) + SqlQueryParameterQuery._configure( + lambda key, value: pulumi.set(__self__, key, value), + query_id=query_id, + multiple=multiple, + value=value, + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + query_id: str, + multiple: Optional['outputs.SqlQueryParameterQueryMultiple'] = None, + value: Optional[str] = None, + values: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("query_id", query_id) if multiple is not None: - pulumi.set(__self__, "multiple", multiple) + _setter("multiple", multiple) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) if values is not None: - pulumi.set(__self__, "values", values) + _setter("values", values) @property @pulumi.getter(name="queryId") @@ -12629,9 +16015,22 @@ def __init__(__self__, *, prefix: str, separator: str, suffix: str): - pulumi.set(__self__, "prefix", prefix) - pulumi.set(__self__, "separator", separator) - pulumi.set(__self__, "suffix", suffix) + SqlQueryParameterQueryMultiple._configure( + lambda key, value: pulumi.set(__self__, key, value), + prefix=prefix, + separator=separator, + suffix=suffix, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + prefix: str, + separator: str, + suffix: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("prefix", prefix) + _setter("separator", separator) + _setter("suffix", suffix) @property @pulumi.getter @@ -12653,7 +16052,16 @@ def suffix(self) -> str: class SqlQueryParameterText(dict): def __init__(__self__, *, value: str): - pulumi.set(__self__, "value", value) + SqlQueryParameterText._configure( + lambda key, value: pulumi.set(__self__, key, value), + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + value: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("value", value) @property @pulumi.getter @@ -12667,12 +16075,25 @@ def __init__(__self__, *, continuous: Optional['outputs.SqlQueryScheduleContinuous'] = None, daily: Optional['outputs.SqlQueryScheduleDaily'] = None, weekly: Optional['outputs.SqlQueryScheduleWeekly'] = None): + SqlQuerySchedule._configure( + lambda key, value: pulumi.set(__self__, key, value), + continuous=continuous, + daily=daily, + weekly=weekly, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + continuous: Optional['outputs.SqlQueryScheduleContinuous'] = None, + daily: Optional['outputs.SqlQueryScheduleDaily'] = None, + weekly: Optional['outputs.SqlQueryScheduleWeekly'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if daily is not None: - pulumi.set(__self__, "daily", daily) + _setter("daily", daily) if weekly is not None: - pulumi.set(__self__, "weekly", weekly) + _setter("weekly", weekly) @property @pulumi.getter @@ -12714,9 +16135,20 @@ def get(self, key: str, default = None) -> Any: def __init__(__self__, *, interval_seconds: int, until_date: Optional[str] = None): - pulumi.set(__self__, "interval_seconds", interval_seconds) + SqlQueryScheduleContinuous._configure( + lambda key, value: pulumi.set(__self__, key, value), + interval_seconds=interval_seconds, + until_date=until_date, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + interval_seconds: int, + until_date: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("interval_seconds", interval_seconds) if until_date is not None: - pulumi.set(__self__, "until_date", until_date) + _setter("until_date", until_date) @property @pulumi.getter(name="intervalSeconds") @@ -12756,10 +16188,23 @@ def __init__(__self__, *, interval_days: int, time_of_day: str, until_date: Optional[str] = None): - pulumi.set(__self__, "interval_days", interval_days) - pulumi.set(__self__, "time_of_day", time_of_day) + SqlQueryScheduleDaily._configure( + lambda key, value: pulumi.set(__self__, key, value), + interval_days=interval_days, + time_of_day=time_of_day, + until_date=until_date, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + interval_days: int, + time_of_day: str, + until_date: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("interval_days", interval_days) + _setter("time_of_day", time_of_day) if until_date is not None: - pulumi.set(__self__, "until_date", until_date) + _setter("until_date", until_date) @property @pulumi.getter(name="intervalDays") @@ -12807,11 +16252,26 @@ def __init__(__self__, *, interval_weeks: int, time_of_day: str, until_date: Optional[str] = None): - pulumi.set(__self__, "day_of_week", day_of_week) - pulumi.set(__self__, "interval_weeks", interval_weeks) - pulumi.set(__self__, "time_of_day", time_of_day) + SqlQueryScheduleWeekly._configure( + lambda key, value: pulumi.set(__self__, key, value), + day_of_week=day_of_week, + interval_weeks=interval_weeks, + time_of_day=time_of_day, + until_date=until_date, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + day_of_week: str, + interval_weeks: int, + time_of_day: str, + until_date: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("day_of_week", day_of_week) + _setter("interval_weeks", interval_weeks) + _setter("time_of_day", time_of_day) if until_date is not None: - pulumi.set(__self__, "until_date", until_date) + _setter("until_date", until_date) @property @pulumi.getter(name="dayOfWeek") @@ -12847,13 +16307,28 @@ def __init__(__self__, *, :param bool nullable: Whether field is nullable (Default: `true`) :param str type: Column type spec (with metadata) as SQL text. Not supported for `VIEW` table_type. """ - pulumi.set(__self__, "name", name) + SqlTableColumn._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + comment=comment, + nullable=nullable, + type=type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: str, + comment: Optional[str] = None, + nullable: Optional[bool] = None, + type: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if nullable is not None: - pulumi.set(__self__, "nullable", nullable) + _setter("nullable", nullable) if type is not None: - pulumi.set(__self__, "type", type) + _setter("type", type) @property @pulumi.getter @@ -12914,16 +16389,35 @@ def __init__(__self__, *, title: Optional[str] = None, value: Optional[str] = None, values: Optional[Sequence[str]] = None): - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "type", type) + SqlWidgetParameter._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + type=type, + map_to=map_to, + title=title, + value=value, + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: str, + type: str, + map_to: Optional[str] = None, + title: Optional[str] = None, + value: Optional[str] = None, + values: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("type", type) if map_to is not None: - pulumi.set(__self__, "map_to", map_to) + _setter("map_to", map_to) if title is not None: - pulumi.set(__self__, "title", title) + _setter("title", title) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) if values is not None: - pulumi.set(__self__, "values", values) + _setter("values", values) @property @pulumi.getter @@ -12989,14 +16483,31 @@ def __init__(__self__, *, auto_height: Optional[bool] = None, pos_x: Optional[int] = None, pos_y: Optional[int] = None): - pulumi.set(__self__, "size_x", size_x) - pulumi.set(__self__, "size_y", size_y) + SqlWidgetPosition._configure( + lambda key, value: pulumi.set(__self__, key, value), + size_x=size_x, + size_y=size_y, + auto_height=auto_height, + pos_x=pos_x, + pos_y=pos_y, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + size_x: int, + size_y: int, + auto_height: Optional[bool] = None, + pos_x: Optional[int] = None, + pos_y: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("size_x", size_x) + _setter("size_y", size_y) if auto_height is not None: - pulumi.set(__self__, "auto_height", auto_height) + _setter("auto_height", auto_height) if pos_x is not None: - pulumi.set(__self__, "pos_x", pos_x) + _setter("pos_x", pos_x) if pos_y is not None: - pulumi.set(__self__, "pos_y", pos_y) + _setter("pos_y", pos_y) @property @pulumi.getter(name="sizeX") @@ -13050,7 +16561,16 @@ def __init__(__self__, *, `azure_managed_identity` optional configuration block for using managed identity as credential details for Azure (recommended over service principal): """ - pulumi.set(__self__, "role_arn", role_arn) + StorageCredentialAwsIamRole._configure( + lambda key, value: pulumi.set(__self__, key, value), + role_arn=role_arn, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role_arn: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role_arn", role_arn) @property @pulumi.getter(name="roleArn") @@ -13096,11 +16616,24 @@ def __init__(__self__, *, `databricks_gcp_service_account` optional configuration block for creating a Databricks-managed GCP Service Account: """ - pulumi.set(__self__, "access_connector_id", access_connector_id) + StorageCredentialAzureManagedIdentity._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_connector_id=access_connector_id, + credential_id=credential_id, + managed_identity_id=managed_identity_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_connector_id: str, + credential_id: Optional[str] = None, + managed_identity_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("access_connector_id", access_connector_id) if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if managed_identity_id is not None: - pulumi.set(__self__, "managed_identity_id", managed_identity_id) + _setter("managed_identity_id", managed_identity_id) @property @pulumi.getter(name="accessConnectorId") @@ -13158,9 +16691,22 @@ def __init__(__self__, *, :param str client_secret: The client secret generated for the above app ID in AAD. **This field is redacted on output** :param str directory_id: The directory ID corresponding to the Azure Active Directory (AAD) tenant of the application """ - pulumi.set(__self__, "application_id", application_id) - pulumi.set(__self__, "client_secret", client_secret) - pulumi.set(__self__, "directory_id", directory_id) + StorageCredentialAzureServicePrincipal._configure( + lambda key, value: pulumi.set(__self__, key, value), + application_id=application_id, + client_secret=client_secret, + directory_id=directory_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + application_id: str, + client_secret: str, + directory_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("application_id", application_id) + _setter("client_secret", client_secret) + _setter("directory_id", directory_id) @property @pulumi.getter(name="applicationId") @@ -13212,10 +16758,21 @@ def __init__(__self__, *, """ :param str email: The email of the GCP service account created, to be granted access to relevant buckets. """ + StorageCredentialDatabricksGcpServiceAccount._configure( + lambda key, value: pulumi.set(__self__, key, value), + credential_id=credential_id, + email=email, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + credential_id: Optional[str] = None, + email: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if credential_id is not None: - pulumi.set(__self__, "credential_id", credential_id) + _setter("credential_id", credential_id) if email is not None: - pulumi.set(__self__, "email", email) + _setter("email", email) @property @pulumi.getter(name="credentialId") @@ -13259,9 +16816,22 @@ def __init__(__self__, *, """ :param str email: The email of the GCP service account created, to be granted access to relevant buckets. """ - pulumi.set(__self__, "email", email) - pulumi.set(__self__, "private_key", private_key) - pulumi.set(__self__, "private_key_id", private_key_id) + StorageCredentialGcpServiceAccountKey._configure( + lambda key, value: pulumi.set(__self__, key, value), + email=email, + private_key=private_key, + private_key_id=private_key_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + email: str, + private_key: str, + private_key_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("email", email) + _setter("private_key", private_key) + _setter("private_key_id", private_key_id) @property @pulumi.getter @@ -13325,24 +16895,53 @@ def __init__(__self__, *, type_json: Optional[str] = None, type_precision: Optional[int] = None, type_scale: Optional[int] = None): - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "position", position) - pulumi.set(__self__, "type_name", type_name) - pulumi.set(__self__, "type_text", type_text) + TableColumn._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + position=position, + type_name=type_name, + type_text=type_text, + comment=comment, + nullable=nullable, + partition_index=partition_index, + type_interval_type=type_interval_type, + type_json=type_json, + type_precision=type_precision, + type_scale=type_scale, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: str, + position: int, + type_name: str, + type_text: str, + comment: Optional[str] = None, + nullable: Optional[bool] = None, + partition_index: Optional[int] = None, + type_interval_type: Optional[str] = None, + type_json: Optional[str] = None, + type_precision: Optional[int] = None, + type_scale: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("position", position) + _setter("type_name", type_name) + _setter("type_text", type_text) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if nullable is not None: - pulumi.set(__self__, "nullable", nullable) + _setter("nullable", nullable) if partition_index is not None: - pulumi.set(__self__, "partition_index", partition_index) + _setter("partition_index", partition_index) if type_interval_type is not None: - pulumi.set(__self__, "type_interval_type", type_interval_type) + _setter("type_interval_type", type_interval_type) if type_json is not None: - pulumi.set(__self__, "type_json", type_json) + _setter("type_json", type_json) if type_precision is not None: - pulumi.set(__self__, "type_precision", type_precision) + _setter("type_precision", type_precision) if type_scale is not None: - pulumi.set(__self__, "type_scale", type_scale) + _setter("type_scale", type_scale) @property @pulumi.getter @@ -13466,88 +17065,181 @@ def __init__(__self__, *, :param Mapping[str, Any] spark_env_vars: Map with environment variable key-value pairs to fine-tune Spark clusters. Key-value pairs of the form (X,Y) are exported (i.e., X='Y') while launching the driver and workers. :param Sequence[str] ssh_public_keys: SSH public key contents that will be added to each Spark node in this cluster. """ - pulumi.set(__self__, "default_tags", default_tags) - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "spark_version", spark_version) - pulumi.set(__self__, "state", state) + GetClusterClusterInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + default_tags=default_tags, + driver_instance_pool_id=driver_instance_pool_id, + spark_version=spark_version, + state=state, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_cores=cluster_cores, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_log_status=cluster_log_status, + cluster_memory_mb=cluster_memory_mb, + cluster_name=cluster_name, + cluster_source=cluster_source, + creator_user_name=creator_user_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + driver=driver, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + executors=executors, + gcp_attributes=gcp_attributes, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + jdbc_port=jdbc_port, + last_activity_time=last_activity_time, + last_state_loss_time=last_state_loss_time, + node_type_id=node_type_id, + num_workers=num_workers, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_context_id=spark_context_id, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + start_time=start_time, + state_message=state_message, + terminate_time=terminate_time, + termination_reason=termination_reason, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default_tags: Mapping[str, Any], + driver_instance_pool_id: str, + spark_version: str, + state: str, + autoscale: Optional['outputs.GetClusterClusterInfoAutoscaleResult'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['outputs.GetClusterClusterInfoAwsAttributesResult'] = None, + azure_attributes: Optional['outputs.GetClusterClusterInfoAzureAttributesResult'] = None, + cluster_cores: Optional[float] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['outputs.GetClusterClusterInfoClusterLogConfResult'] = None, + cluster_log_status: Optional['outputs.GetClusterClusterInfoClusterLogStatusResult'] = None, + cluster_memory_mb: Optional[int] = None, + cluster_name: Optional[str] = None, + cluster_source: Optional[str] = None, + creator_user_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['outputs.GetClusterClusterInfoDockerImageResult'] = None, + driver: Optional['outputs.GetClusterClusterInfoDriverResult'] = None, + driver_node_type_id: Optional[str] = None, + enable_elastic_disk: Optional[bool] = None, + enable_local_disk_encryption: Optional[bool] = None, + executors: Optional[Sequence['outputs.GetClusterClusterInfoExecutorResult']] = None, + gcp_attributes: Optional['outputs.GetClusterClusterInfoGcpAttributesResult'] = None, + init_scripts: Optional[Sequence['outputs.GetClusterClusterInfoInitScriptResult']] = None, + instance_pool_id: Optional[str] = None, + jdbc_port: Optional[int] = None, + last_activity_time: Optional[int] = None, + last_state_loss_time: Optional[int] = None, + node_type_id: Optional[str] = None, + num_workers: Optional[int] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_context_id: Optional[int] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + start_time: Optional[int] = None, + state_message: Optional[str] = None, + terminate_time: Optional[int] = None, + termination_reason: Optional['outputs.GetClusterClusterInfoTerminationReasonResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("default_tags", default_tags) + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("spark_version", spark_version) + _setter("state", state) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_cores is not None: - pulumi.set(__self__, "cluster_cores", cluster_cores) + _setter("cluster_cores", cluster_cores) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_log_status is not None: - pulumi.set(__self__, "cluster_log_status", cluster_log_status) + _setter("cluster_log_status", cluster_log_status) if cluster_memory_mb is not None: - pulumi.set(__self__, "cluster_memory_mb", cluster_memory_mb) + _setter("cluster_memory_mb", cluster_memory_mb) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if cluster_source is not None: - pulumi.set(__self__, "cluster_source", cluster_source) + _setter("cluster_source", cluster_source) if creator_user_name is not None: - pulumi.set(__self__, "creator_user_name", creator_user_name) + _setter("creator_user_name", creator_user_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if driver is not None: - pulumi.set(__self__, "driver", driver) + _setter("driver", driver) if driver_node_type_id is not None: - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) + _setter("driver_node_type_id", driver_node_type_id) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if enable_local_disk_encryption is not None: - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) if executors is not None: - pulumi.set(__self__, "executors", executors) + _setter("executors", executors) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if jdbc_port is not None: - pulumi.set(__self__, "jdbc_port", jdbc_port) + _setter("jdbc_port", jdbc_port) if last_activity_time is not None: - pulumi.set(__self__, "last_activity_time", last_activity_time) + _setter("last_activity_time", last_activity_time) if last_state_loss_time is not None: - pulumi.set(__self__, "last_state_loss_time", last_state_loss_time) + _setter("last_state_loss_time", last_state_loss_time) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if num_workers is not None: - pulumi.set(__self__, "num_workers", num_workers) + _setter("num_workers", num_workers) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_context_id is not None: - pulumi.set(__self__, "spark_context_id", spark_context_id) + _setter("spark_context_id", spark_context_id) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if start_time is not None: - pulumi.set(__self__, "start_time", start_time) + _setter("start_time", start_time) if state_message is not None: - pulumi.set(__self__, "state_message", state_message) + _setter("state_message", state_message) if terminate_time is not None: - pulumi.set(__self__, "terminate_time", terminate_time) + _setter("terminate_time", terminate_time) if termination_reason is not None: - pulumi.set(__self__, "termination_reason", termination_reason) + _setter("termination_reason", termination_reason) @property @pulumi.getter(name="defaultTags") @@ -13824,10 +17516,21 @@ class GetClusterClusterInfoAutoscaleResult(dict): def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetClusterClusterInfoAutoscaleResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -13851,22 +17554,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetClusterClusterInfoAwsAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -13915,12 +17641,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetClusterClusterInfoAzureAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -13943,10 +17682,21 @@ class GetClusterClusterInfoClusterLogConfResult(dict): def __init__(__self__, *, dbfs: Optional['outputs.GetClusterClusterInfoClusterLogConfDbfsResult'] = None, s3: Optional['outputs.GetClusterClusterInfoClusterLogConfS3Result'] = None): + GetClusterClusterInfoClusterLogConfResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.GetClusterClusterInfoClusterLogConfDbfsResult'] = None, + s3: Optional['outputs.GetClusterClusterInfoClusterLogConfS3Result'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -13963,7 +17713,16 @@ def s3(self) -> Optional['outputs.GetClusterClusterInfoClusterLogConfS3Result']: class GetClusterClusterInfoClusterLogConfDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoClusterLogConfDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -13981,19 +17740,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoClusterLogConfS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -14036,10 +17816,21 @@ class GetClusterClusterInfoClusterLogStatusResult(dict): def __init__(__self__, *, last_attempted: Optional[int] = None, last_exception: Optional[str] = None): + GetClusterClusterInfoClusterLogStatusResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + last_attempted=last_attempted, + last_exception=last_exception, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + last_attempted: Optional[int] = None, + last_exception: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if last_attempted is not None: - pulumi.set(__self__, "last_attempted", last_attempted) + _setter("last_attempted", last_attempted) if last_exception is not None: - pulumi.set(__self__, "last_exception", last_exception) + _setter("last_exception", last_exception) @property @pulumi.getter(name="lastAttempted") @@ -14057,9 +17848,20 @@ class GetClusterClusterInfoDockerImageResult(dict): def __init__(__self__, *, url: str, basic_auth: Optional['outputs.GetClusterClusterInfoDockerImageBasicAuthResult'] = None): - pulumi.set(__self__, "url", url) + GetClusterClusterInfoDockerImageResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.GetClusterClusterInfoDockerImageBasicAuthResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -14077,8 +17879,19 @@ class GetClusterClusterInfoDockerImageBasicAuthResult(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetClusterClusterInfoDockerImageBasicAuthResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -14101,20 +17914,41 @@ def __init__(__self__, *, private_ip: Optional[str] = None, public_dns: Optional[str] = None, start_timestamp: Optional[int] = None): + GetClusterClusterInfoDriverResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + host_private_ip=host_private_ip, + instance_id=instance_id, + node_aws_attributes=node_aws_attributes, + node_id=node_id, + private_ip=private_ip, + public_dns=public_dns, + start_timestamp=start_timestamp, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + host_private_ip: Optional[str] = None, + instance_id: Optional[str] = None, + node_aws_attributes: Optional['outputs.GetClusterClusterInfoDriverNodeAwsAttributesResult'] = None, + node_id: Optional[str] = None, + private_ip: Optional[str] = None, + public_dns: Optional[str] = None, + start_timestamp: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if host_private_ip is not None: - pulumi.set(__self__, "host_private_ip", host_private_ip) + _setter("host_private_ip", host_private_ip) if instance_id is not None: - pulumi.set(__self__, "instance_id", instance_id) + _setter("instance_id", instance_id) if node_aws_attributes is not None: - pulumi.set(__self__, "node_aws_attributes", node_aws_attributes) + _setter("node_aws_attributes", node_aws_attributes) if node_id is not None: - pulumi.set(__self__, "node_id", node_id) + _setter("node_id", node_id) if private_ip is not None: - pulumi.set(__self__, "private_ip", private_ip) + _setter("private_ip", private_ip) if public_dns is not None: - pulumi.set(__self__, "public_dns", public_dns) + _setter("public_dns", public_dns) if start_timestamp is not None: - pulumi.set(__self__, "start_timestamp", start_timestamp) + _setter("start_timestamp", start_timestamp) @property @pulumi.getter(name="hostPrivateIp") @@ -14156,8 +17990,17 @@ def start_timestamp(self) -> Optional[int]: class GetClusterClusterInfoDriverNodeAwsAttributesResult(dict): def __init__(__self__, *, is_spot: Optional[bool] = None): + GetClusterClusterInfoDriverNodeAwsAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + is_spot=is_spot, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + is_spot: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if is_spot is not None: - pulumi.set(__self__, "is_spot", is_spot) + _setter("is_spot", is_spot) @property @pulumi.getter(name="isSpot") @@ -14175,20 +18018,41 @@ def __init__(__self__, *, private_ip: Optional[str] = None, public_dns: Optional[str] = None, start_timestamp: Optional[int] = None): + GetClusterClusterInfoExecutorResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + host_private_ip=host_private_ip, + instance_id=instance_id, + node_aws_attributes=node_aws_attributes, + node_id=node_id, + private_ip=private_ip, + public_dns=public_dns, + start_timestamp=start_timestamp, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + host_private_ip: Optional[str] = None, + instance_id: Optional[str] = None, + node_aws_attributes: Optional['outputs.GetClusterClusterInfoExecutorNodeAwsAttributesResult'] = None, + node_id: Optional[str] = None, + private_ip: Optional[str] = None, + public_dns: Optional[str] = None, + start_timestamp: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if host_private_ip is not None: - pulumi.set(__self__, "host_private_ip", host_private_ip) + _setter("host_private_ip", host_private_ip) if instance_id is not None: - pulumi.set(__self__, "instance_id", instance_id) + _setter("instance_id", instance_id) if node_aws_attributes is not None: - pulumi.set(__self__, "node_aws_attributes", node_aws_attributes) + _setter("node_aws_attributes", node_aws_attributes) if node_id is not None: - pulumi.set(__self__, "node_id", node_id) + _setter("node_id", node_id) if private_ip is not None: - pulumi.set(__self__, "private_ip", private_ip) + _setter("private_ip", private_ip) if public_dns is not None: - pulumi.set(__self__, "public_dns", public_dns) + _setter("public_dns", public_dns) if start_timestamp is not None: - pulumi.set(__self__, "start_timestamp", start_timestamp) + _setter("start_timestamp", start_timestamp) @property @pulumi.getter(name="hostPrivateIp") @@ -14230,8 +18094,17 @@ def start_timestamp(self) -> Optional[int]: class GetClusterClusterInfoExecutorNodeAwsAttributesResult(dict): def __init__(__self__, *, is_spot: Optional[bool] = None): + GetClusterClusterInfoExecutorNodeAwsAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + is_spot=is_spot, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + is_spot: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if is_spot is not None: - pulumi.set(__self__, "is_spot", is_spot) + _setter("is_spot", is_spot) @property @pulumi.getter(name="isSpot") @@ -14248,18 +18121,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetClusterClusterInfoGcpAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -14302,20 +18194,41 @@ def __init__(__self__, *, s3: Optional['outputs.GetClusterClusterInfoInitScriptS3Result'] = None, volumes: Optional['outputs.GetClusterClusterInfoInitScriptVolumesResult'] = None, workspace: Optional['outputs.GetClusterClusterInfoInitScriptWorkspaceResult'] = None): + GetClusterClusterInfoInitScriptResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.GetClusterClusterInfoInitScriptAbfssResult'] = None, + dbfs: Optional['outputs.GetClusterClusterInfoInitScriptDbfsResult'] = None, + file: Optional['outputs.GetClusterClusterInfoInitScriptFileResult'] = None, + gcs: Optional['outputs.GetClusterClusterInfoInitScriptGcsResult'] = None, + s3: Optional['outputs.GetClusterClusterInfoInitScriptS3Result'] = None, + volumes: Optional['outputs.GetClusterClusterInfoInitScriptVolumesResult'] = None, + workspace: Optional['outputs.GetClusterClusterInfoInitScriptWorkspaceResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -14357,8 +18270,17 @@ def workspace(self) -> Optional['outputs.GetClusterClusterInfoInitScriptWorkspac class GetClusterClusterInfoInitScriptAbfssResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptAbfssResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -14370,7 +18292,16 @@ def destination(self) -> Optional[str]: class GetClusterClusterInfoInitScriptDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoInitScriptDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -14382,8 +18313,17 @@ def destination(self) -> str: class GetClusterClusterInfoInitScriptFileResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptFileResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -14395,8 +18335,17 @@ def destination(self) -> Optional[str]: class GetClusterClusterInfoInitScriptGcsResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptGcsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -14414,19 +18363,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetClusterClusterInfoInitScriptS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -14468,8 +18438,17 @@ def region(self) -> Optional[str]: class GetClusterClusterInfoInitScriptVolumesResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptVolumesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -14481,8 +18460,17 @@ def destination(self) -> Optional[str]: class GetClusterClusterInfoInitScriptWorkspaceResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetClusterClusterInfoInitScriptWorkspaceResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -14496,12 +18484,25 @@ def __init__(__self__, *, code: Optional[str] = None, parameters: Optional[Mapping[str, Any]] = None, type: Optional[str] = None): + GetClusterClusterInfoTerminationReasonResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + code=code, + parameters=parameters, + type=type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + code: Optional[str] = None, + parameters: Optional[Mapping[str, Any]] = None, + type: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if code is not None: - pulumi.set(__self__, "code", code) + _setter("code", code) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if type is not None: - pulumi.set(__self__, "type", type) + _setter("type", type) @property @pulumi.getter @@ -14527,10 +18528,21 @@ def __init__(__self__, *, """ :param str path: Path on DBFS for the file to perform listing """ + GetDbfsFilePathsPathListResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + file_size=file_size, + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + file_size: Optional[int] = None, + path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if file_size is not None: - pulumi.set(__self__, "file_size", file_size) + _setter("file_size", file_size) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) @property @pulumi.getter(name="fileSize") @@ -14567,38 +18579,81 @@ def __init__(__self__, *, preloaded_spark_versions: Optional[Sequence[str]] = None, state: Optional[str] = None, stats: Optional['outputs.GetInstancePoolPoolInfoStatsResult'] = None): - pulumi.set(__self__, "default_tags", default_tags) - pulumi.set(__self__, "idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) - pulumi.set(__self__, "instance_pool_id", instance_pool_id) - pulumi.set(__self__, "instance_pool_name", instance_pool_name) + GetInstancePoolPoolInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + default_tags=default_tags, + idle_instance_autotermination_minutes=idle_instance_autotermination_minutes, + instance_pool_id=instance_pool_id, + instance_pool_name=instance_pool_name, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + custom_tags=custom_tags, + disk_spec=disk_spec, + enable_elastic_disk=enable_elastic_disk, + gcp_attributes=gcp_attributes, + instance_pool_fleet_attributes=instance_pool_fleet_attributes, + max_capacity=max_capacity, + min_idle_instances=min_idle_instances, + node_type_id=node_type_id, + preloaded_docker_images=preloaded_docker_images, + preloaded_spark_versions=preloaded_spark_versions, + state=state, + stats=stats, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default_tags: Mapping[str, Any], + idle_instance_autotermination_minutes: int, + instance_pool_id: str, + instance_pool_name: str, + aws_attributes: Optional['outputs.GetInstancePoolPoolInfoAwsAttributesResult'] = None, + azure_attributes: Optional['outputs.GetInstancePoolPoolInfoAzureAttributesResult'] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + disk_spec: Optional['outputs.GetInstancePoolPoolInfoDiskSpecResult'] = None, + enable_elastic_disk: Optional[bool] = None, + gcp_attributes: Optional['outputs.GetInstancePoolPoolInfoGcpAttributesResult'] = None, + instance_pool_fleet_attributes: Optional[Sequence['outputs.GetInstancePoolPoolInfoInstancePoolFleetAttributeResult']] = None, + max_capacity: Optional[int] = None, + min_idle_instances: Optional[int] = None, + node_type_id: Optional[str] = None, + preloaded_docker_images: Optional[Sequence['outputs.GetInstancePoolPoolInfoPreloadedDockerImageResult']] = None, + preloaded_spark_versions: Optional[Sequence[str]] = None, + state: Optional[str] = None, + stats: Optional['outputs.GetInstancePoolPoolInfoStatsResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("default_tags", default_tags) + _setter("idle_instance_autotermination_minutes", idle_instance_autotermination_minutes) + _setter("instance_pool_id", instance_pool_id) + _setter("instance_pool_name", instance_pool_name) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if disk_spec is not None: - pulumi.set(__self__, "disk_spec", disk_spec) + _setter("disk_spec", disk_spec) if enable_elastic_disk is not None: - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) + _setter("enable_elastic_disk", enable_elastic_disk) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if instance_pool_fleet_attributes is not None: - pulumi.set(__self__, "instance_pool_fleet_attributes", instance_pool_fleet_attributes) + _setter("instance_pool_fleet_attributes", instance_pool_fleet_attributes) if max_capacity is not None: - pulumi.set(__self__, "max_capacity", max_capacity) + _setter("max_capacity", max_capacity) if min_idle_instances is not None: - pulumi.set(__self__, "min_idle_instances", min_idle_instances) + _setter("min_idle_instances", min_idle_instances) if node_type_id is not None: - pulumi.set(__self__, "node_type_id", node_type_id) + _setter("node_type_id", node_type_id) if preloaded_docker_images is not None: - pulumi.set(__self__, "preloaded_docker_images", preloaded_docker_images) + _setter("preloaded_docker_images", preloaded_docker_images) if preloaded_spark_versions is not None: - pulumi.set(__self__, "preloaded_spark_versions", preloaded_spark_versions) + _setter("preloaded_spark_versions", preloaded_spark_versions) if state is not None: - pulumi.set(__self__, "state", state) + _setter("state", state) if stats is not None: - pulumi.set(__self__, "stats", stats) + _setter("stats", stats) @property @pulumi.getter(name="defaultTags") @@ -14697,11 +18752,24 @@ def __init__(__self__, *, zone_id: str, availability: Optional[str] = None, spot_bid_price_percent: Optional[int] = None): - pulumi.set(__self__, "zone_id", zone_id) + GetInstancePoolPoolInfoAwsAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + zone_id=zone_id, + availability=availability, + spot_bid_price_percent=spot_bid_price_percent, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + zone_id: str, + availability: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("zone_id", zone_id) if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) @property @pulumi.getter(name="zoneId") @@ -14724,10 +18792,21 @@ class GetInstancePoolPoolInfoAzureAttributesResult(dict): def __init__(__self__, *, availability: Optional[str] = None, spot_bid_max_price: Optional[float] = None): + GetInstancePoolPoolInfoAzureAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -14746,12 +18825,25 @@ def __init__(__self__, *, disk_count: Optional[int] = None, disk_size: Optional[int] = None, disk_type: Optional['outputs.GetInstancePoolPoolInfoDiskSpecDiskTypeResult'] = None): + GetInstancePoolPoolInfoDiskSpecResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + disk_count=disk_count, + disk_size=disk_size, + disk_type=disk_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + disk_count: Optional[int] = None, + disk_size: Optional[int] = None, + disk_type: Optional['outputs.GetInstancePoolPoolInfoDiskSpecDiskTypeResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if disk_count is not None: - pulumi.set(__self__, "disk_count", disk_count) + _setter("disk_count", disk_count) if disk_size is not None: - pulumi.set(__self__, "disk_size", disk_size) + _setter("disk_size", disk_size) if disk_type is not None: - pulumi.set(__self__, "disk_type", disk_type) + _setter("disk_type", disk_type) @property @pulumi.getter(name="diskCount") @@ -14774,10 +18866,21 @@ class GetInstancePoolPoolInfoDiskSpecDiskTypeResult(dict): def __init__(__self__, *, azure_disk_volume_type: Optional[str] = None, ebs_volume_type: Optional[str] = None): + GetInstancePoolPoolInfoDiskSpecDiskTypeResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + azure_disk_volume_type=azure_disk_volume_type, + ebs_volume_type=ebs_volume_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + azure_disk_volume_type: Optional[str] = None, + ebs_volume_type: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if azure_disk_volume_type is not None: - pulumi.set(__self__, "azure_disk_volume_type", azure_disk_volume_type) + _setter("azure_disk_volume_type", azure_disk_volume_type) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) @property @pulumi.getter(name="azureDiskVolumeType") @@ -14795,10 +18898,21 @@ class GetInstancePoolPoolInfoGcpAttributesResult(dict): def __init__(__self__, *, gcp_availability: Optional[str] = None, local_ssd_count: Optional[int] = None): + GetInstancePoolPoolInfoGcpAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + gcp_availability=gcp_availability, + local_ssd_count=local_ssd_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + gcp_availability: Optional[str] = None, + local_ssd_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if gcp_availability is not None: - pulumi.set(__self__, "gcp_availability", gcp_availability) + _setter("gcp_availability", gcp_availability) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) @property @pulumi.getter(name="gcpAvailability") @@ -14817,11 +18931,24 @@ def __init__(__self__, *, launch_template_overrides: Sequence['outputs.GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideResult'], fleet_on_demand_option: Optional['outputs.GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionResult'] = None, fleet_spot_option: Optional['outputs.GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionResult'] = None): - pulumi.set(__self__, "launch_template_overrides", launch_template_overrides) + GetInstancePoolPoolInfoInstancePoolFleetAttributeResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + launch_template_overrides=launch_template_overrides, + fleet_on_demand_option=fleet_on_demand_option, + fleet_spot_option=fleet_spot_option, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + launch_template_overrides: Sequence['outputs.GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideResult'], + fleet_on_demand_option: Optional['outputs.GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionResult'] = None, + fleet_spot_option: Optional['outputs.GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("launch_template_overrides", launch_template_overrides) if fleet_on_demand_option is not None: - pulumi.set(__self__, "fleet_on_demand_option", fleet_on_demand_option) + _setter("fleet_on_demand_option", fleet_on_demand_option) if fleet_spot_option is not None: - pulumi.set(__self__, "fleet_spot_option", fleet_spot_option) + _setter("fleet_spot_option", fleet_spot_option) @property @pulumi.getter(name="launchTemplateOverrides") @@ -14844,9 +18971,20 @@ class GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionResult def __init__(__self__, *, allocation_strategy: str, instance_pools_to_use_count: Optional[int] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetOnDemandOptionResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: str, + instance_pools_to_use_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -14864,9 +19002,20 @@ class GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionResult(dic def __init__(__self__, *, allocation_strategy: str, instance_pools_to_use_count: Optional[int] = None): - pulumi.set(__self__, "allocation_strategy", allocation_strategy) + GetInstancePoolPoolInfoInstancePoolFleetAttributeFleetSpotOptionResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + allocation_strategy=allocation_strategy, + instance_pools_to_use_count=instance_pools_to_use_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allocation_strategy: str, + instance_pools_to_use_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("allocation_strategy", allocation_strategy) if instance_pools_to_use_count is not None: - pulumi.set(__self__, "instance_pools_to_use_count", instance_pools_to_use_count) + _setter("instance_pools_to_use_count", instance_pools_to_use_count) @property @pulumi.getter(name="allocationStrategy") @@ -14884,8 +19033,19 @@ class GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideRes def __init__(__self__, *, availability_zone: str, instance_type: str): - pulumi.set(__self__, "availability_zone", availability_zone) - pulumi.set(__self__, "instance_type", instance_type) + GetInstancePoolPoolInfoInstancePoolFleetAttributeLaunchTemplateOverrideResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability_zone=availability_zone, + instance_type=instance_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability_zone: str, + instance_type: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("availability_zone", availability_zone) + _setter("instance_type", instance_type) @property @pulumi.getter(name="availabilityZone") @@ -14903,9 +19063,20 @@ class GetInstancePoolPoolInfoPreloadedDockerImageResult(dict): def __init__(__self__, *, url: str, basic_auth: Optional['outputs.GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthResult'] = None): - pulumi.set(__self__, "url", url) + GetInstancePoolPoolInfoPreloadedDockerImageResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -14923,8 +19094,19 @@ class GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthResult(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetInstancePoolPoolInfoPreloadedDockerImageBasicAuthResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -14944,14 +19126,29 @@ def __init__(__self__, *, pending_idle_count: Optional[int] = None, pending_used_count: Optional[int] = None, used_count: Optional[int] = None): + GetInstancePoolPoolInfoStatsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + idle_count=idle_count, + pending_idle_count=pending_idle_count, + pending_used_count=pending_used_count, + used_count=used_count, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + idle_count: Optional[int] = None, + pending_idle_count: Optional[int] = None, + pending_used_count: Optional[int] = None, + used_count: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if idle_count is not None: - pulumi.set(__self__, "idle_count", idle_count) + _setter("idle_count", idle_count) if pending_idle_count is not None: - pulumi.set(__self__, "pending_idle_count", pending_idle_count) + _setter("pending_idle_count", pending_idle_count) if pending_used_count is not None: - pulumi.set(__self__, "pending_used_count", pending_used_count) + _setter("pending_used_count", pending_used_count) if used_count is not None: - pulumi.set(__self__, "used_count", used_count) + _setter("used_count", used_count) @property @pulumi.getter(name="idleCount") @@ -14982,15 +19179,32 @@ def __init__(__self__, *, creator_user_name: Optional[str] = None, job_id: Optional[int] = None, settings: Optional['outputs.GetJobJobSettingsSettingsResult'] = None): - pulumi.set(__self__, "run_as_user_name", run_as_user_name) + GetJobJobSettingsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + run_as_user_name=run_as_user_name, + created_time=created_time, + creator_user_name=creator_user_name, + job_id=job_id, + settings=settings, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + run_as_user_name: str, + created_time: Optional[int] = None, + creator_user_name: Optional[str] = None, + job_id: Optional[int] = None, + settings: Optional['outputs.GetJobJobSettingsSettingsResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("run_as_user_name", run_as_user_name) if created_time is not None: - pulumi.set(__self__, "created_time", created_time) + _setter("created_time", created_time) if creator_user_name is not None: - pulumi.set(__self__, "creator_user_name", creator_user_name) + _setter("creator_user_name", creator_user_name) if job_id is not None: - pulumi.set(__self__, "job_id", job_id) + _setter("job_id", job_id) if settings is not None: - pulumi.set(__self__, "settings", settings) + _setter("settings", settings) @property @pulumi.getter(name="runAsUserName") @@ -15057,71 +19271,144 @@ def __init__(__self__, *, """ :param str name: the job name of Job if the resource was matched by id. """ - pulumi.set(__self__, "format", format) + GetJobJobSettingsSettingsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + format=format, + computes=computes, + continuous=continuous, + dbt_task=dbt_task, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + git_source=git_source, + health=health, + job_clusters=job_clusters, + libraries=libraries, + max_concurrent_runs=max_concurrent_runs, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + name=name, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + parameters=parameters, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + queue=queue, + retry_on_timeout=retry_on_timeout, + run_as=run_as, + run_job_task=run_job_task, + schedule=schedule, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + tags=tags, + tasks=tasks, + timeout_seconds=timeout_seconds, + trigger=trigger, + webhook_notifications=webhook_notifications, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + format: str, + computes: Optional[Sequence['outputs.GetJobJobSettingsSettingsComputeResult']] = None, + continuous: Optional['outputs.GetJobJobSettingsSettingsContinuousResult'] = None, + dbt_task: Optional['outputs.GetJobJobSettingsSettingsDbtTaskResult'] = None, + email_notifications: Optional['outputs.GetJobJobSettingsSettingsEmailNotificationsResult'] = None, + existing_cluster_id: Optional[str] = None, + git_source: Optional['outputs.GetJobJobSettingsSettingsGitSourceResult'] = None, + health: Optional['outputs.GetJobJobSettingsSettingsHealthResult'] = None, + job_clusters: Optional[Sequence['outputs.GetJobJobSettingsSettingsJobClusterResult']] = None, + libraries: Optional[Sequence['outputs.GetJobJobSettingsSettingsLibraryResult']] = None, + max_concurrent_runs: Optional[int] = None, + max_retries: Optional[int] = None, + min_retry_interval_millis: Optional[int] = None, + name: Optional[str] = None, + new_cluster: Optional['outputs.GetJobJobSettingsSettingsNewClusterResult'] = None, + notebook_task: Optional['outputs.GetJobJobSettingsSettingsNotebookTaskResult'] = None, + notification_settings: Optional['outputs.GetJobJobSettingsSettingsNotificationSettingsResult'] = None, + parameters: Optional[Sequence['outputs.GetJobJobSettingsSettingsParameterResult']] = None, + pipeline_task: Optional['outputs.GetJobJobSettingsSettingsPipelineTaskResult'] = None, + python_wheel_task: Optional['outputs.GetJobJobSettingsSettingsPythonWheelTaskResult'] = None, + queue: Optional['outputs.GetJobJobSettingsSettingsQueueResult'] = None, + retry_on_timeout: Optional[bool] = None, + run_as: Optional['outputs.GetJobJobSettingsSettingsRunAsResult'] = None, + run_job_task: Optional['outputs.GetJobJobSettingsSettingsRunJobTaskResult'] = None, + schedule: Optional['outputs.GetJobJobSettingsSettingsScheduleResult'] = None, + spark_jar_task: Optional['outputs.GetJobJobSettingsSettingsSparkJarTaskResult'] = None, + spark_python_task: Optional['outputs.GetJobJobSettingsSettingsSparkPythonTaskResult'] = None, + spark_submit_task: Optional['outputs.GetJobJobSettingsSettingsSparkSubmitTaskResult'] = None, + tags: Optional[Mapping[str, Any]] = None, + tasks: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskResult']] = None, + timeout_seconds: Optional[int] = None, + trigger: Optional['outputs.GetJobJobSettingsSettingsTriggerResult'] = None, + webhook_notifications: Optional['outputs.GetJobJobSettingsSettingsWebhookNotificationsResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("format", format) if computes is not None: - pulumi.set(__self__, "computes", computes) + _setter("computes", computes) if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if git_source is not None: - pulumi.set(__self__, "git_source", git_source) + _setter("git_source", git_source) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_clusters is not None: - pulumi.set(__self__, "job_clusters", job_clusters) + _setter("job_clusters", job_clusters) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_concurrent_runs is not None: - pulumi.set(__self__, "max_concurrent_runs", max_concurrent_runs) + _setter("max_concurrent_runs", max_concurrent_runs) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if queue is not None: - pulumi.set(__self__, "queue", queue) + _setter("queue", queue) if retry_on_timeout is not None: - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + _setter("retry_on_timeout", retry_on_timeout) if run_as is not None: - pulumi.set(__self__, "run_as", run_as) + _setter("run_as", run_as) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if schedule is not None: - pulumi.set(__self__, "schedule", schedule) + _setter("schedule", schedule) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if tasks is not None: - pulumi.set(__self__, "tasks", tasks) + _setter("tasks", tasks) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) if trigger is not None: - pulumi.set(__self__, "trigger", trigger) + _setter("trigger", trigger) if webhook_notifications is not None: - pulumi.set(__self__, "webhook_notifications", webhook_notifications) + _setter("webhook_notifications", webhook_notifications) @property @pulumi.getter @@ -15297,10 +19584,21 @@ class GetJobJobSettingsSettingsComputeResult(dict): def __init__(__self__, *, compute_key: Optional[str] = None, spec: Optional['outputs.GetJobJobSettingsSettingsComputeSpecResult'] = None): + GetJobJobSettingsSettingsComputeResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + compute_key=compute_key, + spec=spec, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + compute_key: Optional[str] = None, + spec: Optional['outputs.GetJobJobSettingsSettingsComputeSpecResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if spec is not None: - pulumi.set(__self__, "spec", spec) + _setter("spec", spec) @property @pulumi.getter(name="computeKey") @@ -15317,8 +19615,17 @@ def spec(self) -> Optional['outputs.GetJobJobSettingsSettingsComputeSpecResult'] class GetJobJobSettingsSettingsComputeSpecResult(dict): def __init__(__self__, *, kind: Optional[str] = None): + GetJobJobSettingsSettingsComputeSpecResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + kind=kind, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + kind: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if kind is not None: - pulumi.set(__self__, "kind", kind) + _setter("kind", kind) @property @pulumi.getter @@ -15330,8 +19637,17 @@ def kind(self) -> Optional[str]: class GetJobJobSettingsSettingsContinuousResult(dict): def __init__(__self__, *, pause_status: Optional[str] = None): + GetJobJobSettingsSettingsContinuousResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="pauseStatus") @@ -15348,17 +19664,36 @@ def __init__(__self__, *, project_directory: Optional[str] = None, schema: Optional[str] = None, warehouse_id: Optional[str] = None): - pulumi.set(__self__, "commands", commands) + GetJobJobSettingsSettingsDbtTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: Sequence[str], + catalog: Optional[str] = None, + profiles_directory: Optional[str] = None, + project_directory: Optional[str] = None, + schema: Optional[str] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -15400,18 +19735,37 @@ def __init__(__self__, *, on_failures: Optional[Sequence[str]] = None, on_starts: Optional[Sequence[str]] = None, on_successes: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsEmailNotificationsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + on_duration_warning_threshold_exceededs: Optional[Sequence[str]] = None, + on_failures: Optional[Sequence[str]] = None, + on_starts: Optional[Sequence[str]] = None, + on_successes: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -15453,17 +19807,36 @@ def __init__(__self__, *, job_source: Optional['outputs.GetJobJobSettingsSettingsGitSourceJobSourceResult'] = None, provider: Optional[str] = None, tag: Optional[str] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsGitSourceResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + branch=branch, + commit=commit, + job_source=job_source, + provider=provider, + tag=tag, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + branch: Optional[str] = None, + commit: Optional[str] = None, + job_source: Optional['outputs.GetJobJobSettingsSettingsGitSourceJobSourceResult'] = None, + provider: Optional[str] = None, + tag: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if branch is not None: - pulumi.set(__self__, "branch", branch) + _setter("branch", branch) if commit is not None: - pulumi.set(__self__, "commit", commit) + _setter("commit", commit) if job_source is not None: - pulumi.set(__self__, "job_source", job_source) + _setter("job_source", job_source) if provider is not None: - pulumi.set(__self__, "provider", provider) + _setter("provider", provider) if tag is not None: - pulumi.set(__self__, "tag", tag) + _setter("tag", tag) @property @pulumi.getter @@ -15502,10 +19875,23 @@ def __init__(__self__, *, import_from_git_branch: str, job_config_path: str, dirty_state: Optional[str] = None): - pulumi.set(__self__, "import_from_git_branch", import_from_git_branch) - pulumi.set(__self__, "job_config_path", job_config_path) + GetJobJobSettingsSettingsGitSourceJobSourceResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + import_from_git_branch=import_from_git_branch, + job_config_path=job_config_path, + dirty_state=dirty_state, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + import_from_git_branch: str, + job_config_path: str, + dirty_state: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("import_from_git_branch", import_from_git_branch) + _setter("job_config_path", job_config_path) if dirty_state is not None: - pulumi.set(__self__, "dirty_state", dirty_state) + _setter("dirty_state", dirty_state) @property @pulumi.getter(name="importFromGitBranch") @@ -15527,7 +19913,16 @@ def dirty_state(self) -> Optional[str]: class GetJobJobSettingsSettingsHealthResult(dict): def __init__(__self__, *, rules: Sequence['outputs.GetJobJobSettingsSettingsHealthRuleResult']): - pulumi.set(__self__, "rules", rules) + GetJobJobSettingsSettingsHealthResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: Sequence['outputs.GetJobJobSettingsSettingsHealthRuleResult'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -15541,12 +19936,25 @@ def __init__(__self__, *, metric: Optional[str] = None, op: Optional[str] = None, value: Optional[int] = None): + GetJobJobSettingsSettingsHealthRuleResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[str] = None, + op: Optional[str] = None, + value: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -15569,10 +19977,21 @@ class GetJobJobSettingsSettingsJobClusterResult(dict): def __init__(__self__, *, job_cluster_key: Optional[str] = None, new_cluster: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult'] = None): + GetJobJobSettingsSettingsJobClusterResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_cluster_key=job_cluster_key, + new_cluster=new_cluster, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_cluster_key: Optional[str] = None, + new_cluster: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) @property @pulumi.getter(name="jobClusterKey") @@ -15618,59 +20037,126 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeResult'] = None): - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) - pulumi.set(__self__, "node_type_id", node_type_id) - pulumi.set(__self__, "num_workers", num_workers) - pulumi.set(__self__, "spark_version", spark_version) + GetJobJobSettingsSettingsJobClusterNewClusterResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + node_type_id=node_type_id, + num_workers=num_workers, + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + driver_instance_pool_id: str, + driver_node_type_id: str, + enable_elastic_disk: bool, + enable_local_disk_encryption: bool, + node_type_id: str, + num_workers: int, + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleResult'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesResult'] = None, + azure_attributes: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterAzureAttributesResult'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfResult'] = None, + cluster_mount_infos: Optional[Sequence['outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoResult']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterDockerImageResult'] = None, + gcp_attributes: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterGcpAttributesResult'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptResult']] = None, + instance_pool_id: Optional[str] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_node_type_id", driver_node_type_id) + _setter("enable_elastic_disk", enable_elastic_disk) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) + _setter("node_type_id", node_type_id) + _setter("num_workers", num_workers) + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="driverInstancePoolId") @@ -15828,10 +20314,21 @@ class GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleResult(dict): def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetJobJobSettingsSettingsJobClusterNewClusterAutoscaleResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -15855,22 +20352,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterAwsAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -15919,12 +20439,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetJobJobSettingsSettingsJobClusterNewClusterAzureAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -15947,10 +20480,21 @@ class GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfResult(dict): def __init__(__self__, *, dbfs: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsResult'] = None, s3: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfS3Result'] = None): + GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsResult'] = None, + s3: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfS3Result'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -15967,7 +20511,16 @@ def s3(self) -> Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterC class GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -15985,19 +20538,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterClusterLogConfS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -16041,10 +20615,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoResult', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'outputs.GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoResult', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -16067,9 +20654,20 @@ class GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesy def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + GetJobJobSettingsSettingsJobClusterNewClusterClusterMountInfoNetworkFilesystemInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -16087,9 +20685,20 @@ class GetJobJobSettingsSettingsJobClusterNewClusterDockerImageResult(dict): def __init__(__self__, *, url: str, basic_auth: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthResult'] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsJobClusterNewClusterDockerImageResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -16107,8 +20716,19 @@ class GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthResult(di def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetJobJobSettingsSettingsJobClusterNewClusterDockerImageBasicAuthResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -16130,18 +20750,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterGcpAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -16184,20 +20823,41 @@ def __init__(__self__, *, s3: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptS3Result'] = None, volumes: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesResult'] = None, workspace: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceResult'] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptAbfssResult'] = None, + dbfs: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptDbfsResult'] = None, + file: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptFileResult'] = None, + gcs: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptGcsResult'] = None, + s3: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptS3Result'] = None, + volumes: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesResult'] = None, + workspace: Optional['outputs.GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -16239,8 +20899,17 @@ def workspace(self) -> Optional['outputs.GetJobJobSettingsSettingsJobClusterNewC class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptAbfssResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptAbfssResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -16252,7 +20921,16 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -16264,8 +20942,17 @@ def destination(self) -> str: class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptFileResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptFileResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -16277,8 +20964,17 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptGcsResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptGcsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -16296,19 +20992,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -16350,8 +21067,17 @@ def region(self) -> Optional[str]: class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptVolumesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -16363,8 +21089,17 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsJobClusterNewClusterInitScriptWorkspaceResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -16376,7 +21111,16 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeResult(dict): def __init__(__self__, *, clients: 'outputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsResult'): - pulumi.set(__self__, "clients", clients) + GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'outputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsResult', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -16389,10 +21133,21 @@ class GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsResult(dic def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClientsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -16414,18 +21169,37 @@ def __init__(__self__, *, maven: Optional['outputs.GetJobJobSettingsSettingsLibraryMavenResult'] = None, pypi: Optional['outputs.GetJobJobSettingsSettingsLibraryPypiResult'] = None, whl: Optional[str] = None): + GetJobJobSettingsSettingsLibraryResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional['outputs.GetJobJobSettingsSettingsLibraryCranResult'] = None, + egg: Optional[str] = None, + jar: Optional[str] = None, + maven: Optional['outputs.GetJobJobSettingsSettingsLibraryMavenResult'] = None, + pypi: Optional['outputs.GetJobJobSettingsSettingsLibraryPypiResult'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -16463,9 +21237,20 @@ class GetJobJobSettingsSettingsLibraryCranResult(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsLibraryCranResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -16484,11 +21269,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + GetJobJobSettingsSettingsLibraryMavenResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -16511,9 +21309,20 @@ class GetJobJobSettingsSettingsLibraryPypiResult(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsLibraryPypiResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -16559,59 +21368,126 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['outputs.GetJobJobSettingsSettingsNewClusterWorkloadTypeResult'] = None): - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) - pulumi.set(__self__, "node_type_id", node_type_id) - pulumi.set(__self__, "num_workers", num_workers) - pulumi.set(__self__, "spark_version", spark_version) + GetJobJobSettingsSettingsNewClusterResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + node_type_id=node_type_id, + num_workers=num_workers, + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + driver_instance_pool_id: str, + driver_node_type_id: str, + enable_elastic_disk: bool, + enable_local_disk_encryption: bool, + node_type_id: str, + num_workers: int, + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['outputs.GetJobJobSettingsSettingsNewClusterAutoscaleResult'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['outputs.GetJobJobSettingsSettingsNewClusterAwsAttributesResult'] = None, + azure_attributes: Optional['outputs.GetJobJobSettingsSettingsNewClusterAzureAttributesResult'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['outputs.GetJobJobSettingsSettingsNewClusterClusterLogConfResult'] = None, + cluster_mount_infos: Optional[Sequence['outputs.GetJobJobSettingsSettingsNewClusterClusterMountInfoResult']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['outputs.GetJobJobSettingsSettingsNewClusterDockerImageResult'] = None, + gcp_attributes: Optional['outputs.GetJobJobSettingsSettingsNewClusterGcpAttributesResult'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['outputs.GetJobJobSettingsSettingsNewClusterInitScriptResult']] = None, + instance_pool_id: Optional[str] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['outputs.GetJobJobSettingsSettingsNewClusterWorkloadTypeResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_node_type_id", driver_node_type_id) + _setter("enable_elastic_disk", enable_elastic_disk) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) + _setter("node_type_id", node_type_id) + _setter("num_workers", num_workers) + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="driverInstancePoolId") @@ -16769,10 +21645,21 @@ class GetJobJobSettingsSettingsNewClusterAutoscaleResult(dict): def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetJobJobSettingsSettingsNewClusterAutoscaleResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -16796,22 +21683,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterAwsAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -16860,12 +21770,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetJobJobSettingsSettingsNewClusterAzureAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -16888,10 +21811,21 @@ class GetJobJobSettingsSettingsNewClusterClusterLogConfResult(dict): def __init__(__self__, *, dbfs: Optional['outputs.GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsResult'] = None, s3: Optional['outputs.GetJobJobSettingsSettingsNewClusterClusterLogConfS3Result'] = None): + GetJobJobSettingsSettingsNewClusterClusterLogConfResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsResult'] = None, + s3: Optional['outputs.GetJobJobSettingsSettingsNewClusterClusterLogConfS3Result'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -16908,7 +21842,16 @@ def s3(self) -> Optional['outputs.GetJobJobSettingsSettingsNewClusterClusterLogC class GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterClusterLogConfDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -16926,19 +21869,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterClusterLogConfS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -16982,10 +21946,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'outputs.GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoResult', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + GetJobJobSettingsSettingsNewClusterClusterMountInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'outputs.GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoResult', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -17008,9 +21985,20 @@ class GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoRe def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + GetJobJobSettingsSettingsNewClusterClusterMountInfoNetworkFilesystemInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -17028,9 +22016,20 @@ class GetJobJobSettingsSettingsNewClusterDockerImageResult(dict): def __init__(__self__, *, url: str, basic_auth: Optional['outputs.GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthResult'] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsNewClusterDockerImageResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -17048,8 +22047,19 @@ class GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthResult(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetJobJobSettingsSettingsNewClusterDockerImageBasicAuthResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -17071,18 +22081,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterGcpAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -17125,20 +22154,41 @@ def __init__(__self__, *, s3: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptS3Result'] = None, volumes: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptVolumesResult'] = None, workspace: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceResult'] = None): + GetJobJobSettingsSettingsNewClusterInitScriptResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptAbfssResult'] = None, + dbfs: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptDbfsResult'] = None, + file: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptFileResult'] = None, + gcs: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptGcsResult'] = None, + s3: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptS3Result'] = None, + volumes: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptVolumesResult'] = None, + workspace: Optional['outputs.GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -17180,8 +22230,17 @@ def workspace(self) -> Optional['outputs.GetJobJobSettingsSettingsNewClusterInit class GetJobJobSettingsSettingsNewClusterInitScriptAbfssResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptAbfssResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17193,7 +22252,16 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsNewClusterInitScriptDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterInitScriptDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -17205,8 +22273,17 @@ def destination(self) -> str: class GetJobJobSettingsSettingsNewClusterInitScriptFileResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptFileResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17218,8 +22295,17 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsNewClusterInitScriptGcsResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptGcsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17237,19 +22323,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsNewClusterInitScriptS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -17291,8 +22398,17 @@ def region(self) -> Optional[str]: class GetJobJobSettingsSettingsNewClusterInitScriptVolumesResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptVolumesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17304,8 +22420,17 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsNewClusterInitScriptWorkspaceResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -17317,7 +22442,16 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsNewClusterWorkloadTypeResult(dict): def __init__(__self__, *, clients: 'outputs.GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsResult'): - pulumi.set(__self__, "clients", clients) + GetJobJobSettingsSettingsNewClusterWorkloadTypeResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'outputs.GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsResult', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -17330,10 +22464,21 @@ class GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsResult(dict): def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + GetJobJobSettingsSettingsNewClusterWorkloadTypeClientsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -17352,11 +22497,24 @@ def __init__(__self__, *, notebook_path: str, base_parameters: Optional[Mapping[str, Any]] = None, source: Optional[str] = None): - pulumi.set(__self__, "notebook_path", notebook_path) + GetJobJobSettingsSettingsNotebookTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: str, + base_parameters: Optional[Mapping[str, Any]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -17379,10 +22537,21 @@ class GetJobJobSettingsSettingsNotificationSettingsResult(dict): def __init__(__self__, *, no_alert_for_canceled_runs: Optional[bool] = None, no_alert_for_skipped_runs: Optional[bool] = None): + GetJobJobSettingsSettingsNotificationSettingsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + no_alert_for_canceled_runs: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="noAlertForCanceledRuns") @@ -17403,10 +22572,21 @@ def __init__(__self__, *, """ :param str name: the job name of Job if the resource was matched by id. """ + GetJobJobSettingsSettingsParameterResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + default=default, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + default: Optional[str] = None, + name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if default is not None: - pulumi.set(__self__, "default", default) + _setter("default", default) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -17427,9 +22607,20 @@ class GetJobJobSettingsSettingsPipelineTaskResult(dict): def __init__(__self__, *, pipeline_id: str, full_refresh: Optional[bool] = None): - pulumi.set(__self__, "pipeline_id", pipeline_id) + GetJobJobSettingsSettingsPipelineTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: str, + full_refresh: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -17449,14 +22640,29 @@ def __init__(__self__, *, named_parameters: Optional[Mapping[str, Any]] = None, package_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsPythonWheelTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[str] = None, + named_parameters: Optional[Mapping[str, Any]] = None, + package_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -17483,6 +22689,11 @@ def parameters(self) -> Optional[Sequence[str]]: class GetJobJobSettingsSettingsQueueResult(dict): def __init__(__self__): pass + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + opts: Optional[pulumi.ResourceOptions]=None): + pass @pulumi.output_type @@ -17490,10 +22701,21 @@ class GetJobJobSettingsSettingsRunAsResult(dict): def __init__(__self__, *, service_principal_name: Optional[str] = None, user_name: Optional[str] = None): + GetJobJobSettingsSettingsRunAsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + service_principal_name=service_principal_name, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + service_principal_name: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if service_principal_name is not None: - pulumi.set(__self__, "service_principal_name", service_principal_name) + _setter("service_principal_name", service_principal_name) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="servicePrincipalName") @@ -17511,9 +22733,20 @@ class GetJobJobSettingsSettingsRunJobTaskResult(dict): def __init__(__self__, *, job_id: int, job_parameters: Optional[Mapping[str, Any]] = None): - pulumi.set(__self__, "job_id", job_id) + GetJobJobSettingsSettingsRunJobTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: int, + job_parameters: Optional[Mapping[str, Any]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -17532,10 +22765,23 @@ def __init__(__self__, *, quartz_cron_expression: str, timezone_id: str, pause_status: Optional[str] = None): - pulumi.set(__self__, "quartz_cron_expression", quartz_cron_expression) - pulumi.set(__self__, "timezone_id", timezone_id) + GetJobJobSettingsSettingsScheduleResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + quartz_cron_expression=quartz_cron_expression, + timezone_id=timezone_id, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + quartz_cron_expression: str, + timezone_id: str, + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("quartz_cron_expression", quartz_cron_expression) + _setter("timezone_id", timezone_id) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="quartzCronExpression") @@ -17559,12 +22805,25 @@ def __init__(__self__, *, jar_uri: Optional[str] = None, main_class_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsSparkJarTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[str] = None, + main_class_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -17588,11 +22847,24 @@ def __init__(__self__, *, python_file: str, parameters: Optional[Sequence[str]] = None, source: Optional[str] = None): - pulumi.set(__self__, "python_file", python_file) + GetJobJobSettingsSettingsSparkPythonTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: str, + parameters: Optional[Sequence[str]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -17614,8 +22886,17 @@ def source(self) -> Optional[str]: class GetJobJobSettingsSettingsSparkSubmitTaskResult(dict): def __init__(__self__, *, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsSparkSubmitTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -17652,57 +22933,116 @@ def __init__(__self__, *, sql_task: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskResult'] = None, task_key: Optional[str] = None, timeout_seconds: Optional[int] = None): - pulumi.set(__self__, "retry_on_timeout", retry_on_timeout) + GetJobJobSettingsSettingsTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + retry_on_timeout=retry_on_timeout, + compute_key=compute_key, + condition_task=condition_task, + dbt_task=dbt_task, + depends_ons=depends_ons, + description=description, + email_notifications=email_notifications, + existing_cluster_id=existing_cluster_id, + health=health, + job_cluster_key=job_cluster_key, + libraries=libraries, + max_retries=max_retries, + min_retry_interval_millis=min_retry_interval_millis, + new_cluster=new_cluster, + notebook_task=notebook_task, + notification_settings=notification_settings, + pipeline_task=pipeline_task, + python_wheel_task=python_wheel_task, + run_if=run_if, + run_job_task=run_job_task, + spark_jar_task=spark_jar_task, + spark_python_task=spark_python_task, + spark_submit_task=spark_submit_task, + sql_task=sql_task, + task_key=task_key, + timeout_seconds=timeout_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + retry_on_timeout: bool, + compute_key: Optional[str] = None, + condition_task: Optional['outputs.GetJobJobSettingsSettingsTaskConditionTaskResult'] = None, + dbt_task: Optional['outputs.GetJobJobSettingsSettingsTaskDbtTaskResult'] = None, + depends_ons: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskDependsOnResult']] = None, + description: Optional[str] = None, + email_notifications: Optional['outputs.GetJobJobSettingsSettingsTaskEmailNotificationsResult'] = None, + existing_cluster_id: Optional[str] = None, + health: Optional['outputs.GetJobJobSettingsSettingsTaskHealthResult'] = None, + job_cluster_key: Optional[str] = None, + libraries: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskLibraryResult']] = None, + max_retries: Optional[int] = None, + min_retry_interval_millis: Optional[int] = None, + new_cluster: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterResult'] = None, + notebook_task: Optional['outputs.GetJobJobSettingsSettingsTaskNotebookTaskResult'] = None, + notification_settings: Optional['outputs.GetJobJobSettingsSettingsTaskNotificationSettingsResult'] = None, + pipeline_task: Optional['outputs.GetJobJobSettingsSettingsTaskPipelineTaskResult'] = None, + python_wheel_task: Optional['outputs.GetJobJobSettingsSettingsTaskPythonWheelTaskResult'] = None, + run_if: Optional[str] = None, + run_job_task: Optional['outputs.GetJobJobSettingsSettingsTaskRunJobTaskResult'] = None, + spark_jar_task: Optional['outputs.GetJobJobSettingsSettingsTaskSparkJarTaskResult'] = None, + spark_python_task: Optional['outputs.GetJobJobSettingsSettingsTaskSparkPythonTaskResult'] = None, + spark_submit_task: Optional['outputs.GetJobJobSettingsSettingsTaskSparkSubmitTaskResult'] = None, + sql_task: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskResult'] = None, + task_key: Optional[str] = None, + timeout_seconds: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("retry_on_timeout", retry_on_timeout) if compute_key is not None: - pulumi.set(__self__, "compute_key", compute_key) + _setter("compute_key", compute_key) if condition_task is not None: - pulumi.set(__self__, "condition_task", condition_task) + _setter("condition_task", condition_task) if dbt_task is not None: - pulumi.set(__self__, "dbt_task", dbt_task) + _setter("dbt_task", dbt_task) if depends_ons is not None: - pulumi.set(__self__, "depends_ons", depends_ons) + _setter("depends_ons", depends_ons) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if email_notifications is not None: - pulumi.set(__self__, "email_notifications", email_notifications) + _setter("email_notifications", email_notifications) if existing_cluster_id is not None: - pulumi.set(__self__, "existing_cluster_id", existing_cluster_id) + _setter("existing_cluster_id", existing_cluster_id) if health is not None: - pulumi.set(__self__, "health", health) + _setter("health", health) if job_cluster_key is not None: - pulumi.set(__self__, "job_cluster_key", job_cluster_key) + _setter("job_cluster_key", job_cluster_key) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if max_retries is not None: - pulumi.set(__self__, "max_retries", max_retries) + _setter("max_retries", max_retries) if min_retry_interval_millis is not None: - pulumi.set(__self__, "min_retry_interval_millis", min_retry_interval_millis) + _setter("min_retry_interval_millis", min_retry_interval_millis) if new_cluster is not None: - pulumi.set(__self__, "new_cluster", new_cluster) + _setter("new_cluster", new_cluster) if notebook_task is not None: - pulumi.set(__self__, "notebook_task", notebook_task) + _setter("notebook_task", notebook_task) if notification_settings is not None: - pulumi.set(__self__, "notification_settings", notification_settings) + _setter("notification_settings", notification_settings) if pipeline_task is not None: - pulumi.set(__self__, "pipeline_task", pipeline_task) + _setter("pipeline_task", pipeline_task) if python_wheel_task is not None: - pulumi.set(__self__, "python_wheel_task", python_wheel_task) + _setter("python_wheel_task", python_wheel_task) if run_if is not None: - pulumi.set(__self__, "run_if", run_if) + _setter("run_if", run_if) if run_job_task is not None: - pulumi.set(__self__, "run_job_task", run_job_task) + _setter("run_job_task", run_job_task) if spark_jar_task is not None: - pulumi.set(__self__, "spark_jar_task", spark_jar_task) + _setter("spark_jar_task", spark_jar_task) if spark_python_task is not None: - pulumi.set(__self__, "spark_python_task", spark_python_task) + _setter("spark_python_task", spark_python_task) if spark_submit_task is not None: - pulumi.set(__self__, "spark_submit_task", spark_submit_task) + _setter("spark_submit_task", spark_submit_task) if sql_task is not None: - pulumi.set(__self__, "sql_task", sql_task) + _setter("sql_task", sql_task) if task_key is not None: - pulumi.set(__self__, "task_key", task_key) + _setter("task_key", task_key) if timeout_seconds is not None: - pulumi.set(__self__, "timeout_seconds", timeout_seconds) + _setter("timeout_seconds", timeout_seconds) @property @pulumi.getter(name="retryOnTimeout") @@ -17841,12 +23181,25 @@ def __init__(__self__, *, left: Optional[str] = None, op: Optional[str] = None, right: Optional[str] = None): + GetJobJobSettingsSettingsTaskConditionTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + left=left, + op=op, + right=right, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + left: Optional[str] = None, + op: Optional[str] = None, + right: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if left is not None: - pulumi.set(__self__, "left", left) + _setter("left", left) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if right is not None: - pulumi.set(__self__, "right", right) + _setter("right", right) @property @pulumi.getter @@ -17873,17 +23226,36 @@ def __init__(__self__, *, project_directory: Optional[str] = None, schema: Optional[str] = None, warehouse_id: Optional[str] = None): - pulumi.set(__self__, "commands", commands) + GetJobJobSettingsSettingsTaskDbtTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + commands=commands, + catalog=catalog, + profiles_directory=profiles_directory, + project_directory=project_directory, + schema=schema, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + commands: Sequence[str], + catalog: Optional[str] = None, + profiles_directory: Optional[str] = None, + project_directory: Optional[str] = None, + schema: Optional[str] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("commands", commands) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if profiles_directory is not None: - pulumi.set(__self__, "profiles_directory", profiles_directory) + _setter("profiles_directory", profiles_directory) if project_directory is not None: - pulumi.set(__self__, "project_directory", project_directory) + _setter("project_directory", project_directory) if schema is not None: - pulumi.set(__self__, "schema", schema) + _setter("schema", schema) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -17921,9 +23293,20 @@ class GetJobJobSettingsSettingsTaskDependsOnResult(dict): def __init__(__self__, *, task_key: str, outcome: Optional[str] = None): - pulumi.set(__self__, "task_key", task_key) + GetJobJobSettingsSettingsTaskDependsOnResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + task_key=task_key, + outcome=outcome, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + task_key: str, + outcome: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("task_key", task_key) if outcome is not None: - pulumi.set(__self__, "outcome", outcome) + _setter("outcome", outcome) @property @pulumi.getter(name="taskKey") @@ -17945,18 +23328,37 @@ def __init__(__self__, *, on_failures: Optional[Sequence[str]] = None, on_starts: Optional[Sequence[str]] = None, on_successes: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskEmailNotificationsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + on_duration_warning_threshold_exceededs: Optional[Sequence[str]] = None, + on_failures: Optional[Sequence[str]] = None, + on_starts: Optional[Sequence[str]] = None, + on_successes: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="alertOnLastAttempt") @@ -17993,7 +23395,16 @@ def on_successes(self) -> Optional[Sequence[str]]: class GetJobJobSettingsSettingsTaskHealthResult(dict): def __init__(__self__, *, rules: Sequence['outputs.GetJobJobSettingsSettingsTaskHealthRuleResult']): - pulumi.set(__self__, "rules", rules) + GetJobJobSettingsSettingsTaskHealthResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + rules=rules, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + rules: Sequence['outputs.GetJobJobSettingsSettingsTaskHealthRuleResult'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("rules", rules) @property @pulumi.getter @@ -18007,12 +23418,25 @@ def __init__(__self__, *, metric: Optional[str] = None, op: Optional[str] = None, value: Optional[int] = None): + GetJobJobSettingsSettingsTaskHealthRuleResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + metric=metric, + op=op, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + metric: Optional[str] = None, + op: Optional[str] = None, + value: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if metric is not None: - pulumi.set(__self__, "metric", metric) + _setter("metric", metric) if op is not None: - pulumi.set(__self__, "op", op) + _setter("op", op) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -18039,18 +23463,37 @@ def __init__(__self__, *, maven: Optional['outputs.GetJobJobSettingsSettingsTaskLibraryMavenResult'] = None, pypi: Optional['outputs.GetJobJobSettingsSettingsTaskLibraryPypiResult'] = None, whl: Optional[str] = None): + GetJobJobSettingsSettingsTaskLibraryResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + cran=cran, + egg=egg, + jar=jar, + maven=maven, + pypi=pypi, + whl=whl, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cran: Optional['outputs.GetJobJobSettingsSettingsTaskLibraryCranResult'] = None, + egg: Optional[str] = None, + jar: Optional[str] = None, + maven: Optional['outputs.GetJobJobSettingsSettingsTaskLibraryMavenResult'] = None, + pypi: Optional['outputs.GetJobJobSettingsSettingsTaskLibraryPypiResult'] = None, + whl: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cran is not None: - pulumi.set(__self__, "cran", cran) + _setter("cran", cran) if egg is not None: - pulumi.set(__self__, "egg", egg) + _setter("egg", egg) if jar is not None: - pulumi.set(__self__, "jar", jar) + _setter("jar", jar) if maven is not None: - pulumi.set(__self__, "maven", maven) + _setter("maven", maven) if pypi is not None: - pulumi.set(__self__, "pypi", pypi) + _setter("pypi", pypi) if whl is not None: - pulumi.set(__self__, "whl", whl) + _setter("whl", whl) @property @pulumi.getter @@ -18088,9 +23531,20 @@ class GetJobJobSettingsSettingsTaskLibraryCranResult(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsTaskLibraryCranResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -18109,11 +23563,24 @@ def __init__(__self__, *, coordinates: str, exclusions: Optional[Sequence[str]] = None, repo: Optional[str] = None): - pulumi.set(__self__, "coordinates", coordinates) + GetJobJobSettingsSettingsTaskLibraryMavenResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + coordinates=coordinates, + exclusions=exclusions, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + coordinates: str, + exclusions: Optional[Sequence[str]] = None, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("coordinates", coordinates) if exclusions is not None: - pulumi.set(__self__, "exclusions", exclusions) + _setter("exclusions", exclusions) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -18136,9 +23603,20 @@ class GetJobJobSettingsSettingsTaskLibraryPypiResult(dict): def __init__(__self__, *, package: str, repo: Optional[str] = None): - pulumi.set(__self__, "package", package) + GetJobJobSettingsSettingsTaskLibraryPypiResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + package=package, + repo=repo, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + package: str, + repo: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("package", package) if repo is not None: - pulumi.set(__self__, "repo", repo) + _setter("repo", repo) @property @pulumi.getter @@ -18184,59 +23662,126 @@ def __init__(__self__, *, spark_env_vars: Optional[Mapping[str, Any]] = None, ssh_public_keys: Optional[Sequence[str]] = None, workload_type: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeResult'] = None): - pulumi.set(__self__, "driver_instance_pool_id", driver_instance_pool_id) - pulumi.set(__self__, "driver_node_type_id", driver_node_type_id) - pulumi.set(__self__, "enable_elastic_disk", enable_elastic_disk) - pulumi.set(__self__, "enable_local_disk_encryption", enable_local_disk_encryption) - pulumi.set(__self__, "node_type_id", node_type_id) - pulumi.set(__self__, "num_workers", num_workers) - pulumi.set(__self__, "spark_version", spark_version) + GetJobJobSettingsSettingsTaskNewClusterResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + driver_instance_pool_id=driver_instance_pool_id, + driver_node_type_id=driver_node_type_id, + enable_elastic_disk=enable_elastic_disk, + enable_local_disk_encryption=enable_local_disk_encryption, + node_type_id=node_type_id, + num_workers=num_workers, + spark_version=spark_version, + apply_policy_default_values=apply_policy_default_values, + autoscale=autoscale, + autotermination_minutes=autotermination_minutes, + aws_attributes=aws_attributes, + azure_attributes=azure_attributes, + cluster_id=cluster_id, + cluster_log_conf=cluster_log_conf, + cluster_mount_infos=cluster_mount_infos, + cluster_name=cluster_name, + custom_tags=custom_tags, + data_security_mode=data_security_mode, + docker_image=docker_image, + gcp_attributes=gcp_attributes, + idempotency_token=idempotency_token, + init_scripts=init_scripts, + instance_pool_id=instance_pool_id, + policy_id=policy_id, + runtime_engine=runtime_engine, + single_user_name=single_user_name, + spark_conf=spark_conf, + spark_env_vars=spark_env_vars, + ssh_public_keys=ssh_public_keys, + workload_type=workload_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + driver_instance_pool_id: str, + driver_node_type_id: str, + enable_elastic_disk: bool, + enable_local_disk_encryption: bool, + node_type_id: str, + num_workers: int, + spark_version: str, + apply_policy_default_values: Optional[bool] = None, + autoscale: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterAutoscaleResult'] = None, + autotermination_minutes: Optional[int] = None, + aws_attributes: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterAwsAttributesResult'] = None, + azure_attributes: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterAzureAttributesResult'] = None, + cluster_id: Optional[str] = None, + cluster_log_conf: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterClusterLogConfResult'] = None, + cluster_mount_infos: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoResult']] = None, + cluster_name: Optional[str] = None, + custom_tags: Optional[Mapping[str, Any]] = None, + data_security_mode: Optional[str] = None, + docker_image: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterDockerImageResult'] = None, + gcp_attributes: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterGcpAttributesResult'] = None, + idempotency_token: Optional[str] = None, + init_scripts: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptResult']] = None, + instance_pool_id: Optional[str] = None, + policy_id: Optional[str] = None, + runtime_engine: Optional[str] = None, + single_user_name: Optional[str] = None, + spark_conf: Optional[Mapping[str, Any]] = None, + spark_env_vars: Optional[Mapping[str, Any]] = None, + ssh_public_keys: Optional[Sequence[str]] = None, + workload_type: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("driver_instance_pool_id", driver_instance_pool_id) + _setter("driver_node_type_id", driver_node_type_id) + _setter("enable_elastic_disk", enable_elastic_disk) + _setter("enable_local_disk_encryption", enable_local_disk_encryption) + _setter("node_type_id", node_type_id) + _setter("num_workers", num_workers) + _setter("spark_version", spark_version) if apply_policy_default_values is not None: - pulumi.set(__self__, "apply_policy_default_values", apply_policy_default_values) + _setter("apply_policy_default_values", apply_policy_default_values) if autoscale is not None: - pulumi.set(__self__, "autoscale", autoscale) + _setter("autoscale", autoscale) if autotermination_minutes is not None: - pulumi.set(__self__, "autotermination_minutes", autotermination_minutes) + _setter("autotermination_minutes", autotermination_minutes) if aws_attributes is not None: - pulumi.set(__self__, "aws_attributes", aws_attributes) + _setter("aws_attributes", aws_attributes) if azure_attributes is not None: - pulumi.set(__self__, "azure_attributes", azure_attributes) + _setter("azure_attributes", azure_attributes) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_log_conf is not None: - pulumi.set(__self__, "cluster_log_conf", cluster_log_conf) + _setter("cluster_log_conf", cluster_log_conf) if cluster_mount_infos is not None: - pulumi.set(__self__, "cluster_mount_infos", cluster_mount_infos) + _setter("cluster_mount_infos", cluster_mount_infos) if cluster_name is not None: - pulumi.set(__self__, "cluster_name", cluster_name) + _setter("cluster_name", cluster_name) if custom_tags is not None: - pulumi.set(__self__, "custom_tags", custom_tags) + _setter("custom_tags", custom_tags) if data_security_mode is not None: - pulumi.set(__self__, "data_security_mode", data_security_mode) + _setter("data_security_mode", data_security_mode) if docker_image is not None: - pulumi.set(__self__, "docker_image", docker_image) + _setter("docker_image", docker_image) if gcp_attributes is not None: - pulumi.set(__self__, "gcp_attributes", gcp_attributes) + _setter("gcp_attributes", gcp_attributes) if idempotency_token is not None: - pulumi.set(__self__, "idempotency_token", idempotency_token) + _setter("idempotency_token", idempotency_token) if init_scripts is not None: - pulumi.set(__self__, "init_scripts", init_scripts) + _setter("init_scripts", init_scripts) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if policy_id is not None: - pulumi.set(__self__, "policy_id", policy_id) + _setter("policy_id", policy_id) if runtime_engine is not None: - pulumi.set(__self__, "runtime_engine", runtime_engine) + _setter("runtime_engine", runtime_engine) if single_user_name is not None: - pulumi.set(__self__, "single_user_name", single_user_name) + _setter("single_user_name", single_user_name) if spark_conf is not None: - pulumi.set(__self__, "spark_conf", spark_conf) + _setter("spark_conf", spark_conf) if spark_env_vars is not None: - pulumi.set(__self__, "spark_env_vars", spark_env_vars) + _setter("spark_env_vars", spark_env_vars) if ssh_public_keys is not None: - pulumi.set(__self__, "ssh_public_keys", ssh_public_keys) + _setter("ssh_public_keys", ssh_public_keys) if workload_type is not None: - pulumi.set(__self__, "workload_type", workload_type) + _setter("workload_type", workload_type) @property @pulumi.getter(name="driverInstancePoolId") @@ -18394,10 +23939,21 @@ class GetJobJobSettingsSettingsTaskNewClusterAutoscaleResult(dict): def __init__(__self__, *, max_workers: Optional[int] = None, min_workers: Optional[int] = None): + GetJobJobSettingsSettingsTaskNewClusterAutoscaleResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + max_workers=max_workers, + min_workers=min_workers, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + max_workers: Optional[int] = None, + min_workers: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): if max_workers is not None: - pulumi.set(__self__, "max_workers", max_workers) + _setter("max_workers", max_workers) if min_workers is not None: - pulumi.set(__self__, "min_workers", min_workers) + _setter("min_workers", min_workers) @property @pulumi.getter(name="maxWorkers") @@ -18421,22 +23977,45 @@ def __init__(__self__, *, instance_profile_arn: Optional[str] = None, spot_bid_price_percent: Optional[int] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterAwsAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + ebs_volume_count=ebs_volume_count, + ebs_volume_size=ebs_volume_size, + ebs_volume_type=ebs_volume_type, + first_on_demand=first_on_demand, + instance_profile_arn=instance_profile_arn, + spot_bid_price_percent=spot_bid_price_percent, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + ebs_volume_count: Optional[int] = None, + ebs_volume_size: Optional[int] = None, + ebs_volume_type: Optional[str] = None, + first_on_demand: Optional[int] = None, + instance_profile_arn: Optional[str] = None, + spot_bid_price_percent: Optional[int] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if ebs_volume_count is not None: - pulumi.set(__self__, "ebs_volume_count", ebs_volume_count) + _setter("ebs_volume_count", ebs_volume_count) if ebs_volume_size is not None: - pulumi.set(__self__, "ebs_volume_size", ebs_volume_size) + _setter("ebs_volume_size", ebs_volume_size) if ebs_volume_type is not None: - pulumi.set(__self__, "ebs_volume_type", ebs_volume_type) + _setter("ebs_volume_type", ebs_volume_type) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if spot_bid_price_percent is not None: - pulumi.set(__self__, "spot_bid_price_percent", spot_bid_price_percent) + _setter("spot_bid_price_percent", spot_bid_price_percent) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -18485,12 +24064,25 @@ def __init__(__self__, *, availability: Optional[str] = None, first_on_demand: Optional[int] = None, spot_bid_max_price: Optional[float] = None): + GetJobJobSettingsSettingsTaskNewClusterAzureAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + first_on_demand=first_on_demand, + spot_bid_max_price=spot_bid_max_price, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + first_on_demand: Optional[int] = None, + spot_bid_max_price: Optional[float] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if first_on_demand is not None: - pulumi.set(__self__, "first_on_demand", first_on_demand) + _setter("first_on_demand", first_on_demand) if spot_bid_max_price is not None: - pulumi.set(__self__, "spot_bid_max_price", spot_bid_max_price) + _setter("spot_bid_max_price", spot_bid_max_price) @property @pulumi.getter @@ -18513,10 +24105,21 @@ class GetJobJobSettingsSettingsTaskNewClusterClusterLogConfResult(dict): def __init__(__self__, *, dbfs: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsResult'] = None, s3: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterClusterLogConfS3Result'] = None): + GetJobJobSettingsSettingsTaskNewClusterClusterLogConfResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + dbfs=dbfs, + s3=s3, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dbfs: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsResult'] = None, + s3: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterClusterLogConfS3Result'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) @property @pulumi.getter @@ -18533,7 +24136,16 @@ def s3(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterCluster class GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterClusterLogConfDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -18551,19 +24163,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterClusterLogConfS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -18607,10 +24240,23 @@ def __init__(__self__, *, local_mount_dir_path: str, network_filesystem_info: 'outputs.GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemInfoResult', remote_mount_dir_path: Optional[str] = None): - pulumi.set(__self__, "local_mount_dir_path", local_mount_dir_path) - pulumi.set(__self__, "network_filesystem_info", network_filesystem_info) + GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + local_mount_dir_path=local_mount_dir_path, + network_filesystem_info=network_filesystem_info, + remote_mount_dir_path=remote_mount_dir_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + local_mount_dir_path: str, + network_filesystem_info: 'outputs.GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemInfoResult', + remote_mount_dir_path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("local_mount_dir_path", local_mount_dir_path) + _setter("network_filesystem_info", network_filesystem_info) if remote_mount_dir_path is not None: - pulumi.set(__self__, "remote_mount_dir_path", remote_mount_dir_path) + _setter("remote_mount_dir_path", remote_mount_dir_path) @property @pulumi.getter(name="localMountDirPath") @@ -18633,9 +24279,20 @@ class GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemIn def __init__(__self__, *, server_address: str, mount_options: Optional[str] = None): - pulumi.set(__self__, "server_address", server_address) + GetJobJobSettingsSettingsTaskNewClusterClusterMountInfoNetworkFilesystemInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + server_address=server_address, + mount_options=mount_options, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + server_address: str, + mount_options: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("server_address", server_address) if mount_options is not None: - pulumi.set(__self__, "mount_options", mount_options) + _setter("mount_options", mount_options) @property @pulumi.getter(name="serverAddress") @@ -18653,9 +24310,20 @@ class GetJobJobSettingsSettingsTaskNewClusterDockerImageResult(dict): def __init__(__self__, *, url: str, basic_auth: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthResult'] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsTaskNewClusterDockerImageResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + basic_auth=basic_auth, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + basic_auth: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if basic_auth is not None: - pulumi.set(__self__, "basic_auth", basic_auth) + _setter("basic_auth", basic_auth) @property @pulumi.getter @@ -18673,8 +24341,19 @@ class GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthResult(dict): def __init__(__self__, *, password: str, username: str): - pulumi.set(__self__, "password", password) - pulumi.set(__self__, "username", username) + GetJobJobSettingsSettingsTaskNewClusterDockerImageBasicAuthResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + password=password, + username=username, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + password: str, + username: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("password", password) + _setter("username", username) @property @pulumi.getter @@ -18696,18 +24375,37 @@ def __init__(__self__, *, local_ssd_count: Optional[int] = None, use_preemptible_executors: Optional[bool] = None, zone_id: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterGcpAttributesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + availability=availability, + boot_disk_size=boot_disk_size, + google_service_account=google_service_account, + local_ssd_count=local_ssd_count, + use_preemptible_executors=use_preemptible_executors, + zone_id=zone_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + availability: Optional[str] = None, + boot_disk_size: Optional[int] = None, + google_service_account: Optional[str] = None, + local_ssd_count: Optional[int] = None, + use_preemptible_executors: Optional[bool] = None, + zone_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if availability is not None: - pulumi.set(__self__, "availability", availability) + _setter("availability", availability) if boot_disk_size is not None: - pulumi.set(__self__, "boot_disk_size", boot_disk_size) + _setter("boot_disk_size", boot_disk_size) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if local_ssd_count is not None: - pulumi.set(__self__, "local_ssd_count", local_ssd_count) + _setter("local_ssd_count", local_ssd_count) if use_preemptible_executors is not None: - pulumi.set(__self__, "use_preemptible_executors", use_preemptible_executors) + _setter("use_preemptible_executors", use_preemptible_executors) if zone_id is not None: - pulumi.set(__self__, "zone_id", zone_id) + _setter("zone_id", zone_id) @property @pulumi.getter @@ -18750,20 +24448,41 @@ def __init__(__self__, *, s3: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptS3Result'] = None, volumes: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesResult'] = None, workspace: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceResult'] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + abfss=abfss, + dbfs=dbfs, + file=file, + gcs=gcs, + s3=s3, + volumes=volumes, + workspace=workspace, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + abfss: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptAbfssResult'] = None, + dbfs: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptDbfsResult'] = None, + file: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptFileResult'] = None, + gcs: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptGcsResult'] = None, + s3: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptS3Result'] = None, + volumes: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesResult'] = None, + workspace: Optional['outputs.GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceResult'] = None, + opts: Optional[pulumi.ResourceOptions]=None): if abfss is not None: - pulumi.set(__self__, "abfss", abfss) + _setter("abfss", abfss) if dbfs is not None: - pulumi.set(__self__, "dbfs", dbfs) + _setter("dbfs", dbfs) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if gcs is not None: - pulumi.set(__self__, "gcs", gcs) + _setter("gcs", gcs) if s3 is not None: - pulumi.set(__self__, "s3", s3) + _setter("s3", s3) if volumes is not None: - pulumi.set(__self__, "volumes", volumes) + _setter("volumes", volumes) if workspace is not None: - pulumi.set(__self__, "workspace", workspace) + _setter("workspace", workspace) @property @pulumi.getter @@ -18805,8 +24524,17 @@ def workspace(self) -> Optional['outputs.GetJobJobSettingsSettingsTaskNewCluster class GetJobJobSettingsSettingsTaskNewClusterInitScriptAbfssResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptAbfssResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -18818,7 +24546,16 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsTaskNewClusterInitScriptDbfsResult(dict): def __init__(__self__, *, destination: str): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterInitScriptDbfsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) @property @pulumi.getter @@ -18830,8 +24567,17 @@ def destination(self) -> str: class GetJobJobSettingsSettingsTaskNewClusterInitScriptFileResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptFileResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -18843,8 +24589,17 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsTaskNewClusterInitScriptGcsResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptGcsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -18862,19 +24617,40 @@ def __init__(__self__, *, endpoint: Optional[str] = None, kms_key: Optional[str] = None, region: Optional[str] = None): - pulumi.set(__self__, "destination", destination) + GetJobJobSettingsSettingsTaskNewClusterInitScriptS3Result._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + canned_acl=canned_acl, + enable_encryption=enable_encryption, + encryption_type=encryption_type, + endpoint=endpoint, + kms_key=kms_key, + region=region, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: str, + canned_acl: Optional[str] = None, + enable_encryption: Optional[bool] = None, + encryption_type: Optional[str] = None, + endpoint: Optional[str] = None, + kms_key: Optional[str] = None, + region: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("destination", destination) if canned_acl is not None: - pulumi.set(__self__, "canned_acl", canned_acl) + _setter("canned_acl", canned_acl) if enable_encryption is not None: - pulumi.set(__self__, "enable_encryption", enable_encryption) + _setter("enable_encryption", enable_encryption) if encryption_type is not None: - pulumi.set(__self__, "encryption_type", encryption_type) + _setter("encryption_type", encryption_type) if endpoint is not None: - pulumi.set(__self__, "endpoint", endpoint) + _setter("endpoint", endpoint) if kms_key is not None: - pulumi.set(__self__, "kms_key", kms_key) + _setter("kms_key", kms_key) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) @property @pulumi.getter @@ -18916,8 +24692,17 @@ def region(self) -> Optional[str]: class GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptVolumesResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -18929,8 +24714,17 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceResult(dict): def __init__(__self__, *, destination: Optional[str] = None): + GetJobJobSettingsSettingsTaskNewClusterInitScriptWorkspaceResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination=destination, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination is not None: - pulumi.set(__self__, "destination", destination) + _setter("destination", destination) @property @pulumi.getter @@ -18942,7 +24736,16 @@ def destination(self) -> Optional[str]: class GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeResult(dict): def __init__(__self__, *, clients: 'outputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsResult'): - pulumi.set(__self__, "clients", clients) + GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + clients=clients, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + clients: 'outputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsResult', + opts: Optional[pulumi.ResourceOptions]=None): + _setter("clients", clients) @property @pulumi.getter @@ -18955,10 +24758,21 @@ class GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsResult(dict): def __init__(__self__, *, jobs: Optional[bool] = None, notebooks: Optional[bool] = None): + GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeClientsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + jobs=jobs, + notebooks=notebooks, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jobs: Optional[bool] = None, + notebooks: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jobs is not None: - pulumi.set(__self__, "jobs", jobs) + _setter("jobs", jobs) if notebooks is not None: - pulumi.set(__self__, "notebooks", notebooks) + _setter("notebooks", notebooks) @property @pulumi.getter @@ -18977,11 +24791,24 @@ def __init__(__self__, *, notebook_path: str, base_parameters: Optional[Mapping[str, Any]] = None, source: Optional[str] = None): - pulumi.set(__self__, "notebook_path", notebook_path) + GetJobJobSettingsSettingsTaskNotebookTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + notebook_path=notebook_path, + base_parameters=base_parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + notebook_path: str, + base_parameters: Optional[Mapping[str, Any]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("notebook_path", notebook_path) if base_parameters is not None: - pulumi.set(__self__, "base_parameters", base_parameters) + _setter("base_parameters", base_parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="notebookPath") @@ -19005,12 +24832,25 @@ def __init__(__self__, *, alert_on_last_attempt: Optional[bool] = None, no_alert_for_canceled_runs: Optional[bool] = None, no_alert_for_skipped_runs: Optional[bool] = None): + GetJobJobSettingsSettingsTaskNotificationSettingsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_on_last_attempt=alert_on_last_attempt, + no_alert_for_canceled_runs=no_alert_for_canceled_runs, + no_alert_for_skipped_runs=no_alert_for_skipped_runs, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_on_last_attempt: Optional[bool] = None, + no_alert_for_canceled_runs: Optional[bool] = None, + no_alert_for_skipped_runs: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert_on_last_attempt is not None: - pulumi.set(__self__, "alert_on_last_attempt", alert_on_last_attempt) + _setter("alert_on_last_attempt", alert_on_last_attempt) if no_alert_for_canceled_runs is not None: - pulumi.set(__self__, "no_alert_for_canceled_runs", no_alert_for_canceled_runs) + _setter("no_alert_for_canceled_runs", no_alert_for_canceled_runs) if no_alert_for_skipped_runs is not None: - pulumi.set(__self__, "no_alert_for_skipped_runs", no_alert_for_skipped_runs) + _setter("no_alert_for_skipped_runs", no_alert_for_skipped_runs) @property @pulumi.getter(name="alertOnLastAttempt") @@ -19033,9 +24873,20 @@ class GetJobJobSettingsSettingsTaskPipelineTaskResult(dict): def __init__(__self__, *, pipeline_id: str, full_refresh: Optional[bool] = None): - pulumi.set(__self__, "pipeline_id", pipeline_id) + GetJobJobSettingsSettingsTaskPipelineTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + pipeline_id=pipeline_id, + full_refresh=full_refresh, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + pipeline_id: str, + full_refresh: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("pipeline_id", pipeline_id) if full_refresh is not None: - pulumi.set(__self__, "full_refresh", full_refresh) + _setter("full_refresh", full_refresh) @property @pulumi.getter(name="pipelineId") @@ -19055,14 +24906,29 @@ def __init__(__self__, *, named_parameters: Optional[Mapping[str, Any]] = None, package_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskPythonWheelTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + entry_point=entry_point, + named_parameters=named_parameters, + package_name=package_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + entry_point: Optional[str] = None, + named_parameters: Optional[Mapping[str, Any]] = None, + package_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if entry_point is not None: - pulumi.set(__self__, "entry_point", entry_point) + _setter("entry_point", entry_point) if named_parameters is not None: - pulumi.set(__self__, "named_parameters", named_parameters) + _setter("named_parameters", named_parameters) if package_name is not None: - pulumi.set(__self__, "package_name", package_name) + _setter("package_name", package_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="entryPoint") @@ -19090,9 +24956,20 @@ class GetJobJobSettingsSettingsTaskRunJobTaskResult(dict): def __init__(__self__, *, job_id: int, job_parameters: Optional[Mapping[str, Any]] = None): - pulumi.set(__self__, "job_id", job_id) + GetJobJobSettingsSettingsTaskRunJobTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + job_id=job_id, + job_parameters=job_parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + job_id: int, + job_parameters: Optional[Mapping[str, Any]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("job_id", job_id) if job_parameters is not None: - pulumi.set(__self__, "job_parameters", job_parameters) + _setter("job_parameters", job_parameters) @property @pulumi.getter(name="jobId") @@ -19111,12 +24988,25 @@ def __init__(__self__, *, jar_uri: Optional[str] = None, main_class_name: Optional[str] = None, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskSparkJarTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + jar_uri=jar_uri, + main_class_name=main_class_name, + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + jar_uri: Optional[str] = None, + main_class_name: Optional[str] = None, + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if jar_uri is not None: - pulumi.set(__self__, "jar_uri", jar_uri) + _setter("jar_uri", jar_uri) if main_class_name is not None: - pulumi.set(__self__, "main_class_name", main_class_name) + _setter("main_class_name", main_class_name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter(name="jarUri") @@ -19140,11 +25030,24 @@ def __init__(__self__, *, python_file: str, parameters: Optional[Sequence[str]] = None, source: Optional[str] = None): - pulumi.set(__self__, "python_file", python_file) + GetJobJobSettingsSettingsTaskSparkPythonTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + python_file=python_file, + parameters=parameters, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + python_file: str, + parameters: Optional[Sequence[str]] = None, + source: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("python_file", python_file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter(name="pythonFile") @@ -19166,8 +25069,17 @@ def source(self) -> Optional[str]: class GetJobJobSettingsSettingsTaskSparkSubmitTaskResult(dict): def __init__(__self__, *, parameters: Optional[Sequence[str]] = None): + GetJobJobSettingsSettingsTaskSparkSubmitTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + parameters=parameters, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + parameters: Optional[Sequence[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) @property @pulumi.getter @@ -19184,18 +25096,37 @@ def __init__(__self__, *, parameters: Optional[Mapping[str, Any]] = None, query: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskQueryResult'] = None, warehouse_id: Optional[str] = None): + GetJobJobSettingsSettingsTaskSqlTaskResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert=alert, + dashboard=dashboard, + file=file, + parameters=parameters, + query=query, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskAlertResult'] = None, + dashboard: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskDashboardResult'] = None, + file: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskFileResult'] = None, + parameters: Optional[Mapping[str, Any]] = None, + query: Optional['outputs.GetJobJobSettingsSettingsTaskSqlTaskQueryResult'] = None, + warehouse_id: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if alert is not None: - pulumi.set(__self__, "alert", alert) + _setter("alert", alert) if dashboard is not None: - pulumi.set(__self__, "dashboard", dashboard) + _setter("dashboard", dashboard) if file is not None: - pulumi.set(__self__, "file", file) + _setter("file", file) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if query is not None: - pulumi.set(__self__, "query", query) + _setter("query", query) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter @@ -19234,10 +25165,23 @@ def __init__(__self__, *, alert_id: str, subscriptions: Sequence['outputs.GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionResult'], pause_subscriptions: Optional[bool] = None): - pulumi.set(__self__, "alert_id", alert_id) - pulumi.set(__self__, "subscriptions", subscriptions) + GetJobJobSettingsSettingsTaskSqlTaskAlertResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + alert_id=alert_id, + subscriptions=subscriptions, + pause_subscriptions=pause_subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + alert_id: str, + subscriptions: Sequence['outputs.GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionResult'], + pause_subscriptions: Optional[bool] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("alert_id", alert_id) + _setter("subscriptions", subscriptions) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) @property @pulumi.getter(name="alertId") @@ -19260,10 +25204,21 @@ class GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionResult(dict): def __init__(__self__, *, destination_id: Optional[str] = None, user_name: Optional[str] = None): + GetJobJobSettingsSettingsTaskSqlTaskAlertSubscriptionResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -19283,13 +25238,28 @@ def __init__(__self__, *, custom_subject: Optional[str] = None, pause_subscriptions: Optional[bool] = None, subscriptions: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionResult']] = None): - pulumi.set(__self__, "dashboard_id", dashboard_id) + GetJobJobSettingsSettingsTaskSqlTaskDashboardResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + dashboard_id=dashboard_id, + custom_subject=custom_subject, + pause_subscriptions=pause_subscriptions, + subscriptions=subscriptions, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dashboard_id: str, + custom_subject: Optional[str] = None, + pause_subscriptions: Optional[bool] = None, + subscriptions: Optional[Sequence['outputs.GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionResult']] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dashboard_id", dashboard_id) if custom_subject is not None: - pulumi.set(__self__, "custom_subject", custom_subject) + _setter("custom_subject", custom_subject) if pause_subscriptions is not None: - pulumi.set(__self__, "pause_subscriptions", pause_subscriptions) + _setter("pause_subscriptions", pause_subscriptions) if subscriptions is not None: - pulumi.set(__self__, "subscriptions", subscriptions) + _setter("subscriptions", subscriptions) @property @pulumi.getter(name="dashboardId") @@ -19317,10 +25287,21 @@ class GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionResult(dict): def __init__(__self__, *, destination_id: Optional[str] = None, user_name: Optional[str] = None): + GetJobJobSettingsSettingsTaskSqlTaskDashboardSubscriptionResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + destination_id=destination_id, + user_name=user_name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + destination_id: Optional[str] = None, + user_name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if destination_id is not None: - pulumi.set(__self__, "destination_id", destination_id) + _setter("destination_id", destination_id) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) @property @pulumi.getter(name="destinationId") @@ -19337,7 +25318,16 @@ def user_name(self) -> Optional[str]: class GetJobJobSettingsSettingsTaskSqlTaskFileResult(dict): def __init__(__self__, *, path: str): - pulumi.set(__self__, "path", path) + GetJobJobSettingsSettingsTaskSqlTaskFileResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) @property @pulumi.getter @@ -19349,7 +25339,16 @@ def path(self) -> str: class GetJobJobSettingsSettingsTaskSqlTaskQueryResult(dict): def __init__(__self__, *, query_id: str): - pulumi.set(__self__, "query_id", query_id) + GetJobJobSettingsSettingsTaskSqlTaskQueryResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + query_id=query_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + query_id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("query_id", query_id) @property @pulumi.getter(name="queryId") @@ -19362,9 +25361,20 @@ class GetJobJobSettingsSettingsTriggerResult(dict): def __init__(__self__, *, file_arrival: 'outputs.GetJobJobSettingsSettingsTriggerFileArrivalResult', pause_status: Optional[str] = None): - pulumi.set(__self__, "file_arrival", file_arrival) + GetJobJobSettingsSettingsTriggerResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + file_arrival=file_arrival, + pause_status=pause_status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + file_arrival: 'outputs.GetJobJobSettingsSettingsTriggerFileArrivalResult', + pause_status: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("file_arrival", file_arrival) if pause_status is not None: - pulumi.set(__self__, "pause_status", pause_status) + _setter("pause_status", pause_status) @property @pulumi.getter(name="fileArrival") @@ -19383,11 +25393,24 @@ def __init__(__self__, *, url: str, min_time_between_triggers_seconds: Optional[int] = None, wait_after_last_change_seconds: Optional[int] = None): - pulumi.set(__self__, "url", url) + GetJobJobSettingsSettingsTriggerFileArrivalResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + min_time_between_triggers_seconds=min_time_between_triggers_seconds, + wait_after_last_change_seconds=wait_after_last_change_seconds, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: str, + min_time_between_triggers_seconds: Optional[int] = None, + wait_after_last_change_seconds: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if min_time_between_triggers_seconds is not None: - pulumi.set(__self__, "min_time_between_triggers_seconds", min_time_between_triggers_seconds) + _setter("min_time_between_triggers_seconds", min_time_between_triggers_seconds) if wait_after_last_change_seconds is not None: - pulumi.set(__self__, "wait_after_last_change_seconds", wait_after_last_change_seconds) + _setter("wait_after_last_change_seconds", wait_after_last_change_seconds) @property @pulumi.getter @@ -19412,14 +25435,29 @@ def __init__(__self__, *, on_failures: Optional[Sequence['outputs.GetJobJobSettingsSettingsWebhookNotificationsOnFailureResult']] = None, on_starts: Optional[Sequence['outputs.GetJobJobSettingsSettingsWebhookNotificationsOnStartResult']] = None, on_successes: Optional[Sequence['outputs.GetJobJobSettingsSettingsWebhookNotificationsOnSuccessResult']] = None): + GetJobJobSettingsSettingsWebhookNotificationsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + on_duration_warning_threshold_exceededs=on_duration_warning_threshold_exceededs, + on_failures=on_failures, + on_starts=on_starts, + on_successes=on_successes, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + on_duration_warning_threshold_exceededs: Optional[Sequence['outputs.GetJobJobSettingsSettingsWebhookNotificationsOnDurationWarningThresholdExceededResult']] = None, + on_failures: Optional[Sequence['outputs.GetJobJobSettingsSettingsWebhookNotificationsOnFailureResult']] = None, + on_starts: Optional[Sequence['outputs.GetJobJobSettingsSettingsWebhookNotificationsOnStartResult']] = None, + on_successes: Optional[Sequence['outputs.GetJobJobSettingsSettingsWebhookNotificationsOnSuccessResult']] = None, + opts: Optional[pulumi.ResourceOptions]=None): if on_duration_warning_threshold_exceededs is not None: - pulumi.set(__self__, "on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) + _setter("on_duration_warning_threshold_exceededs", on_duration_warning_threshold_exceededs) if on_failures is not None: - pulumi.set(__self__, "on_failures", on_failures) + _setter("on_failures", on_failures) if on_starts is not None: - pulumi.set(__self__, "on_starts", on_starts) + _setter("on_starts", on_starts) if on_successes is not None: - pulumi.set(__self__, "on_successes", on_successes) + _setter("on_successes", on_successes) @property @pulumi.getter(name="onDurationWarningThresholdExceededs") @@ -19449,7 +25487,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnDurationWarningThresholdExceededResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -19467,7 +25514,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnFailureResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -19485,7 +25541,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnStartResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -19503,7 +25568,16 @@ def __init__(__self__, *, """ :param str id: the id of Job if the resource was matched by name. """ - pulumi.set(__self__, "id", id) + GetJobJobSettingsSettingsWebhookNotificationsOnSuccessResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + id=id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + id: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("id", id) @property @pulumi.getter @@ -19544,42 +25618,85 @@ def __init__(__self__, *, :param str owner: Username/groupname/sp application_id of the metastore owner. :param str storage_root: Path on cloud storage account, where managed `Table` are stored. Change forces creation of a new resource. """ + GetMetastoreMetastoreInfoResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + cloud=cloud, + created_at=created_at, + created_by=created_by, + default_data_access_config_id=default_data_access_config_id, + delta_sharing_organization_name=delta_sharing_organization_name, + delta_sharing_recipient_token_lifetime_in_seconds=delta_sharing_recipient_token_lifetime_in_seconds, + delta_sharing_scope=delta_sharing_scope, + global_metastore_id=global_metastore_id, + metastore_id=metastore_id, + name=name, + owner=owner, + privilege_model_version=privilege_model_version, + region=region, + storage_root=storage_root, + storage_root_credential_id=storage_root_credential_id, + storage_root_credential_name=storage_root_credential_name, + updated_at=updated_at, + updated_by=updated_by, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cloud: Optional[str] = None, + created_at: Optional[int] = None, + created_by: Optional[str] = None, + default_data_access_config_id: Optional[str] = None, + delta_sharing_organization_name: Optional[str] = None, + delta_sharing_recipient_token_lifetime_in_seconds: Optional[int] = None, + delta_sharing_scope: Optional[str] = None, + global_metastore_id: Optional[str] = None, + metastore_id: Optional[str] = None, + name: Optional[str] = None, + owner: Optional[str] = None, + privilege_model_version: Optional[str] = None, + region: Optional[str] = None, + storage_root: Optional[str] = None, + storage_root_credential_id: Optional[str] = None, + storage_root_credential_name: Optional[str] = None, + updated_at: Optional[int] = None, + updated_by: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if cloud is not None: - pulumi.set(__self__, "cloud", cloud) + _setter("cloud", cloud) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if default_data_access_config_id is not None: - pulumi.set(__self__, "default_data_access_config_id", default_data_access_config_id) + _setter("default_data_access_config_id", default_data_access_config_id) if delta_sharing_organization_name is not None: - pulumi.set(__self__, "delta_sharing_organization_name", delta_sharing_organization_name) + _setter("delta_sharing_organization_name", delta_sharing_organization_name) if delta_sharing_recipient_token_lifetime_in_seconds is not None: - pulumi.set(__self__, "delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) + _setter("delta_sharing_recipient_token_lifetime_in_seconds", delta_sharing_recipient_token_lifetime_in_seconds) if delta_sharing_scope is not None: - pulumi.set(__self__, "delta_sharing_scope", delta_sharing_scope) + _setter("delta_sharing_scope", delta_sharing_scope) if global_metastore_id is not None: - pulumi.set(__self__, "global_metastore_id", global_metastore_id) + _setter("global_metastore_id", global_metastore_id) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if privilege_model_version is not None: - pulumi.set(__self__, "privilege_model_version", privilege_model_version) + _setter("privilege_model_version", privilege_model_version) if region is not None: - pulumi.set(__self__, "region", region) + _setter("region", region) if storage_root is not None: - pulumi.set(__self__, "storage_root", storage_root) + _setter("storage_root", storage_root) if storage_root_credential_id is not None: - pulumi.set(__self__, "storage_root_credential_id", storage_root_credential_id) + _setter("storage_root_credential_id", storage_root_credential_id) if storage_root_credential_name is not None: - pulumi.set(__self__, "storage_root_credential_name", storage_root_credential_name) + _setter("storage_root_credential_name", storage_root_credential_name) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) if updated_by is not None: - pulumi.set(__self__, "updated_by", updated_by) + _setter("updated_by", updated_by) @property @pulumi.getter @@ -19701,10 +25818,21 @@ def __init__(__self__, *, """ :param str path: Path to workspace directory """ + GetNotebookPathsNotebookPathListResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + language=language, + path=path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + language: Optional[str] = None, + path: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if language is not None: - pulumi.set(__self__, "language", language) + _setter("language", language) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) @property @pulumi.getter @@ -19739,23 +25867,52 @@ def __init__(__self__, *, :param str name: The name of the share :param str comment: Description about the object. """ - pulumi.set(__self__, "added_at", added_at) - pulumi.set(__self__, "added_by", added_by) - pulumi.set(__self__, "data_object_type", data_object_type) - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "status", status) + GetShareObjectResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + added_at=added_at, + added_by=added_by, + data_object_type=data_object_type, + name=name, + status=status, + cdf_enabled=cdf_enabled, + comment=comment, + history_data_sharing_status=history_data_sharing_status, + partitions=partitions, + shared_as=shared_as, + start_version=start_version, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + added_at: int, + added_by: str, + data_object_type: str, + name: str, + status: str, + cdf_enabled: Optional[bool] = None, + comment: Optional[str] = None, + history_data_sharing_status: Optional[str] = None, + partitions: Optional[Sequence['outputs.GetShareObjectPartitionResult']] = None, + shared_as: Optional[str] = None, + start_version: Optional[int] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("added_at", added_at) + _setter("added_by", added_by) + _setter("data_object_type", data_object_type) + _setter("name", name) + _setter("status", status) if cdf_enabled is not None: - pulumi.set(__self__, "cdf_enabled", cdf_enabled) + _setter("cdf_enabled", cdf_enabled) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if history_data_sharing_status is not None: - pulumi.set(__self__, "history_data_sharing_status", history_data_sharing_status) + _setter("history_data_sharing_status", history_data_sharing_status) if partitions is not None: - pulumi.set(__self__, "partitions", partitions) + _setter("partitions", partitions) if shared_as is not None: - pulumi.set(__self__, "shared_as", shared_as) + _setter("shared_as", shared_as) if start_version is not None: - pulumi.set(__self__, "start_version", start_version) + _setter("start_version", start_version) @property @pulumi.getter(name="addedAt") @@ -19826,7 +25983,16 @@ def start_version(self) -> Optional[int]: class GetShareObjectPartitionResult(dict): def __init__(__self__, *, values: Sequence['outputs.GetShareObjectPartitionValueResult']): - pulumi.set(__self__, "values", values) + GetShareObjectPartitionResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + values=values, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + values: Sequence['outputs.GetShareObjectPartitionValueResult'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("values", values) @property @pulumi.getter @@ -19844,12 +26010,27 @@ def __init__(__self__, *, """ :param str name: The name of the share """ - pulumi.set(__self__, "name", name) - pulumi.set(__self__, "op", op) + GetShareObjectPartitionValueResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + op=op, + recipient_property_key=recipient_property_key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: str, + op: str, + recipient_property_key: Optional[str] = None, + value: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("name", name) + _setter("op", op) if recipient_property_key is not None: - pulumi.set(__self__, "recipient_property_key", recipient_property_key) + _setter("recipient_property_key", recipient_property_key) if value is not None: - pulumi.set(__self__, "value", value) + _setter("value", value) @property @pulumi.getter @@ -19882,8 +26063,17 @@ def __init__(__self__, *, """ :param str name: Name of the SQL warehouse to search (case-sensitive). """ + GetSqlWarehouseChannelResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + name: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter @@ -19902,13 +26092,30 @@ def __init__(__self__, *, protocol: str, host: Optional[str] = None, hostname: Optional[str] = None): - pulumi.set(__self__, "path", path) - pulumi.set(__self__, "port", port) - pulumi.set(__self__, "protocol", protocol) + GetSqlWarehouseOdbcParamsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + port=port, + protocol=protocol, + host=host, + hostname=hostname, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: str, + port: int, + protocol: str, + host: Optional[str] = None, + hostname: Optional[str] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) + _setter("port", port) + _setter("protocol", protocol) if host is not None: - pulumi.set(__self__, "host", host) + _setter("host", host) if hostname is not None: - pulumi.set(__self__, "hostname", hostname) + _setter("hostname", hostname) @property @pulumi.getter @@ -19940,7 +26147,16 @@ def hostname(self) -> Optional[str]: class GetSqlWarehouseTagsResult(dict): def __init__(__self__, *, custom_tags: Sequence['outputs.GetSqlWarehouseTagsCustomTagResult']): - pulumi.set(__self__, "custom_tags", custom_tags) + GetSqlWarehouseTagsResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + custom_tags=custom_tags, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + custom_tags: Sequence['outputs.GetSqlWarehouseTagsCustomTagResult'], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("custom_tags", custom_tags) @property @pulumi.getter(name="customTags") @@ -19953,8 +26169,19 @@ class GetSqlWarehouseTagsCustomTagResult(dict): def __init__(__self__, *, key: str, value: str): - pulumi.set(__self__, "key", key) - pulumi.set(__self__, "value", value) + GetSqlWarehouseTagsCustomTagResult._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + value=value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: str, + value: str, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key", key) + _setter("value", value) @property @pulumi.getter diff --git a/sdk/python/pulumi_databricks/permission_assignment.py b/sdk/python/pulumi_databricks/permission_assignment.py index 663de2ef..8bde0e65 100644 --- a/sdk/python/pulumi_databricks/permission_assignment.py +++ b/sdk/python/pulumi_databricks/permission_assignment.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['PermissionAssignmentArgs', 'PermissionAssignment'] @@ -19,8 +19,19 @@ def __init__(__self__, *, """ The set of arguments for constructing a PermissionAssignment resource. """ - pulumi.set(__self__, "permissions", permissions) - pulumi.set(__self__, "principal_id", principal_id) + PermissionAssignmentArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + permissions=permissions, + principal_id=principal_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permissions: pulumi.Input[Sequence[pulumi.Input[str]]], + principal_id: pulumi.Input[int], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("permissions", permissions) + _setter("principal_id", principal_id) @property @pulumi.getter @@ -49,10 +60,21 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering PermissionAssignment resources. """ + _PermissionAssignmentState._configure( + lambda key, value: pulumi.set(__self__, key, value), + permissions=permissions, + principal_id=principal_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permissions: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + principal_id: Optional[pulumi.Input[int]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if permissions is not None: - pulumi.set(__self__, "permissions", permissions) + _setter("permissions", permissions) if principal_id is not None: - pulumi.set(__self__, "principal_id", principal_id) + _setter("principal_id", principal_id) @property @pulumi.getter @@ -104,6 +126,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + PermissionAssignmentArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/permissions.py b/sdk/python/pulumi_databricks/permissions.py index fe686272..4780ca7c 100644 --- a/sdk/python/pulumi_databricks/permissions.py +++ b/sdk/python/pulumi_databricks/permissions.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -62,51 +62,104 @@ def __init__(__self__, *, :param pulumi.Input[str] sql_endpoint_id: SQL warehouse id :param pulumi.Input[str] sql_query_id: SQL query id """ - pulumi.set(__self__, "access_controls", access_controls) + PermissionsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_controls=access_controls, + authorization=authorization, + cluster_id=cluster_id, + cluster_policy_id=cluster_policy_id, + directory_id=directory_id, + directory_path=directory_path, + experiment_id=experiment_id, + instance_pool_id=instance_pool_id, + job_id=job_id, + notebook_id=notebook_id, + notebook_path=notebook_path, + object_type=object_type, + pipeline_id=pipeline_id, + registered_model_id=registered_model_id, + repo_id=repo_id, + repo_path=repo_path, + serving_endpoint_id=serving_endpoint_id, + sql_alert_id=sql_alert_id, + sql_dashboard_id=sql_dashboard_id, + sql_endpoint_id=sql_endpoint_id, + sql_query_id=sql_query_id, + workspace_file_id=workspace_file_id, + workspace_file_path=workspace_file_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_controls: pulumi.Input[Sequence[pulumi.Input['PermissionsAccessControlArgs']]], + authorization: Optional[pulumi.Input[str]] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + cluster_policy_id: Optional[pulumi.Input[str]] = None, + directory_id: Optional[pulumi.Input[str]] = None, + directory_path: Optional[pulumi.Input[str]] = None, + experiment_id: Optional[pulumi.Input[str]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + job_id: Optional[pulumi.Input[str]] = None, + notebook_id: Optional[pulumi.Input[str]] = None, + notebook_path: Optional[pulumi.Input[str]] = None, + object_type: Optional[pulumi.Input[str]] = None, + pipeline_id: Optional[pulumi.Input[str]] = None, + registered_model_id: Optional[pulumi.Input[str]] = None, + repo_id: Optional[pulumi.Input[str]] = None, + repo_path: Optional[pulumi.Input[str]] = None, + serving_endpoint_id: Optional[pulumi.Input[str]] = None, + sql_alert_id: Optional[pulumi.Input[str]] = None, + sql_dashboard_id: Optional[pulumi.Input[str]] = None, + sql_endpoint_id: Optional[pulumi.Input[str]] = None, + sql_query_id: Optional[pulumi.Input[str]] = None, + workspace_file_id: Optional[pulumi.Input[str]] = None, + workspace_file_path: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("access_controls", access_controls) if authorization is not None: - pulumi.set(__self__, "authorization", authorization) + _setter("authorization", authorization) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_policy_id is not None: - pulumi.set(__self__, "cluster_policy_id", cluster_policy_id) + _setter("cluster_policy_id", cluster_policy_id) if directory_id is not None: - pulumi.set(__self__, "directory_id", directory_id) + _setter("directory_id", directory_id) if directory_path is not None: - pulumi.set(__self__, "directory_path", directory_path) + _setter("directory_path", directory_path) if experiment_id is not None: - pulumi.set(__self__, "experiment_id", experiment_id) + _setter("experiment_id", experiment_id) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if job_id is not None: - pulumi.set(__self__, "job_id", job_id) + _setter("job_id", job_id) if notebook_id is not None: - pulumi.set(__self__, "notebook_id", notebook_id) + _setter("notebook_id", notebook_id) if notebook_path is not None: - pulumi.set(__self__, "notebook_path", notebook_path) + _setter("notebook_path", notebook_path) if object_type is not None: - pulumi.set(__self__, "object_type", object_type) + _setter("object_type", object_type) if pipeline_id is not None: - pulumi.set(__self__, "pipeline_id", pipeline_id) + _setter("pipeline_id", pipeline_id) if registered_model_id is not None: - pulumi.set(__self__, "registered_model_id", registered_model_id) + _setter("registered_model_id", registered_model_id) if repo_id is not None: - pulumi.set(__self__, "repo_id", repo_id) + _setter("repo_id", repo_id) if repo_path is not None: - pulumi.set(__self__, "repo_path", repo_path) + _setter("repo_path", repo_path) if serving_endpoint_id is not None: - pulumi.set(__self__, "serving_endpoint_id", serving_endpoint_id) + _setter("serving_endpoint_id", serving_endpoint_id) if sql_alert_id is not None: - pulumi.set(__self__, "sql_alert_id", sql_alert_id) + _setter("sql_alert_id", sql_alert_id) if sql_dashboard_id is not None: - pulumi.set(__self__, "sql_dashboard_id", sql_dashboard_id) + _setter("sql_dashboard_id", sql_dashboard_id) if sql_endpoint_id is not None: - pulumi.set(__self__, "sql_endpoint_id", sql_endpoint_id) + _setter("sql_endpoint_id", sql_endpoint_id) if sql_query_id is not None: - pulumi.set(__self__, "sql_query_id", sql_query_id) + _setter("sql_query_id", sql_query_id) if workspace_file_id is not None: - pulumi.set(__self__, "workspace_file_id", workspace_file_id) + _setter("workspace_file_id", workspace_file_id) if workspace_file_path is not None: - pulumi.set(__self__, "workspace_file_path", workspace_file_path) + _setter("workspace_file_path", workspace_file_path) @property @pulumi.getter(name="accessControls") @@ -425,52 +478,105 @@ def __init__(__self__, *, :param pulumi.Input[str] sql_endpoint_id: SQL warehouse id :param pulumi.Input[str] sql_query_id: SQL query id """ + _PermissionsState._configure( + lambda key, value: pulumi.set(__self__, key, value), + access_controls=access_controls, + authorization=authorization, + cluster_id=cluster_id, + cluster_policy_id=cluster_policy_id, + directory_id=directory_id, + directory_path=directory_path, + experiment_id=experiment_id, + instance_pool_id=instance_pool_id, + job_id=job_id, + notebook_id=notebook_id, + notebook_path=notebook_path, + object_type=object_type, + pipeline_id=pipeline_id, + registered_model_id=registered_model_id, + repo_id=repo_id, + repo_path=repo_path, + serving_endpoint_id=serving_endpoint_id, + sql_alert_id=sql_alert_id, + sql_dashboard_id=sql_dashboard_id, + sql_endpoint_id=sql_endpoint_id, + sql_query_id=sql_query_id, + workspace_file_id=workspace_file_id, + workspace_file_path=workspace_file_path, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + access_controls: Optional[pulumi.Input[Sequence[pulumi.Input['PermissionsAccessControlArgs']]]] = None, + authorization: Optional[pulumi.Input[str]] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + cluster_policy_id: Optional[pulumi.Input[str]] = None, + directory_id: Optional[pulumi.Input[str]] = None, + directory_path: Optional[pulumi.Input[str]] = None, + experiment_id: Optional[pulumi.Input[str]] = None, + instance_pool_id: Optional[pulumi.Input[str]] = None, + job_id: Optional[pulumi.Input[str]] = None, + notebook_id: Optional[pulumi.Input[str]] = None, + notebook_path: Optional[pulumi.Input[str]] = None, + object_type: Optional[pulumi.Input[str]] = None, + pipeline_id: Optional[pulumi.Input[str]] = None, + registered_model_id: Optional[pulumi.Input[str]] = None, + repo_id: Optional[pulumi.Input[str]] = None, + repo_path: Optional[pulumi.Input[str]] = None, + serving_endpoint_id: Optional[pulumi.Input[str]] = None, + sql_alert_id: Optional[pulumi.Input[str]] = None, + sql_dashboard_id: Optional[pulumi.Input[str]] = None, + sql_endpoint_id: Optional[pulumi.Input[str]] = None, + sql_query_id: Optional[pulumi.Input[str]] = None, + workspace_file_id: Optional[pulumi.Input[str]] = None, + workspace_file_path: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if access_controls is not None: - pulumi.set(__self__, "access_controls", access_controls) + _setter("access_controls", access_controls) if authorization is not None: - pulumi.set(__self__, "authorization", authorization) + _setter("authorization", authorization) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if cluster_policy_id is not None: - pulumi.set(__self__, "cluster_policy_id", cluster_policy_id) + _setter("cluster_policy_id", cluster_policy_id) if directory_id is not None: - pulumi.set(__self__, "directory_id", directory_id) + _setter("directory_id", directory_id) if directory_path is not None: - pulumi.set(__self__, "directory_path", directory_path) + _setter("directory_path", directory_path) if experiment_id is not None: - pulumi.set(__self__, "experiment_id", experiment_id) + _setter("experiment_id", experiment_id) if instance_pool_id is not None: - pulumi.set(__self__, "instance_pool_id", instance_pool_id) + _setter("instance_pool_id", instance_pool_id) if job_id is not None: - pulumi.set(__self__, "job_id", job_id) + _setter("job_id", job_id) if notebook_id is not None: - pulumi.set(__self__, "notebook_id", notebook_id) + _setter("notebook_id", notebook_id) if notebook_path is not None: - pulumi.set(__self__, "notebook_path", notebook_path) + _setter("notebook_path", notebook_path) if object_type is not None: - pulumi.set(__self__, "object_type", object_type) + _setter("object_type", object_type) if pipeline_id is not None: - pulumi.set(__self__, "pipeline_id", pipeline_id) + _setter("pipeline_id", pipeline_id) if registered_model_id is not None: - pulumi.set(__self__, "registered_model_id", registered_model_id) + _setter("registered_model_id", registered_model_id) if repo_id is not None: - pulumi.set(__self__, "repo_id", repo_id) + _setter("repo_id", repo_id) if repo_path is not None: - pulumi.set(__self__, "repo_path", repo_path) + _setter("repo_path", repo_path) if serving_endpoint_id is not None: - pulumi.set(__self__, "serving_endpoint_id", serving_endpoint_id) + _setter("serving_endpoint_id", serving_endpoint_id) if sql_alert_id is not None: - pulumi.set(__self__, "sql_alert_id", sql_alert_id) + _setter("sql_alert_id", sql_alert_id) if sql_dashboard_id is not None: - pulumi.set(__self__, "sql_dashboard_id", sql_dashboard_id) + _setter("sql_dashboard_id", sql_dashboard_id) if sql_endpoint_id is not None: - pulumi.set(__self__, "sql_endpoint_id", sql_endpoint_id) + _setter("sql_endpoint_id", sql_endpoint_id) if sql_query_id is not None: - pulumi.set(__self__, "sql_query_id", sql_query_id) + _setter("sql_query_id", sql_query_id) if workspace_file_id is not None: - pulumi.set(__self__, "workspace_file_id", workspace_file_id) + _setter("workspace_file_id", workspace_file_id) if workspace_file_path is not None: - pulumi.set(__self__, "workspace_file_path", workspace_file_path) + _setter("workspace_file_path", workspace_file_path) @property @pulumi.getter(name="accessControls") @@ -862,6 +968,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + PermissionsArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/pipeline.py b/sdk/python/pulumi_databricks/pipeline.py index 829aa978..f71d787b 100644 --- a/sdk/python/pulumi_databricks/pipeline.py +++ b/sdk/python/pulumi_databricks/pipeline.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -47,38 +47,77 @@ def __init__(__self__, *, :param pulumi.Input[str] storage: A location on DBFS or cloud storage where output data and metadata required for pipeline execution are stored. By default, tables are stored in a subdirectory of this location. *Change of this parameter forces recreation of the pipeline.* (Conflicts with `catalog`). :param pulumi.Input[str] target: The name of a database (in either the Hive metastore or in a UC catalog) for persisting pipeline output data. Configuring the target setting allows you to view and query the pipeline output data from the Databricks UI. """ + PipelineArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + allow_duplicate_names=allow_duplicate_names, + catalog=catalog, + channel=channel, + clusters=clusters, + configuration=configuration, + continuous=continuous, + development=development, + edition=edition, + filters=filters, + libraries=libraries, + name=name, + notifications=notifications, + photon=photon, + serverless=serverless, + storage=storage, + target=target, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allow_duplicate_names: Optional[pulumi.Input[bool]] = None, + catalog: Optional[pulumi.Input[str]] = None, + channel: Optional[pulumi.Input[str]] = None, + clusters: Optional[pulumi.Input[Sequence[pulumi.Input['PipelineClusterArgs']]]] = None, + configuration: Optional[pulumi.Input[Mapping[str, Any]]] = None, + continuous: Optional[pulumi.Input[bool]] = None, + development: Optional[pulumi.Input[bool]] = None, + edition: Optional[pulumi.Input[str]] = None, + filters: Optional[pulumi.Input['PipelineFiltersArgs']] = None, + libraries: Optional[pulumi.Input[Sequence[pulumi.Input['PipelineLibraryArgs']]]] = None, + name: Optional[pulumi.Input[str]] = None, + notifications: Optional[pulumi.Input[Sequence[pulumi.Input['PipelineNotificationArgs']]]] = None, + photon: Optional[pulumi.Input[bool]] = None, + serverless: Optional[pulumi.Input[bool]] = None, + storage: Optional[pulumi.Input[str]] = None, + target: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if allow_duplicate_names is not None: - pulumi.set(__self__, "allow_duplicate_names", allow_duplicate_names) + _setter("allow_duplicate_names", allow_duplicate_names) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if channel is not None: - pulumi.set(__self__, "channel", channel) + _setter("channel", channel) if clusters is not None: - pulumi.set(__self__, "clusters", clusters) + _setter("clusters", clusters) if configuration is not None: - pulumi.set(__self__, "configuration", configuration) + _setter("configuration", configuration) if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if development is not None: - pulumi.set(__self__, "development", development) + _setter("development", development) if edition is not None: - pulumi.set(__self__, "edition", edition) + _setter("edition", edition) if filters is not None: - pulumi.set(__self__, "filters", filters) + _setter("filters", filters) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if notifications is not None: - pulumi.set(__self__, "notifications", notifications) + _setter("notifications", notifications) if photon is not None: - pulumi.set(__self__, "photon", photon) + _setter("photon", photon) if serverless is not None: - pulumi.set(__self__, "serverless", serverless) + _setter("serverless", serverless) if storage is not None: - pulumi.set(__self__, "storage", storage) + _setter("storage", storage) if target is not None: - pulumi.set(__self__, "target", target) + _setter("target", target) @property @pulumi.getter(name="allowDuplicateNames") @@ -296,40 +335,81 @@ def __init__(__self__, *, :param pulumi.Input[str] storage: A location on DBFS or cloud storage where output data and metadata required for pipeline execution are stored. By default, tables are stored in a subdirectory of this location. *Change of this parameter forces recreation of the pipeline.* (Conflicts with `catalog`). :param pulumi.Input[str] target: The name of a database (in either the Hive metastore or in a UC catalog) for persisting pipeline output data. Configuring the target setting allows you to view and query the pipeline output data from the Databricks UI. """ + _PipelineState._configure( + lambda key, value: pulumi.set(__self__, key, value), + allow_duplicate_names=allow_duplicate_names, + catalog=catalog, + channel=channel, + clusters=clusters, + configuration=configuration, + continuous=continuous, + development=development, + edition=edition, + filters=filters, + libraries=libraries, + name=name, + notifications=notifications, + photon=photon, + serverless=serverless, + storage=storage, + target=target, + url=url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + allow_duplicate_names: Optional[pulumi.Input[bool]] = None, + catalog: Optional[pulumi.Input[str]] = None, + channel: Optional[pulumi.Input[str]] = None, + clusters: Optional[pulumi.Input[Sequence[pulumi.Input['PipelineClusterArgs']]]] = None, + configuration: Optional[pulumi.Input[Mapping[str, Any]]] = None, + continuous: Optional[pulumi.Input[bool]] = None, + development: Optional[pulumi.Input[bool]] = None, + edition: Optional[pulumi.Input[str]] = None, + filters: Optional[pulumi.Input['PipelineFiltersArgs']] = None, + libraries: Optional[pulumi.Input[Sequence[pulumi.Input['PipelineLibraryArgs']]]] = None, + name: Optional[pulumi.Input[str]] = None, + notifications: Optional[pulumi.Input[Sequence[pulumi.Input['PipelineNotificationArgs']]]] = None, + photon: Optional[pulumi.Input[bool]] = None, + serverless: Optional[pulumi.Input[bool]] = None, + storage: Optional[pulumi.Input[str]] = None, + target: Optional[pulumi.Input[str]] = None, + url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if allow_duplicate_names is not None: - pulumi.set(__self__, "allow_duplicate_names", allow_duplicate_names) + _setter("allow_duplicate_names", allow_duplicate_names) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if channel is not None: - pulumi.set(__self__, "channel", channel) + _setter("channel", channel) if clusters is not None: - pulumi.set(__self__, "clusters", clusters) + _setter("clusters", clusters) if configuration is not None: - pulumi.set(__self__, "configuration", configuration) + _setter("configuration", configuration) if continuous is not None: - pulumi.set(__self__, "continuous", continuous) + _setter("continuous", continuous) if development is not None: - pulumi.set(__self__, "development", development) + _setter("development", development) if edition is not None: - pulumi.set(__self__, "edition", edition) + _setter("edition", edition) if filters is not None: - pulumi.set(__self__, "filters", filters) + _setter("filters", filters) if libraries is not None: - pulumi.set(__self__, "libraries", libraries) + _setter("libraries", libraries) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if notifications is not None: - pulumi.set(__self__, "notifications", notifications) + _setter("notifications", notifications) if photon is not None: - pulumi.set(__self__, "photon", photon) + _setter("photon", photon) if serverless is not None: - pulumi.set(__self__, "serverless", serverless) + _setter("serverless", serverless) if storage is not None: - pulumi.set(__self__, "storage", storage) + _setter("storage", storage) if target is not None: - pulumi.set(__self__, "target", target) + _setter("target", target) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) @property @pulumi.getter(name="allowDuplicateNames") @@ -732,6 +812,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + PipelineArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -770,6 +854,11 @@ def _internal_init(__self__, __props__.__dict__["continuous"] = continuous __props__.__dict__["development"] = development __props__.__dict__["edition"] = edition + if filters is not None and not isinstance(filters, PipelineFiltersArgs): + filters = filters or {} + def _setter(key, value): + filters[key] = value + PipelineFiltersArgs._configure(_setter, **filters) __props__.__dict__["filters"] = filters __props__.__dict__["libraries"] = libraries __props__.__dict__["name"] = name diff --git a/sdk/python/pulumi_databricks/provider.py b/sdk/python/pulumi_databricks/provider.py index 48130e91..1af38ffb 100644 --- a/sdk/python/pulumi_databricks/provider.py +++ b/sdk/python/pulumi_databricks/provider.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['ProviderArgs', 'Provider'] @@ -46,64 +46,129 @@ def __init__(__self__, *, """ The set of arguments for constructing a Provider resource. """ + ProviderArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + account_id=account_id, + auth_type=auth_type, + azure_client_id=azure_client_id, + azure_client_secret=azure_client_secret, + azure_environment=azure_environment, + azure_login_app_id=azure_login_app_id, + azure_tenant_id=azure_tenant_id, + azure_use_msi=azure_use_msi, + azure_workspace_resource_id=azure_workspace_resource_id, + client_id=client_id, + client_secret=client_secret, + cluster_id=cluster_id, + config_file=config_file, + databricks_cli_path=databricks_cli_path, + debug_headers=debug_headers, + debug_truncate_bytes=debug_truncate_bytes, + google_credentials=google_credentials, + google_service_account=google_service_account, + host=host, + http_timeout_seconds=http_timeout_seconds, + metadata_service_url=metadata_service_url, + password=password, + profile=profile, + rate_limit=rate_limit, + retry_timeout_seconds=retry_timeout_seconds, + skip_verify=skip_verify, + token=token, + username=username, + warehouse_id=warehouse_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + account_id: Optional[pulumi.Input[str]] = None, + auth_type: Optional[pulumi.Input[str]] = None, + azure_client_id: Optional[pulumi.Input[str]] = None, + azure_client_secret: Optional[pulumi.Input[str]] = None, + azure_environment: Optional[pulumi.Input[str]] = None, + azure_login_app_id: Optional[pulumi.Input[str]] = None, + azure_tenant_id: Optional[pulumi.Input[str]] = None, + azure_use_msi: Optional[pulumi.Input[bool]] = None, + azure_workspace_resource_id: Optional[pulumi.Input[str]] = None, + client_id: Optional[pulumi.Input[str]] = None, + client_secret: Optional[pulumi.Input[str]] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + config_file: Optional[pulumi.Input[str]] = None, + databricks_cli_path: Optional[pulumi.Input[str]] = None, + debug_headers: Optional[pulumi.Input[bool]] = None, + debug_truncate_bytes: Optional[pulumi.Input[int]] = None, + google_credentials: Optional[pulumi.Input[str]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + host: Optional[pulumi.Input[str]] = None, + http_timeout_seconds: Optional[pulumi.Input[int]] = None, + metadata_service_url: Optional[pulumi.Input[str]] = None, + password: Optional[pulumi.Input[str]] = None, + profile: Optional[pulumi.Input[str]] = None, + rate_limit: Optional[pulumi.Input[int]] = None, + retry_timeout_seconds: Optional[pulumi.Input[int]] = None, + skip_verify: Optional[pulumi.Input[bool]] = None, + token: Optional[pulumi.Input[str]] = None, + username: Optional[pulumi.Input[str]] = None, + warehouse_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if account_id is not None: - pulumi.set(__self__, "account_id", account_id) + _setter("account_id", account_id) if auth_type is not None: - pulumi.set(__self__, "auth_type", auth_type) + _setter("auth_type", auth_type) if azure_client_id is not None: - pulumi.set(__self__, "azure_client_id", azure_client_id) + _setter("azure_client_id", azure_client_id) if azure_client_secret is not None: - pulumi.set(__self__, "azure_client_secret", azure_client_secret) + _setter("azure_client_secret", azure_client_secret) if azure_environment is not None: - pulumi.set(__self__, "azure_environment", azure_environment) + _setter("azure_environment", azure_environment) if azure_login_app_id is not None: - pulumi.set(__self__, "azure_login_app_id", azure_login_app_id) + _setter("azure_login_app_id", azure_login_app_id) if azure_tenant_id is not None: - pulumi.set(__self__, "azure_tenant_id", azure_tenant_id) + _setter("azure_tenant_id", azure_tenant_id) if azure_use_msi is not None: - pulumi.set(__self__, "azure_use_msi", azure_use_msi) + _setter("azure_use_msi", azure_use_msi) if azure_workspace_resource_id is not None: - pulumi.set(__self__, "azure_workspace_resource_id", azure_workspace_resource_id) + _setter("azure_workspace_resource_id", azure_workspace_resource_id) if client_id is not None: - pulumi.set(__self__, "client_id", client_id) + _setter("client_id", client_id) if client_secret is not None: - pulumi.set(__self__, "client_secret", client_secret) + _setter("client_secret", client_secret) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if config_file is not None: - pulumi.set(__self__, "config_file", config_file) + _setter("config_file", config_file) if databricks_cli_path is not None: - pulumi.set(__self__, "databricks_cli_path", databricks_cli_path) + _setter("databricks_cli_path", databricks_cli_path) if debug_headers is not None: - pulumi.set(__self__, "debug_headers", debug_headers) + _setter("debug_headers", debug_headers) if debug_truncate_bytes is not None: - pulumi.set(__self__, "debug_truncate_bytes", debug_truncate_bytes) + _setter("debug_truncate_bytes", debug_truncate_bytes) if google_credentials is not None: - pulumi.set(__self__, "google_credentials", google_credentials) + _setter("google_credentials", google_credentials) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if host is not None: - pulumi.set(__self__, "host", host) + _setter("host", host) if http_timeout_seconds is not None: - pulumi.set(__self__, "http_timeout_seconds", http_timeout_seconds) + _setter("http_timeout_seconds", http_timeout_seconds) if metadata_service_url is not None: - pulumi.set(__self__, "metadata_service_url", metadata_service_url) + _setter("metadata_service_url", metadata_service_url) if password is not None: - pulumi.set(__self__, "password", password) + _setter("password", password) if profile is not None: - pulumi.set(__self__, "profile", profile) + _setter("profile", profile) if rate_limit is not None: - pulumi.set(__self__, "rate_limit", rate_limit) + _setter("rate_limit", rate_limit) if retry_timeout_seconds is not None: - pulumi.set(__self__, "retry_timeout_seconds", retry_timeout_seconds) + _setter("retry_timeout_seconds", retry_timeout_seconds) if skip_verify is not None: - pulumi.set(__self__, "skip_verify", skip_verify) + _setter("skip_verify", skip_verify) if token is not None: - pulumi.set(__self__, "token", token) + _setter("token", token) if username is not None: - pulumi.set(__self__, "username", username) + _setter("username", username) if warehouse_id is not None: - pulumi.set(__self__, "warehouse_id", warehouse_id) + _setter("warehouse_id", warehouse_id) @property @pulumi.getter(name="accountId") @@ -433,6 +498,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ProviderArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/recipient.py b/sdk/python/pulumi_databricks/recipient.py index 85518a1d..51b791e4 100644 --- a/sdk/python/pulumi_databricks/recipient.py +++ b/sdk/python/pulumi_databricks/recipient.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -33,19 +33,40 @@ def __init__(__self__, *, :param pulumi.Input[str] sharing_code: The one-time sharing code provided by the data recipient. :param pulumi.Input[Sequence[pulumi.Input['RecipientTokenArgs']]] tokens: List of Recipient Tokens. This field is only present when the authentication_type is TOKEN. Each list element is an object with following attributes: """ - pulumi.set(__self__, "authentication_type", authentication_type) + RecipientArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + authentication_type=authentication_type, + comment=comment, + data_recipient_global_metastore_id=data_recipient_global_metastore_id, + ip_access_list=ip_access_list, + name=name, + sharing_code=sharing_code, + tokens=tokens, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + authentication_type: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + data_recipient_global_metastore_id: Optional[pulumi.Input[str]] = None, + ip_access_list: Optional[pulumi.Input['RecipientIpAccessListArgs']] = None, + name: Optional[pulumi.Input[str]] = None, + sharing_code: Optional[pulumi.Input[str]] = None, + tokens: Optional[pulumi.Input[Sequence[pulumi.Input['RecipientTokenArgs']]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("authentication_type", authentication_type) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if data_recipient_global_metastore_id is not None: - pulumi.set(__self__, "data_recipient_global_metastore_id", data_recipient_global_metastore_id) + _setter("data_recipient_global_metastore_id", data_recipient_global_metastore_id) if ip_access_list is not None: - pulumi.set(__self__, "ip_access_list", ip_access_list) + _setter("ip_access_list", ip_access_list) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if sharing_code is not None: - pulumi.set(__self__, "sharing_code", sharing_code) + _setter("sharing_code", sharing_code) if tokens is not None: - pulumi.set(__self__, "tokens", tokens) + _setter("tokens", tokens) @property @pulumi.getter(name="authenticationType") @@ -152,20 +173,41 @@ def __init__(__self__, *, :param pulumi.Input[str] sharing_code: The one-time sharing code provided by the data recipient. :param pulumi.Input[Sequence[pulumi.Input['RecipientTokenArgs']]] tokens: List of Recipient Tokens. This field is only present when the authentication_type is TOKEN. Each list element is an object with following attributes: """ + _RecipientState._configure( + lambda key, value: pulumi.set(__self__, key, value), + authentication_type=authentication_type, + comment=comment, + data_recipient_global_metastore_id=data_recipient_global_metastore_id, + ip_access_list=ip_access_list, + name=name, + sharing_code=sharing_code, + tokens=tokens, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + authentication_type: Optional[pulumi.Input[str]] = None, + comment: Optional[pulumi.Input[str]] = None, + data_recipient_global_metastore_id: Optional[pulumi.Input[str]] = None, + ip_access_list: Optional[pulumi.Input['RecipientIpAccessListArgs']] = None, + name: Optional[pulumi.Input[str]] = None, + sharing_code: Optional[pulumi.Input[str]] = None, + tokens: Optional[pulumi.Input[Sequence[pulumi.Input['RecipientTokenArgs']]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if authentication_type is not None: - pulumi.set(__self__, "authentication_type", authentication_type) + _setter("authentication_type", authentication_type) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if data_recipient_global_metastore_id is not None: - pulumi.set(__self__, "data_recipient_global_metastore_id", data_recipient_global_metastore_id) + _setter("data_recipient_global_metastore_id", data_recipient_global_metastore_id) if ip_access_list is not None: - pulumi.set(__self__, "ip_access_list", ip_access_list) + _setter("ip_access_list", ip_access_list) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if sharing_code is not None: - pulumi.set(__self__, "sharing_code", sharing_code) + _setter("sharing_code", sharing_code) if tokens is not None: - pulumi.set(__self__, "tokens", tokens) + _setter("tokens", tokens) @property @pulumi.getter(name="authenticationType") @@ -363,6 +405,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + RecipientArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -389,6 +435,11 @@ def _internal_init(__self__, __props__.__dict__["authentication_type"] = authentication_type __props__.__dict__["comment"] = comment __props__.__dict__["data_recipient_global_metastore_id"] = data_recipient_global_metastore_id + if ip_access_list is not None and not isinstance(ip_access_list, RecipientIpAccessListArgs): + ip_access_list = ip_access_list or {} + def _setter(key, value): + ip_access_list[key] = value + RecipientIpAccessListArgs._configure(_setter, **ip_access_list) __props__.__dict__["ip_access_list"] = ip_access_list __props__.__dict__["name"] = name __props__.__dict__["sharing_code"] = None if sharing_code is None else pulumi.Output.secret(sharing_code) diff --git a/sdk/python/pulumi_databricks/repo.py b/sdk/python/pulumi_databricks/repo.py index 7e3b4979..f6601e74 100644 --- a/sdk/python/pulumi_databricks/repo.py +++ b/sdk/python/pulumi_databricks/repo.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -32,19 +32,40 @@ def __init__(__self__, *, :param pulumi.Input[str] path: path to put the checked out Repo. If not specified, then repo will be created in the user's repo directory (`/Repos//...`). If the value changes, repo is re-created. :param pulumi.Input[str] tag: name of the tag for initial checkout. Conflicts with `branch`. """ - pulumi.set(__self__, "url", url) + RepoArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + url=url, + branch=branch, + commit_hash=commit_hash, + git_provider=git_provider, + path=path, + sparse_checkout=sparse_checkout, + tag=tag, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + url: pulumi.Input[str], + branch: Optional[pulumi.Input[str]] = None, + commit_hash: Optional[pulumi.Input[str]] = None, + git_provider: Optional[pulumi.Input[str]] = None, + path: Optional[pulumi.Input[str]] = None, + sparse_checkout: Optional[pulumi.Input['RepoSparseCheckoutArgs']] = None, + tag: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("url", url) if branch is not None: - pulumi.set(__self__, "branch", branch) + _setter("branch", branch) if commit_hash is not None: - pulumi.set(__self__, "commit_hash", commit_hash) + _setter("commit_hash", commit_hash) if git_provider is not None: - pulumi.set(__self__, "git_provider", git_provider) + _setter("git_provider", git_provider) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) if sparse_checkout is not None: - pulumi.set(__self__, "sparse_checkout", sparse_checkout) + _setter("sparse_checkout", sparse_checkout) if tag is not None: - pulumi.set(__self__, "tag", tag) + _setter("tag", tag) @property @pulumi.getter @@ -147,20 +168,41 @@ def __init__(__self__, *, :param pulumi.Input[str] tag: name of the tag for initial checkout. Conflicts with `branch`. :param pulumi.Input[str] url: The URL of the Git Repository to clone from. If the value changes, repo is re-created. """ + _RepoState._configure( + lambda key, value: pulumi.set(__self__, key, value), + branch=branch, + commit_hash=commit_hash, + git_provider=git_provider, + path=path, + sparse_checkout=sparse_checkout, + tag=tag, + url=url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + branch: Optional[pulumi.Input[str]] = None, + commit_hash: Optional[pulumi.Input[str]] = None, + git_provider: Optional[pulumi.Input[str]] = None, + path: Optional[pulumi.Input[str]] = None, + sparse_checkout: Optional[pulumi.Input['RepoSparseCheckoutArgs']] = None, + tag: Optional[pulumi.Input[str]] = None, + url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if branch is not None: - pulumi.set(__self__, "branch", branch) + _setter("branch", branch) if commit_hash is not None: - pulumi.set(__self__, "commit_hash", commit_hash) + _setter("commit_hash", commit_hash) if git_provider is not None: - pulumi.set(__self__, "git_provider", git_provider) + _setter("git_provider", git_provider) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) if sparse_checkout is not None: - pulumi.set(__self__, "sparse_checkout", sparse_checkout) + _setter("sparse_checkout", sparse_checkout) if tag is not None: - pulumi.set(__self__, "tag", tag) + _setter("tag", tag) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) @property @pulumi.getter @@ -300,6 +342,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + RepoArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -325,6 +371,11 @@ def _internal_init(__self__, __props__.__dict__["commit_hash"] = commit_hash __props__.__dict__["git_provider"] = git_provider __props__.__dict__["path"] = path + if sparse_checkout is not None and not isinstance(sparse_checkout, RepoSparseCheckoutArgs): + sparse_checkout = sparse_checkout or {} + def _setter(key, value): + sparse_checkout[key] = value + RepoSparseCheckoutArgs._configure(_setter, **sparse_checkout) __props__.__dict__["sparse_checkout"] = sparse_checkout __props__.__dict__["tag"] = tag if url is None and not opts.urn: diff --git a/sdk/python/pulumi_databricks/schema.py b/sdk/python/pulumi_databricks/schema.py index 296528b8..2056b87c 100644 --- a/sdk/python/pulumi_databricks/schema.py +++ b/sdk/python/pulumi_databricks/schema.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['SchemaArgs', 'Schema'] @@ -32,21 +32,44 @@ def __init__(__self__, *, :param pulumi.Input[Mapping[str, Any]] properties: Extensible Schema properties. :param pulumi.Input[str] storage_root: Managed location of the schema. Location in cloud storage where data for managed tables will be stored. If not specified, the location will default to the metastore root location. Change forces creation of a new resource. """ - pulumi.set(__self__, "catalog_name", catalog_name) + SchemaArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + comment=comment, + force_destroy=force_destroy, + metastore_id=metastore_id, + name=name, + owner=owner, + properties=properties, + storage_root=storage_root, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + storage_root: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("catalog_name", catalog_name) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if storage_root is not None: - pulumi.set(__self__, "storage_root", storage_root) + _setter("storage_root", storage_root) @property @pulumi.getter(name="catalogName") @@ -163,22 +186,45 @@ def __init__(__self__, *, :param pulumi.Input[Mapping[str, Any]] properties: Extensible Schema properties. :param pulumi.Input[str] storage_root: Managed location of the schema. Location in cloud storage where data for managed tables will be stored. If not specified, the location will default to the metastore root location. Change forces creation of a new resource. """ + _SchemaState._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + comment=comment, + force_destroy=force_destroy, + metastore_id=metastore_id, + name=name, + owner=owner, + properties=properties, + storage_root=storage_root, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: Optional[pulumi.Input[str]] = None, + comment: Optional[pulumi.Input[str]] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + storage_root: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if catalog_name is not None: - pulumi.set(__self__, "catalog_name", catalog_name) + _setter("catalog_name", catalog_name) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if storage_root is not None: - pulumi.set(__self__, "storage_root", storage_root) + _setter("storage_root", storage_root) @property @pulumi.getter(name="catalogName") @@ -394,6 +440,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SchemaArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/secret.py b/sdk/python/pulumi_databricks/secret.py index 7a3852e9..30968869 100644 --- a/sdk/python/pulumi_databricks/secret.py +++ b/sdk/python/pulumi_databricks/secret.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['SecretArgs', 'Secret'] @@ -23,9 +23,22 @@ def __init__(__self__, *, :param pulumi.Input[str] scope: (String) name of databricks secret scope. Must consist of alphanumeric characters, dashes, underscores, and periods, and may not exceed 128 characters. :param pulumi.Input[str] string_value: (String) super secret sensitive value. """ - pulumi.set(__self__, "key", key) - pulumi.set(__self__, "scope", scope) - pulumi.set(__self__, "string_value", string_value) + SecretArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + key=key, + scope=scope, + string_value=string_value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + key: pulumi.Input[str], + scope: pulumi.Input[str], + string_value: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("key", key) + _setter("scope", scope) + _setter("string_value", string_value) @property @pulumi.getter @@ -80,16 +93,33 @@ def __init__(__self__, *, :param pulumi.Input[str] scope: (String) name of databricks secret scope. Must consist of alphanumeric characters, dashes, underscores, and periods, and may not exceed 128 characters. :param pulumi.Input[str] string_value: (String) super secret sensitive value. """ + _SecretState._configure( + lambda key, value: pulumi.set(__self__, key, value), + config_reference=config_reference, + key=key, + last_updated_timestamp=last_updated_timestamp, + scope=scope, + string_value=string_value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + config_reference: Optional[pulumi.Input[str]] = None, + key: Optional[pulumi.Input[str]] = None, + last_updated_timestamp: Optional[pulumi.Input[int]] = None, + scope: Optional[pulumi.Input[str]] = None, + string_value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if config_reference is not None: - pulumi.set(__self__, "config_reference", config_reference) + _setter("config_reference", config_reference) if key is not None: - pulumi.set(__self__, "key", key) + _setter("key", key) if last_updated_timestamp is not None: - pulumi.set(__self__, "last_updated_timestamp", last_updated_timestamp) + _setter("last_updated_timestamp", last_updated_timestamp) if scope is not None: - pulumi.set(__self__, "scope", scope) + _setter("scope", scope) if string_value is not None: - pulumi.set(__self__, "string_value", string_value) + _setter("string_value", string_value) @property @pulumi.getter(name="configReference") @@ -257,6 +287,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SecretArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/secret_acl.py b/sdk/python/pulumi_databricks/secret_acl.py index 6ec65cdd..ad63a0a0 100644 --- a/sdk/python/pulumi_databricks/secret_acl.py +++ b/sdk/python/pulumi_databricks/secret_acl.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['SecretAclArgs', 'SecretAcl'] @@ -23,9 +23,22 @@ def __init__(__self__, *, :param pulumi.Input[str] principal: name of the principals. It can be `users` for all users or name or `display_name` of databricks_group :param pulumi.Input[str] scope: name of the scope """ - pulumi.set(__self__, "permission", permission) - pulumi.set(__self__, "principal", principal) - pulumi.set(__self__, "scope", scope) + SecretAclArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + permission=permission, + principal=principal, + scope=scope, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permission: pulumi.Input[str], + principal: pulumi.Input[str], + scope: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("permission", permission) + _setter("principal", principal) + _setter("scope", scope) @property @pulumi.getter @@ -76,12 +89,25 @@ def __init__(__self__, *, :param pulumi.Input[str] principal: name of the principals. It can be `users` for all users or name or `display_name` of databricks_group :param pulumi.Input[str] scope: name of the scope """ + _SecretAclState._configure( + lambda key, value: pulumi.set(__self__, key, value), + permission=permission, + principal=principal, + scope=scope, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + permission: Optional[pulumi.Input[str]] = None, + principal: Optional[pulumi.Input[str]] = None, + scope: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if permission is not None: - pulumi.set(__self__, "permission", permission) + _setter("permission", permission) if principal is not None: - pulumi.set(__self__, "principal", principal) + _setter("principal", principal) if scope is not None: - pulumi.set(__self__, "scope", scope) + _setter("scope", scope) @property @pulumi.getter @@ -233,6 +259,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SecretAclArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/secret_scope.py b/sdk/python/pulumi_databricks/secret_scope.py index e57d3286..5ebb81b9 100644 --- a/sdk/python/pulumi_databricks/secret_scope.py +++ b/sdk/python/pulumi_databricks/secret_scope.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -26,14 +26,29 @@ def __init__(__self__, *, :param pulumi.Input[str] initial_manage_principal: The principal with the only possible value `users` that is initially granted `MANAGE` permission to the created scope. If it's omitted, then the SecretAcl with `MANAGE` permission applied to the scope is assigned to the API request issuer's user identity (see [documentation](https://docs.databricks.com/dev-tools/api/latest/secrets.html#create-secret-scope)). This part of the state cannot be imported. :param pulumi.Input[str] name: Scope name requested by the user. Must be unique within a workspace. Must consist of alphanumeric characters, dashes, underscores, and periods, and may not exceed 128 characters. """ + SecretScopeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + backend_type=backend_type, + initial_manage_principal=initial_manage_principal, + keyvault_metadata=keyvault_metadata, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + backend_type: Optional[pulumi.Input[str]] = None, + initial_manage_principal: Optional[pulumi.Input[str]] = None, + keyvault_metadata: Optional[pulumi.Input['SecretScopeKeyvaultMetadataArgs']] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if backend_type is not None: - pulumi.set(__self__, "backend_type", backend_type) + _setter("backend_type", backend_type) if initial_manage_principal is not None: - pulumi.set(__self__, "initial_manage_principal", initial_manage_principal) + _setter("initial_manage_principal", initial_manage_principal) if keyvault_metadata is not None: - pulumi.set(__self__, "keyvault_metadata", keyvault_metadata) + _setter("keyvault_metadata", keyvault_metadata) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter(name="backendType") @@ -94,14 +109,29 @@ def __init__(__self__, *, :param pulumi.Input[str] initial_manage_principal: The principal with the only possible value `users` that is initially granted `MANAGE` permission to the created scope. If it's omitted, then the SecretAcl with `MANAGE` permission applied to the scope is assigned to the API request issuer's user identity (see [documentation](https://docs.databricks.com/dev-tools/api/latest/secrets.html#create-secret-scope)). This part of the state cannot be imported. :param pulumi.Input[str] name: Scope name requested by the user. Must be unique within a workspace. Must consist of alphanumeric characters, dashes, underscores, and periods, and may not exceed 128 characters. """ + _SecretScopeState._configure( + lambda key, value: pulumi.set(__self__, key, value), + backend_type=backend_type, + initial_manage_principal=initial_manage_principal, + keyvault_metadata=keyvault_metadata, + name=name, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + backend_type: Optional[pulumi.Input[str]] = None, + initial_manage_principal: Optional[pulumi.Input[str]] = None, + keyvault_metadata: Optional[pulumi.Input['SecretScopeKeyvaultMetadataArgs']] = None, + name: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if backend_type is not None: - pulumi.set(__self__, "backend_type", backend_type) + _setter("backend_type", backend_type) if initial_manage_principal is not None: - pulumi.set(__self__, "initial_manage_principal", initial_manage_principal) + _setter("initial_manage_principal", initial_manage_principal) if keyvault_metadata is not None: - pulumi.set(__self__, "keyvault_metadata", keyvault_metadata) + _setter("keyvault_metadata", keyvault_metadata) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) @property @pulumi.getter(name="backendType") @@ -199,6 +229,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SecretScopeArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -219,6 +253,11 @@ def _internal_init(__self__, __props__.__dict__["backend_type"] = backend_type __props__.__dict__["initial_manage_principal"] = initial_manage_principal + if keyvault_metadata is not None and not isinstance(keyvault_metadata, SecretScopeKeyvaultMetadataArgs): + keyvault_metadata = keyvault_metadata or {} + def _setter(key, value): + keyvault_metadata[key] = value + SecretScopeKeyvaultMetadataArgs._configure(_setter, **keyvault_metadata) __props__.__dict__["keyvault_metadata"] = keyvault_metadata __props__.__dict__["name"] = name super(SecretScope, __self__).__init__( diff --git a/sdk/python/pulumi_databricks/service_principal.py b/sdk/python/pulumi_databricks/service_principal.py index b2096fd1..6f62d95e 100644 --- a/sdk/python/pulumi_databricks/service_principal.py +++ b/sdk/python/pulumi_databricks/service_principal.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['ServicePrincipalArgs', 'ServicePrincipal'] @@ -46,36 +46,73 @@ def __init__(__self__, *, :param pulumi.Input[str] repos: Personal Repos location of the service principal, e.g. `/Repos/00000000-0000-0000-0000-000000000000`. :param pulumi.Input[bool] workspace_access: This is a field to allow the group to have access to Databricks Workspace. """ + ServicePrincipalArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + acl_principal_id=acl_principal_id, + active=active, + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + application_id=application_id, + databricks_sql_access=databricks_sql_access, + disable_as_user_deletion=disable_as_user_deletion, + display_name=display_name, + external_id=external_id, + force=force, + force_delete_home_dir=force_delete_home_dir, + force_delete_repos=force_delete_repos, + home=home, + repos=repos, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + acl_principal_id: Optional[pulumi.Input[str]] = None, + active: Optional[pulumi.Input[bool]] = None, + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + application_id: Optional[pulumi.Input[str]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + disable_as_user_deletion: Optional[pulumi.Input[bool]] = None, + display_name: Optional[pulumi.Input[str]] = None, + external_id: Optional[pulumi.Input[str]] = None, + force: Optional[pulumi.Input[bool]] = None, + force_delete_home_dir: Optional[pulumi.Input[bool]] = None, + force_delete_repos: Optional[pulumi.Input[bool]] = None, + home: Optional[pulumi.Input[str]] = None, + repos: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if acl_principal_id is not None: - pulumi.set(__self__, "acl_principal_id", acl_principal_id) + _setter("acl_principal_id", acl_principal_id) if active is not None: - pulumi.set(__self__, "active", active) + _setter("active", active) if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if application_id is not None: - pulumi.set(__self__, "application_id", application_id) + _setter("application_id", application_id) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if disable_as_user_deletion is not None: - pulumi.set(__self__, "disable_as_user_deletion", disable_as_user_deletion) + _setter("disable_as_user_deletion", disable_as_user_deletion) if display_name is not None: - pulumi.set(__self__, "display_name", display_name) + _setter("display_name", display_name) if external_id is not None: - pulumi.set(__self__, "external_id", external_id) + _setter("external_id", external_id) if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if force_delete_home_dir is not None: - pulumi.set(__self__, "force_delete_home_dir", force_delete_home_dir) + _setter("force_delete_home_dir", force_delete_home_dir) if force_delete_repos is not None: - pulumi.set(__self__, "force_delete_repos", force_delete_repos) + _setter("force_delete_repos", force_delete_repos) if home is not None: - pulumi.set(__self__, "home", home) + _setter("home", home) if repos is not None: - pulumi.set(__self__, "repos", repos) + _setter("repos", repos) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="aclPrincipalId") @@ -290,36 +327,73 @@ def __init__(__self__, *, :param pulumi.Input[str] repos: Personal Repos location of the service principal, e.g. `/Repos/00000000-0000-0000-0000-000000000000`. :param pulumi.Input[bool] workspace_access: This is a field to allow the group to have access to Databricks Workspace. """ + _ServicePrincipalState._configure( + lambda key, value: pulumi.set(__self__, key, value), + acl_principal_id=acl_principal_id, + active=active, + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + application_id=application_id, + databricks_sql_access=databricks_sql_access, + disable_as_user_deletion=disable_as_user_deletion, + display_name=display_name, + external_id=external_id, + force=force, + force_delete_home_dir=force_delete_home_dir, + force_delete_repos=force_delete_repos, + home=home, + repos=repos, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + acl_principal_id: Optional[pulumi.Input[str]] = None, + active: Optional[pulumi.Input[bool]] = None, + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + application_id: Optional[pulumi.Input[str]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + disable_as_user_deletion: Optional[pulumi.Input[bool]] = None, + display_name: Optional[pulumi.Input[str]] = None, + external_id: Optional[pulumi.Input[str]] = None, + force: Optional[pulumi.Input[bool]] = None, + force_delete_home_dir: Optional[pulumi.Input[bool]] = None, + force_delete_repos: Optional[pulumi.Input[bool]] = None, + home: Optional[pulumi.Input[str]] = None, + repos: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if acl_principal_id is not None: - pulumi.set(__self__, "acl_principal_id", acl_principal_id) + _setter("acl_principal_id", acl_principal_id) if active is not None: - pulumi.set(__self__, "active", active) + _setter("active", active) if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if application_id is not None: - pulumi.set(__self__, "application_id", application_id) + _setter("application_id", application_id) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if disable_as_user_deletion is not None: - pulumi.set(__self__, "disable_as_user_deletion", disable_as_user_deletion) + _setter("disable_as_user_deletion", disable_as_user_deletion) if display_name is not None: - pulumi.set(__self__, "display_name", display_name) + _setter("display_name", display_name) if external_id is not None: - pulumi.set(__self__, "external_id", external_id) + _setter("external_id", external_id) if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if force_delete_home_dir is not None: - pulumi.set(__self__, "force_delete_home_dir", force_delete_home_dir) + _setter("force_delete_home_dir", force_delete_home_dir) if force_delete_repos is not None: - pulumi.set(__self__, "force_delete_repos", force_delete_repos) + _setter("force_delete_repos", force_delete_repos) if home is not None: - pulumi.set(__self__, "home", home) + _setter("home", home) if repos is not None: - pulumi.set(__self__, "repos", repos) + _setter("repos", repos) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="aclPrincipalId") @@ -737,6 +811,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ServicePrincipalArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/service_principal_role.py b/sdk/python/pulumi_databricks/service_principal_role.py index 8a6c5d98..f968f43b 100644 --- a/sdk/python/pulumi_databricks/service_principal_role.py +++ b/sdk/python/pulumi_databricks/service_principal_role.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['ServicePrincipalRoleArgs', 'ServicePrincipalRole'] @@ -21,8 +21,19 @@ def __init__(__self__, *, :param pulumi.Input[str] role: This is the id of the role or instance profile resource. :param pulumi.Input[str] service_principal_id: This is the id of the service principal resource. """ - pulumi.set(__self__, "role", role) - pulumi.set(__self__, "service_principal_id", service_principal_id) + ServicePrincipalRoleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + role=role, + service_principal_id=service_principal_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role: pulumi.Input[str], + service_principal_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role", role) + _setter("service_principal_id", service_principal_id) @property @pulumi.getter @@ -59,10 +70,21 @@ def __init__(__self__, *, :param pulumi.Input[str] role: This is the id of the role or instance profile resource. :param pulumi.Input[str] service_principal_id: This is the id of the service principal resource. """ + _ServicePrincipalRoleState._configure( + lambda key, value: pulumi.set(__self__, key, value), + role=role, + service_principal_id=service_principal_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role: Optional[pulumi.Input[str]] = None, + service_principal_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if role is not None: - pulumi.set(__self__, "role", role) + _setter("role", role) if service_principal_id is not None: - pulumi.set(__self__, "service_principal_id", service_principal_id) + _setter("service_principal_id", service_principal_id) @property @pulumi.getter @@ -180,6 +202,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ServicePrincipalRoleArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/service_principal_secret.py b/sdk/python/pulumi_databricks/service_principal_secret.py index e27406d2..6b3a81d7 100644 --- a/sdk/python/pulumi_databricks/service_principal_secret.py +++ b/sdk/python/pulumi_databricks/service_principal_secret.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['ServicePrincipalSecretArgs', 'ServicePrincipalSecret'] @@ -20,11 +20,24 @@ def __init__(__self__, *, """ The set of arguments for constructing a ServicePrincipalSecret resource. """ - pulumi.set(__self__, "service_principal_id", service_principal_id) + ServicePrincipalSecretArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + service_principal_id=service_principal_id, + secret=secret, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + service_principal_id: pulumi.Input[str], + secret: Optional[pulumi.Input[str]] = None, + status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("service_principal_id", service_principal_id) if secret is not None: - pulumi.set(__self__, "secret", secret) + _setter("secret", secret) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter(name="servicePrincipalId") @@ -63,12 +76,25 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering ServicePrincipalSecret resources. """ + _ServicePrincipalSecretState._configure( + lambda key, value: pulumi.set(__self__, key, value), + secret=secret, + service_principal_id=service_principal_id, + status=status, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + secret: Optional[pulumi.Input[str]] = None, + service_principal_id: Optional[pulumi.Input[str]] = None, + status: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if secret is not None: - pulumi.set(__self__, "secret", secret) + _setter("secret", secret) if service_principal_id is not None: - pulumi.set(__self__, "service_principal_id", service_principal_id) + _setter("service_principal_id", service_principal_id) if status is not None: - pulumi.set(__self__, "status", status) + _setter("status", status) @property @pulumi.getter @@ -130,6 +156,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ServicePrincipalSecretArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/share.py b/sdk/python/pulumi_databricks/share.py index 8e9b4bf5..33281e05 100644 --- a/sdk/python/pulumi_databricks/share.py +++ b/sdk/python/pulumi_databricks/share.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -28,16 +28,33 @@ def __init__(__self__, *, :param pulumi.Input[str] name: Name of share. Change forces creation of a new resource. :param pulumi.Input[str] owner: User name/group name/sp application_id of the share owner. """ + ShareArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + created_at=created_at, + created_by=created_by, + name=name, + objects=objects, + owner=owner, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + created_at: Optional[pulumi.Input[int]] = None, + created_by: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + objects: Optional[pulumi.Input[Sequence[pulumi.Input['ShareObjectArgs']]]] = None, + owner: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if objects is not None: - pulumi.set(__self__, "objects", objects) + _setter("objects", objects) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) @property @pulumi.getter(name="createdAt") @@ -112,16 +129,33 @@ def __init__(__self__, *, :param pulumi.Input[str] name: Name of share. Change forces creation of a new resource. :param pulumi.Input[str] owner: User name/group name/sp application_id of the share owner. """ + _ShareState._configure( + lambda key, value: pulumi.set(__self__, key, value), + created_at=created_at, + created_by=created_by, + name=name, + objects=objects, + owner=owner, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + created_at: Optional[pulumi.Input[int]] = None, + created_by: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + objects: Optional[pulumi.Input[Sequence[pulumi.Input['ShareObjectArgs']]]] = None, + owner: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if created_by is not None: - pulumi.set(__self__, "created_by", created_by) + _setter("created_by", created_by) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if objects is not None: - pulumi.set(__self__, "objects", objects) + _setter("objects", objects) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) @property @pulumi.getter(name="createdAt") @@ -219,6 +253,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + ShareArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/sql_alert.py b/sdk/python/pulumi_databricks/sql_alert.py index 15ed2fe5..ba5226e6 100644 --- a/sdk/python/pulumi_databricks/sql_alert.py +++ b/sdk/python/pulumi_databricks/sql_alert.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -31,18 +31,39 @@ def __init__(__self__, *, :param pulumi.Input[str] parent: The identifier of the workspace folder containing the alert. The default is ther user's home folder. The folder identifier is formatted as `folder/`. :param pulumi.Input[int] rearm: Number of seconds after being triggered before the alert rearms itself and can be triggered again. If not defined, alert will never be triggered again. """ - pulumi.set(__self__, "options", options) - pulumi.set(__self__, "query_id", query_id) + SqlAlertArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + options=options, + query_id=query_id, + created_at=created_at, + name=name, + parent=parent, + rearm=rearm, + updated_at=updated_at, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + options: pulumi.Input['SqlAlertOptionsArgs'], + query_id: pulumi.Input[str], + created_at: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + parent: Optional[pulumi.Input[str]] = None, + rearm: Optional[pulumi.Input[int]] = None, + updated_at: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("options", options) + _setter("query_id", query_id) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if parent is not None: - pulumi.set(__self__, "parent", parent) + _setter("parent", parent) if rearm is not None: - pulumi.set(__self__, "rearm", rearm) + _setter("rearm", rearm) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) @property @pulumi.getter @@ -141,20 +162,41 @@ def __init__(__self__, *, :param pulumi.Input[str] query_id: ID of the query evaluated by the alert. :param pulumi.Input[int] rearm: Number of seconds after being triggered before the alert rearms itself and can be triggered again. If not defined, alert will never be triggered again. """ + _SqlAlertState._configure( + lambda key, value: pulumi.set(__self__, key, value), + created_at=created_at, + name=name, + options=options, + parent=parent, + query_id=query_id, + rearm=rearm, + updated_at=updated_at, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + created_at: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + options: Optional[pulumi.Input['SqlAlertOptionsArgs']] = None, + parent: Optional[pulumi.Input[str]] = None, + query_id: Optional[pulumi.Input[str]] = None, + rearm: Optional[pulumi.Input[int]] = None, + updated_at: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if options is not None: - pulumi.set(__self__, "options", options) + _setter("options", options) if parent is not None: - pulumi.set(__self__, "parent", parent) + _setter("parent", parent) if query_id is not None: - pulumi.set(__self__, "query_id", query_id) + _setter("query_id", query_id) if rearm is not None: - pulumi.set(__self__, "rearm", rearm) + _setter("rearm", rearm) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) @property @pulumi.getter(name="createdAt") @@ -300,6 +342,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlAlertArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -323,6 +369,11 @@ def _internal_init(__self__, __props__.__dict__["created_at"] = created_at __props__.__dict__["name"] = name + if options is not None and not isinstance(options, SqlAlertOptionsArgs): + options = options or {} + def _setter(key, value): + options[key] = value + SqlAlertOptionsArgs._configure(_setter, **options) if options is None and not opts.urn: raise TypeError("Missing required property 'options'") __props__.__dict__["options"] = options diff --git a/sdk/python/pulumi_databricks/sql_dashboard.py b/sdk/python/pulumi_databricks/sql_dashboard.py index 1fa9f220..29fe17bb 100644 --- a/sdk/python/pulumi_databricks/sql_dashboard.py +++ b/sdk/python/pulumi_databricks/sql_dashboard.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['SqlDashboardArgs', 'SqlDashboard'] @@ -22,16 +22,33 @@ def __init__(__self__, *, """ The set of arguments for constructing a SqlDashboard resource. """ + SqlDashboardArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + created_at=created_at, + name=name, + parent=parent, + tags=tags, + updated_at=updated_at, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + created_at: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + parent: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + updated_at: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if parent is not None: - pulumi.set(__self__, "parent", parent) + _setter("parent", parent) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) @property @pulumi.getter(name="createdAt") @@ -90,16 +107,33 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering SqlDashboard resources. """ + _SqlDashboardState._configure( + lambda key, value: pulumi.set(__self__, key, value), + created_at=created_at, + name=name, + parent=parent, + tags=tags, + updated_at=updated_at, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + created_at: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + parent: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + updated_at: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if parent is not None: - pulumi.set(__self__, "parent", parent) + _setter("parent", parent) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) @property @pulumi.getter(name="createdAt") @@ -281,6 +315,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlDashboardArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/sql_endpoint.py b/sdk/python/pulumi_databricks/sql_endpoint.py index 99aa2669..fe4faf60 100644 --- a/sdk/python/pulumi_databricks/sql_endpoint.py +++ b/sdk/python/pulumi_databricks/sql_endpoint.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -54,39 +54,80 @@ def __init__(__self__, *, :param pulumi.Input['SqlEndpointTagsArgs'] tags: Databricks tags all endpoint resources with these tags. :param pulumi.Input[str] warehouse_type: SQL warehouse type. See for [AWS](https://docs.databricks.com/sql/admin/sql-endpoints.html#switch-the-sql-warehouse-type-pro-classic-or-serverless) or [Azure](https://learn.microsoft.com/en-us/azure/databricks/sql/admin/create-sql-warehouse#--upgrade-a-pro-or-classic-sql-warehouse-to-a-serverless-sql-warehouse). Set to `PRO` or `CLASSIC`. If the field `enable_serverless_compute` has the value `true` either explicitly or through the default logic (see that field above for details), the default is `PRO`, which is required for serverless SQL warehouses. Otherwise, the default is `CLASSIC`. """ - pulumi.set(__self__, "cluster_size", cluster_size) + SqlEndpointArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + cluster_size=cluster_size, + auto_stop_mins=auto_stop_mins, + channel=channel, + data_source_id=data_source_id, + enable_photon=enable_photon, + enable_serverless_compute=enable_serverless_compute, + instance_profile_arn=instance_profile_arn, + jdbc_url=jdbc_url, + max_num_clusters=max_num_clusters, + min_num_clusters=min_num_clusters, + name=name, + num_clusters=num_clusters, + odbc_params=odbc_params, + spot_instance_policy=spot_instance_policy, + state=state, + tags=tags, + warehouse_type=warehouse_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + cluster_size: pulumi.Input[str], + auto_stop_mins: Optional[pulumi.Input[int]] = None, + channel: Optional[pulumi.Input['SqlEndpointChannelArgs']] = None, + data_source_id: Optional[pulumi.Input[str]] = None, + enable_photon: Optional[pulumi.Input[bool]] = None, + enable_serverless_compute: Optional[pulumi.Input[bool]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + jdbc_url: Optional[pulumi.Input[str]] = None, + max_num_clusters: Optional[pulumi.Input[int]] = None, + min_num_clusters: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + num_clusters: Optional[pulumi.Input[int]] = None, + odbc_params: Optional[pulumi.Input['SqlEndpointOdbcParamsArgs']] = None, + spot_instance_policy: Optional[pulumi.Input[str]] = None, + state: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input['SqlEndpointTagsArgs']] = None, + warehouse_type: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("cluster_size", cluster_size) if auto_stop_mins is not None: - pulumi.set(__self__, "auto_stop_mins", auto_stop_mins) + _setter("auto_stop_mins", auto_stop_mins) if channel is not None: - pulumi.set(__self__, "channel", channel) + _setter("channel", channel) if data_source_id is not None: - pulumi.set(__self__, "data_source_id", data_source_id) + _setter("data_source_id", data_source_id) if enable_photon is not None: - pulumi.set(__self__, "enable_photon", enable_photon) + _setter("enable_photon", enable_photon) if enable_serverless_compute is not None: - pulumi.set(__self__, "enable_serverless_compute", enable_serverless_compute) + _setter("enable_serverless_compute", enable_serverless_compute) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if jdbc_url is not None: - pulumi.set(__self__, "jdbc_url", jdbc_url) + _setter("jdbc_url", jdbc_url) if max_num_clusters is not None: - pulumi.set(__self__, "max_num_clusters", max_num_clusters) + _setter("max_num_clusters", max_num_clusters) if min_num_clusters is not None: - pulumi.set(__self__, "min_num_clusters", min_num_clusters) + _setter("min_num_clusters", min_num_clusters) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if num_clusters is not None: - pulumi.set(__self__, "num_clusters", num_clusters) + _setter("num_clusters", num_clusters) if odbc_params is not None: - pulumi.set(__self__, "odbc_params", odbc_params) + _setter("odbc_params", odbc_params) if spot_instance_policy is not None: - pulumi.set(__self__, "spot_instance_policy", spot_instance_policy) + _setter("spot_instance_policy", spot_instance_policy) if state is not None: - pulumi.set(__self__, "state", state) + _setter("state", state) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if warehouse_type is not None: - pulumi.set(__self__, "warehouse_type", warehouse_type) + _setter("warehouse_type", warehouse_type) @property @pulumi.getter(name="clusterSize") @@ -329,40 +370,81 @@ def __init__(__self__, *, :param pulumi.Input['SqlEndpointTagsArgs'] tags: Databricks tags all endpoint resources with these tags. :param pulumi.Input[str] warehouse_type: SQL warehouse type. See for [AWS](https://docs.databricks.com/sql/admin/sql-endpoints.html#switch-the-sql-warehouse-type-pro-classic-or-serverless) or [Azure](https://learn.microsoft.com/en-us/azure/databricks/sql/admin/create-sql-warehouse#--upgrade-a-pro-or-classic-sql-warehouse-to-a-serverless-sql-warehouse). Set to `PRO` or `CLASSIC`. If the field `enable_serverless_compute` has the value `true` either explicitly or through the default logic (see that field above for details), the default is `PRO`, which is required for serverless SQL warehouses. Otherwise, the default is `CLASSIC`. """ + _SqlEndpointState._configure( + lambda key, value: pulumi.set(__self__, key, value), + auto_stop_mins=auto_stop_mins, + channel=channel, + cluster_size=cluster_size, + data_source_id=data_source_id, + enable_photon=enable_photon, + enable_serverless_compute=enable_serverless_compute, + instance_profile_arn=instance_profile_arn, + jdbc_url=jdbc_url, + max_num_clusters=max_num_clusters, + min_num_clusters=min_num_clusters, + name=name, + num_clusters=num_clusters, + odbc_params=odbc_params, + spot_instance_policy=spot_instance_policy, + state=state, + tags=tags, + warehouse_type=warehouse_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + auto_stop_mins: Optional[pulumi.Input[int]] = None, + channel: Optional[pulumi.Input['SqlEndpointChannelArgs']] = None, + cluster_size: Optional[pulumi.Input[str]] = None, + data_source_id: Optional[pulumi.Input[str]] = None, + enable_photon: Optional[pulumi.Input[bool]] = None, + enable_serverless_compute: Optional[pulumi.Input[bool]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + jdbc_url: Optional[pulumi.Input[str]] = None, + max_num_clusters: Optional[pulumi.Input[int]] = None, + min_num_clusters: Optional[pulumi.Input[int]] = None, + name: Optional[pulumi.Input[str]] = None, + num_clusters: Optional[pulumi.Input[int]] = None, + odbc_params: Optional[pulumi.Input['SqlEndpointOdbcParamsArgs']] = None, + spot_instance_policy: Optional[pulumi.Input[str]] = None, + state: Optional[pulumi.Input[str]] = None, + tags: Optional[pulumi.Input['SqlEndpointTagsArgs']] = None, + warehouse_type: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if auto_stop_mins is not None: - pulumi.set(__self__, "auto_stop_mins", auto_stop_mins) + _setter("auto_stop_mins", auto_stop_mins) if channel is not None: - pulumi.set(__self__, "channel", channel) + _setter("channel", channel) if cluster_size is not None: - pulumi.set(__self__, "cluster_size", cluster_size) + _setter("cluster_size", cluster_size) if data_source_id is not None: - pulumi.set(__self__, "data_source_id", data_source_id) + _setter("data_source_id", data_source_id) if enable_photon is not None: - pulumi.set(__self__, "enable_photon", enable_photon) + _setter("enable_photon", enable_photon) if enable_serverless_compute is not None: - pulumi.set(__self__, "enable_serverless_compute", enable_serverless_compute) + _setter("enable_serverless_compute", enable_serverless_compute) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if jdbc_url is not None: - pulumi.set(__self__, "jdbc_url", jdbc_url) + _setter("jdbc_url", jdbc_url) if max_num_clusters is not None: - pulumi.set(__self__, "max_num_clusters", max_num_clusters) + _setter("max_num_clusters", max_num_clusters) if min_num_clusters is not None: - pulumi.set(__self__, "min_num_clusters", min_num_clusters) + _setter("min_num_clusters", min_num_clusters) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if num_clusters is not None: - pulumi.set(__self__, "num_clusters", num_clusters) + _setter("num_clusters", num_clusters) if odbc_params is not None: - pulumi.set(__self__, "odbc_params", odbc_params) + _setter("odbc_params", odbc_params) if spot_instance_policy is not None: - pulumi.set(__self__, "spot_instance_policy", spot_instance_policy) + _setter("spot_instance_policy", spot_instance_policy) if state is not None: - pulumi.set(__self__, "state", state) + _setter("state", state) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if warehouse_type is not None: - pulumi.set(__self__, "warehouse_type", warehouse_type) + _setter("warehouse_type", warehouse_type) @property @pulumi.getter(name="autoStopMins") @@ -710,6 +792,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlEndpointArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -742,6 +828,11 @@ def _internal_init(__self__, __props__ = SqlEndpointArgs.__new__(SqlEndpointArgs) __props__.__dict__["auto_stop_mins"] = auto_stop_mins + if channel is not None and not isinstance(channel, SqlEndpointChannelArgs): + channel = channel or {} + def _setter(key, value): + channel[key] = value + SqlEndpointChannelArgs._configure(_setter, **channel) __props__.__dict__["channel"] = channel if cluster_size is None and not opts.urn: raise TypeError("Missing required property 'cluster_size'") @@ -755,9 +846,19 @@ def _internal_init(__self__, __props__.__dict__["min_num_clusters"] = min_num_clusters __props__.__dict__["name"] = name __props__.__dict__["num_clusters"] = num_clusters + if odbc_params is not None and not isinstance(odbc_params, SqlEndpointOdbcParamsArgs): + odbc_params = odbc_params or {} + def _setter(key, value): + odbc_params[key] = value + SqlEndpointOdbcParamsArgs._configure(_setter, **odbc_params) __props__.__dict__["odbc_params"] = odbc_params __props__.__dict__["spot_instance_policy"] = spot_instance_policy __props__.__dict__["state"] = state + if tags is not None and not isinstance(tags, SqlEndpointTagsArgs): + tags = tags or {} + def _setter(key, value): + tags[key] = value + SqlEndpointTagsArgs._configure(_setter, **tags) __props__.__dict__["tags"] = tags __props__.__dict__["warehouse_type"] = warehouse_type super(SqlEndpoint, __self__).__init__( diff --git a/sdk/python/pulumi_databricks/sql_global_config.py b/sdk/python/pulumi_databricks/sql_global_config.py index 194abf71..08774bae 100644 --- a/sdk/python/pulumi_databricks/sql_global_config.py +++ b/sdk/python/pulumi_databricks/sql_global_config.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['SqlGlobalConfigArgs', 'SqlGlobalConfig'] @@ -28,21 +28,40 @@ def __init__(__self__, *, :param pulumi.Input[str] security_policy: The policy for controlling access to datasets. Default value: `DATA_ACCESS_CONTROL`, consult documentation for list of possible values :param pulumi.Input[Mapping[str, Any]] sql_config_params: SQL Configuration Parameters let you override the default behavior for all sessions with all endpoints. """ + SqlGlobalConfigArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + data_access_config=data_access_config, + enable_serverless_compute=enable_serverless_compute, + google_service_account=google_service_account, + instance_profile_arn=instance_profile_arn, + security_policy=security_policy, + sql_config_params=sql_config_params, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + data_access_config: Optional[pulumi.Input[Mapping[str, Any]]] = None, + enable_serverless_compute: Optional[pulumi.Input[bool]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + security_policy: Optional[pulumi.Input[str]] = None, + sql_config_params: Optional[pulumi.Input[Mapping[str, Any]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if data_access_config is not None: - pulumi.set(__self__, "data_access_config", data_access_config) + _setter("data_access_config", data_access_config) if enable_serverless_compute is not None: warnings.warn("""This field is intended as an internal API and may be removed from the Databricks Terraform provider in the future""", DeprecationWarning) pulumi.log.warn("""enable_serverless_compute is deprecated: This field is intended as an internal API and may be removed from the Databricks Terraform provider in the future""") if enable_serverless_compute is not None: - pulumi.set(__self__, "enable_serverless_compute", enable_serverless_compute) + _setter("enable_serverless_compute", enable_serverless_compute) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if security_policy is not None: - pulumi.set(__self__, "security_policy", security_policy) + _setter("security_policy", security_policy) if sql_config_params is not None: - pulumi.set(__self__, "sql_config_params", sql_config_params) + _setter("sql_config_params", sql_config_params) @property @pulumi.getter(name="dataAccessConfig") @@ -134,21 +153,40 @@ def __init__(__self__, *, :param pulumi.Input[str] security_policy: The policy for controlling access to datasets. Default value: `DATA_ACCESS_CONTROL`, consult documentation for list of possible values :param pulumi.Input[Mapping[str, Any]] sql_config_params: SQL Configuration Parameters let you override the default behavior for all sessions with all endpoints. """ + _SqlGlobalConfigState._configure( + lambda key, value: pulumi.set(__self__, key, value), + data_access_config=data_access_config, + enable_serverless_compute=enable_serverless_compute, + google_service_account=google_service_account, + instance_profile_arn=instance_profile_arn, + security_policy=security_policy, + sql_config_params=sql_config_params, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + data_access_config: Optional[pulumi.Input[Mapping[str, Any]]] = None, + enable_serverless_compute: Optional[pulumi.Input[bool]] = None, + google_service_account: Optional[pulumi.Input[str]] = None, + instance_profile_arn: Optional[pulumi.Input[str]] = None, + security_policy: Optional[pulumi.Input[str]] = None, + sql_config_params: Optional[pulumi.Input[Mapping[str, Any]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if data_access_config is not None: - pulumi.set(__self__, "data_access_config", data_access_config) + _setter("data_access_config", data_access_config) if enable_serverless_compute is not None: warnings.warn("""This field is intended as an internal API and may be removed from the Databricks Terraform provider in the future""", DeprecationWarning) pulumi.log.warn("""enable_serverless_compute is deprecated: This field is intended as an internal API and may be removed from the Databricks Terraform provider in the future""") if enable_serverless_compute is not None: - pulumi.set(__self__, "enable_serverless_compute", enable_serverless_compute) + _setter("enable_serverless_compute", enable_serverless_compute) if google_service_account is not None: - pulumi.set(__self__, "google_service_account", google_service_account) + _setter("google_service_account", google_service_account) if instance_profile_arn is not None: - pulumi.set(__self__, "instance_profile_arn", instance_profile_arn) + _setter("instance_profile_arn", instance_profile_arn) if security_policy is not None: - pulumi.set(__self__, "security_policy", security_policy) + _setter("security_policy", security_policy) if sql_config_params is not None: - pulumi.set(__self__, "sql_config_params", sql_config_params) + _setter("sql_config_params", sql_config_params) @property @pulumi.getter(name="dataAccessConfig") @@ -371,6 +409,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlGlobalConfigArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -392,9 +434,6 @@ def _internal_init(__self__, __props__ = SqlGlobalConfigArgs.__new__(SqlGlobalConfigArgs) __props__.__dict__["data_access_config"] = data_access_config - if enable_serverless_compute is not None and not opts.urn: - warnings.warn("""This field is intended as an internal API and may be removed from the Databricks Terraform provider in the future""", DeprecationWarning) - pulumi.log.warn("""enable_serverless_compute is deprecated: This field is intended as an internal API and may be removed from the Databricks Terraform provider in the future""") __props__.__dict__["enable_serverless_compute"] = enable_serverless_compute __props__.__dict__["google_service_account"] = google_service_account __props__.__dict__["instance_profile_arn"] = instance_profile_arn diff --git a/sdk/python/pulumi_databricks/sql_permissions.py b/sdk/python/pulumi_databricks/sql_permissions.py index a95f0c4c..16c6e8ab 100644 --- a/sdk/python/pulumi_databricks/sql_permissions.py +++ b/sdk/python/pulumi_databricks/sql_permissions.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -33,22 +33,45 @@ def __init__(__self__, *, :param pulumi.Input[str] table: Name of the table. Can be combined with `database`. :param pulumi.Input[str] view: Name of the view. Can be combined with `database`. """ + SqlPermissionsArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + anonymous_function=anonymous_function, + any_file=any_file, + catalog=catalog, + cluster_id=cluster_id, + database=database, + privilege_assignments=privilege_assignments, + table=table, + view=view, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + anonymous_function: Optional[pulumi.Input[bool]] = None, + any_file: Optional[pulumi.Input[bool]] = None, + catalog: Optional[pulumi.Input[bool]] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + database: Optional[pulumi.Input[str]] = None, + privilege_assignments: Optional[pulumi.Input[Sequence[pulumi.Input['SqlPermissionsPrivilegeAssignmentArgs']]]] = None, + table: Optional[pulumi.Input[str]] = None, + view: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if anonymous_function is not None: - pulumi.set(__self__, "anonymous_function", anonymous_function) + _setter("anonymous_function", anonymous_function) if any_file is not None: - pulumi.set(__self__, "any_file", any_file) + _setter("any_file", any_file) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if database is not None: - pulumi.set(__self__, "database", database) + _setter("database", database) if privilege_assignments is not None: - pulumi.set(__self__, "privilege_assignments", privilege_assignments) + _setter("privilege_assignments", privilege_assignments) if table is not None: - pulumi.set(__self__, "table", table) + _setter("table", table) if view is not None: - pulumi.set(__self__, "view", view) + _setter("view", view) @property @pulumi.getter(name="anonymousFunction") @@ -161,22 +184,45 @@ def __init__(__self__, *, :param pulumi.Input[str] table: Name of the table. Can be combined with `database`. :param pulumi.Input[str] view: Name of the view. Can be combined with `database`. """ + _SqlPermissionsState._configure( + lambda key, value: pulumi.set(__self__, key, value), + anonymous_function=anonymous_function, + any_file=any_file, + catalog=catalog, + cluster_id=cluster_id, + database=database, + privilege_assignments=privilege_assignments, + table=table, + view=view, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + anonymous_function: Optional[pulumi.Input[bool]] = None, + any_file: Optional[pulumi.Input[bool]] = None, + catalog: Optional[pulumi.Input[bool]] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + database: Optional[pulumi.Input[str]] = None, + privilege_assignments: Optional[pulumi.Input[Sequence[pulumi.Input['SqlPermissionsPrivilegeAssignmentArgs']]]] = None, + table: Optional[pulumi.Input[str]] = None, + view: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if anonymous_function is not None: - pulumi.set(__self__, "anonymous_function", anonymous_function) + _setter("anonymous_function", anonymous_function) if any_file is not None: - pulumi.set(__self__, "any_file", any_file) + _setter("any_file", any_file) if catalog is not None: - pulumi.set(__self__, "catalog", catalog) + _setter("catalog", catalog) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if database is not None: - pulumi.set(__self__, "database", database) + _setter("database", database) if privilege_assignments is not None: - pulumi.set(__self__, "privilege_assignments", privilege_assignments) + _setter("privilege_assignments", privilege_assignments) if table is not None: - pulumi.set(__self__, "table", table) + _setter("table", table) if view is not None: - pulumi.set(__self__, "view", view) + _setter("view", view) @property @pulumi.getter(name="anonymousFunction") @@ -404,6 +450,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlPermissionsArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/sql_query.py b/sdk/python/pulumi_databricks/sql_query.py index 4937bd30..88619da0 100644 --- a/sdk/python/pulumi_databricks/sql_query.py +++ b/sdk/python/pulumi_databricks/sql_query.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -30,29 +30,58 @@ def __init__(__self__, *, """ The set of arguments for constructing a SqlQuery resource. """ - pulumi.set(__self__, "data_source_id", data_source_id) - pulumi.set(__self__, "query", query) + SqlQueryArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + data_source_id=data_source_id, + query=query, + created_at=created_at, + description=description, + name=name, + parameters=parameters, + parent=parent, + run_as_role=run_as_role, + schedule=schedule, + tags=tags, + updated_at=updated_at, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + data_source_id: pulumi.Input[str], + query: pulumi.Input[str], + created_at: Optional[pulumi.Input[str]] = None, + description: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input['SqlQueryParameterArgs']]]] = None, + parent: Optional[pulumi.Input[str]] = None, + run_as_role: Optional[pulumi.Input[str]] = None, + schedule: Optional[pulumi.Input['SqlQueryScheduleArgs']] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + updated_at: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("data_source_id", data_source_id) + _setter("query", query) if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if parent is not None: - pulumi.set(__self__, "parent", parent) + _setter("parent", parent) if run_as_role is not None: - pulumi.set(__self__, "run_as_role", run_as_role) + _setter("run_as_role", run_as_role) if schedule is not None: warnings.warn("""Operations on `databricks_sql_query` schedules are deprecated. Please use `databricks_job` resource to schedule a `sql_task`.""", DeprecationWarning) pulumi.log.warn("""schedule is deprecated: Operations on `databricks_sql_query` schedules are deprecated. Please use `databricks_job` resource to schedule a `sql_task`.""") if schedule is not None: - pulumi.set(__self__, "schedule", schedule) + _setter("schedule", schedule) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) @property @pulumi.getter(name="dataSourceId") @@ -174,31 +203,60 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering SqlQuery resources. """ + _SqlQueryState._configure( + lambda key, value: pulumi.set(__self__, key, value), + created_at=created_at, + data_source_id=data_source_id, + description=description, + name=name, + parameters=parameters, + parent=parent, + query=query, + run_as_role=run_as_role, + schedule=schedule, + tags=tags, + updated_at=updated_at, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + created_at: Optional[pulumi.Input[str]] = None, + data_source_id: Optional[pulumi.Input[str]] = None, + description: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input['SqlQueryParameterArgs']]]] = None, + parent: Optional[pulumi.Input[str]] = None, + query: Optional[pulumi.Input[str]] = None, + run_as_role: Optional[pulumi.Input[str]] = None, + schedule: Optional[pulumi.Input['SqlQueryScheduleArgs']] = None, + tags: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None, + updated_at: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if created_at is not None: - pulumi.set(__self__, "created_at", created_at) + _setter("created_at", created_at) if data_source_id is not None: - pulumi.set(__self__, "data_source_id", data_source_id) + _setter("data_source_id", data_source_id) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if parent is not None: - pulumi.set(__self__, "parent", parent) + _setter("parent", parent) if query is not None: - pulumi.set(__self__, "query", query) + _setter("query", query) if run_as_role is not None: - pulumi.set(__self__, "run_as_role", run_as_role) + _setter("run_as_role", run_as_role) if schedule is not None: warnings.warn("""Operations on `databricks_sql_query` schedules are deprecated. Please use `databricks_job` resource to schedule a `sql_task`.""", DeprecationWarning) pulumi.log.warn("""schedule is deprecated: Operations on `databricks_sql_query` schedules are deprecated. Please use `databricks_job` resource to schedule a `sql_task`.""") if schedule is not None: - pulumi.set(__self__, "schedule", schedule) + _setter("schedule", schedule) if tags is not None: - pulumi.set(__self__, "tags", tags) + _setter("tags", tags) if updated_at is not None: - pulumi.set(__self__, "updated_at", updated_at) + _setter("updated_at", updated_at) @property @pulumi.getter(name="createdAt") @@ -551,6 +609,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlQueryArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -588,9 +650,11 @@ def _internal_init(__self__, raise TypeError("Missing required property 'query'") __props__.__dict__["query"] = query __props__.__dict__["run_as_role"] = run_as_role - if schedule is not None and not opts.urn: - warnings.warn("""Operations on `databricks_sql_query` schedules are deprecated. Please use `databricks_job` resource to schedule a `sql_task`.""", DeprecationWarning) - pulumi.log.warn("""schedule is deprecated: Operations on `databricks_sql_query` schedules are deprecated. Please use `databricks_job` resource to schedule a `sql_task`.""") + if schedule is not None and not isinstance(schedule, SqlQueryScheduleArgs): + schedule = schedule or {} + def _setter(key, value): + schedule[key] = value + SqlQueryScheduleArgs._configure(_setter, **schedule) __props__.__dict__["schedule"] = schedule __props__.__dict__["tags"] = tags __props__.__dict__["updated_at"] = updated_at diff --git a/sdk/python/pulumi_databricks/sql_table.py b/sdk/python/pulumi_databricks/sql_table.py index 3950ac01..322f69bf 100644 --- a/sdk/python/pulumi_databricks/sql_table.py +++ b/sdk/python/pulumi_databricks/sql_table.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -41,27 +41,58 @@ def __init__(__self__, *, :param pulumi.Input[str] storage_location: URL of storage location for Table data (required for EXTERNAL Tables). Not supported for `VIEW` or `MANAGED` table_type. :param pulumi.Input[str] view_definition: SQL text defining the view (for `table_type == "VIEW"`). Not supported for `MANAGED` or `EXTERNAL` table_type. """ - pulumi.set(__self__, "catalog_name", catalog_name) - pulumi.set(__self__, "schema_name", schema_name) - pulumi.set(__self__, "table_type", table_type) + SqlTableArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + schema_name=schema_name, + table_type=table_type, + cluster_id=cluster_id, + columns=columns, + comment=comment, + data_source_format=data_source_format, + name=name, + properties=properties, + storage_credential_name=storage_credential_name, + storage_location=storage_location, + view_definition=view_definition, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: pulumi.Input[str], + schema_name: pulumi.Input[str], + table_type: pulumi.Input[str], + cluster_id: Optional[pulumi.Input[str]] = None, + columns: Optional[pulumi.Input[Sequence[pulumi.Input['SqlTableColumnArgs']]]] = None, + comment: Optional[pulumi.Input[str]] = None, + data_source_format: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + storage_credential_name: Optional[pulumi.Input[str]] = None, + storage_location: Optional[pulumi.Input[str]] = None, + view_definition: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("catalog_name", catalog_name) + _setter("schema_name", schema_name) + _setter("table_type", table_type) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if columns is not None: - pulumi.set(__self__, "columns", columns) + _setter("columns", columns) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if data_source_format is not None: - pulumi.set(__self__, "data_source_format", data_source_format) + _setter("data_source_format", data_source_format) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if storage_credential_name is not None: - pulumi.set(__self__, "storage_credential_name", storage_credential_name) + _setter("storage_credential_name", storage_credential_name) if storage_location is not None: - pulumi.set(__self__, "storage_location", storage_location) + _setter("storage_location", storage_location) if view_definition is not None: - pulumi.set(__self__, "view_definition", view_definition) + _setter("view_definition", view_definition) @property @pulumi.getter(name="catalogName") @@ -230,30 +261,61 @@ def __init__(__self__, *, :param pulumi.Input[str] table_type: Distinguishes a view vs. managed/external Table. `MANAGED`, `EXTERNAL` or `VIEW`. Change forces creation of a new resource. :param pulumi.Input[str] view_definition: SQL text defining the view (for `table_type == "VIEW"`). Not supported for `MANAGED` or `EXTERNAL` table_type. """ + _SqlTableState._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + cluster_id=cluster_id, + columns=columns, + comment=comment, + data_source_format=data_source_format, + name=name, + properties=properties, + schema_name=schema_name, + storage_credential_name=storage_credential_name, + storage_location=storage_location, + table_type=table_type, + view_definition=view_definition, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: Optional[pulumi.Input[str]] = None, + cluster_id: Optional[pulumi.Input[str]] = None, + columns: Optional[pulumi.Input[Sequence[pulumi.Input['SqlTableColumnArgs']]]] = None, + comment: Optional[pulumi.Input[str]] = None, + data_source_format: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + schema_name: Optional[pulumi.Input[str]] = None, + storage_credential_name: Optional[pulumi.Input[str]] = None, + storage_location: Optional[pulumi.Input[str]] = None, + table_type: Optional[pulumi.Input[str]] = None, + view_definition: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if catalog_name is not None: - pulumi.set(__self__, "catalog_name", catalog_name) + _setter("catalog_name", catalog_name) if cluster_id is not None: - pulumi.set(__self__, "cluster_id", cluster_id) + _setter("cluster_id", cluster_id) if columns is not None: - pulumi.set(__self__, "columns", columns) + _setter("columns", columns) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if data_source_format is not None: - pulumi.set(__self__, "data_source_format", data_source_format) + _setter("data_source_format", data_source_format) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if schema_name is not None: - pulumi.set(__self__, "schema_name", schema_name) + _setter("schema_name", schema_name) if storage_credential_name is not None: - pulumi.set(__self__, "storage_credential_name", storage_credential_name) + _setter("storage_credential_name", storage_credential_name) if storage_location is not None: - pulumi.set(__self__, "storage_location", storage_location) + _setter("storage_location", storage_location) if table_type is not None: - pulumi.set(__self__, "table_type", table_type) + _setter("table_type", table_type) if view_definition is not None: - pulumi.set(__self__, "view_definition", view_definition) + _setter("view_definition", view_definition) @property @pulumi.getter(name="catalogName") @@ -471,6 +533,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlTableArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/sql_visualization.py b/sdk/python/pulumi_databricks/sql_visualization.py index e6f36013..97e147fc 100644 --- a/sdk/python/pulumi_databricks/sql_visualization.py +++ b/sdk/python/pulumi_databricks/sql_visualization.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['SqlVisualizationArgs', 'SqlVisualization'] @@ -24,17 +24,38 @@ def __init__(__self__, *, """ The set of arguments for constructing a SqlVisualization resource. """ - pulumi.set(__self__, "options", options) - pulumi.set(__self__, "query_id", query_id) - pulumi.set(__self__, "type", type) + SqlVisualizationArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + options=options, + query_id=query_id, + type=type, + description=description, + name=name, + query_plan=query_plan, + visualization_id=visualization_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + options: pulumi.Input[str], + query_id: pulumi.Input[str], + type: pulumi.Input[str], + description: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + query_plan: Optional[pulumi.Input[str]] = None, + visualization_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("options", options) + _setter("query_id", query_id) + _setter("type", type) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if query_plan is not None: - pulumi.set(__self__, "query_plan", query_plan) + _setter("query_plan", query_plan) if visualization_id is not None: - pulumi.set(__self__, "visualization_id", visualization_id) + _setter("visualization_id", visualization_id) @property @pulumi.getter @@ -113,20 +134,41 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering SqlVisualization resources. """ + _SqlVisualizationState._configure( + lambda key, value: pulumi.set(__self__, key, value), + description=description, + name=name, + options=options, + query_id=query_id, + query_plan=query_plan, + type=type, + visualization_id=visualization_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + description: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + options: Optional[pulumi.Input[str]] = None, + query_id: Optional[pulumi.Input[str]] = None, + query_plan: Optional[pulumi.Input[str]] = None, + type: Optional[pulumi.Input[str]] = None, + visualization_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if options is not None: - pulumi.set(__self__, "options", options) + _setter("options", options) if query_id is not None: - pulumi.set(__self__, "query_id", query_id) + _setter("query_id", query_id) if query_plan is not None: - pulumi.set(__self__, "query_plan", query_plan) + _setter("query_plan", query_plan) if type is not None: - pulumi.set(__self__, "type", type) + _setter("type", type) if visualization_id is not None: - pulumi.set(__self__, "visualization_id", visualization_id) + _setter("visualization_id", visualization_id) @property @pulumi.getter @@ -242,6 +284,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlVisualizationArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/sql_widget.py b/sdk/python/pulumi_databricks/sql_widget.py index ca4e1b0c..1b3a14d1 100644 --- a/sdk/python/pulumi_databricks/sql_widget.py +++ b/sdk/python/pulumi_databricks/sql_widget.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -27,21 +27,44 @@ def __init__(__self__, *, """ The set of arguments for constructing a SqlWidget resource. """ - pulumi.set(__self__, "dashboard_id", dashboard_id) + SqlWidgetArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + dashboard_id=dashboard_id, + description=description, + parameters=parameters, + position=position, + text=text, + title=title, + visualization_id=visualization_id, + widget_id=widget_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dashboard_id: pulumi.Input[str], + description: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input['SqlWidgetParameterArgs']]]] = None, + position: Optional[pulumi.Input['SqlWidgetPositionArgs']] = None, + text: Optional[pulumi.Input[str]] = None, + title: Optional[pulumi.Input[str]] = None, + visualization_id: Optional[pulumi.Input[str]] = None, + widget_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("dashboard_id", dashboard_id) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if position is not None: - pulumi.set(__self__, "position", position) + _setter("position", position) if text is not None: - pulumi.set(__self__, "text", text) + _setter("text", text) if title is not None: - pulumi.set(__self__, "title", title) + _setter("title", title) if visualization_id is not None: - pulumi.set(__self__, "visualization_id", visualization_id) + _setter("visualization_id", visualization_id) if widget_id is not None: - pulumi.set(__self__, "widget_id", widget_id) + _setter("widget_id", widget_id) @property @pulumi.getter(name="dashboardId") @@ -130,22 +153,45 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering SqlWidget resources. """ + _SqlWidgetState._configure( + lambda key, value: pulumi.set(__self__, key, value), + dashboard_id=dashboard_id, + description=description, + parameters=parameters, + position=position, + text=text, + title=title, + visualization_id=visualization_id, + widget_id=widget_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + dashboard_id: Optional[pulumi.Input[str]] = None, + description: Optional[pulumi.Input[str]] = None, + parameters: Optional[pulumi.Input[Sequence[pulumi.Input['SqlWidgetParameterArgs']]]] = None, + position: Optional[pulumi.Input['SqlWidgetPositionArgs']] = None, + text: Optional[pulumi.Input[str]] = None, + title: Optional[pulumi.Input[str]] = None, + visualization_id: Optional[pulumi.Input[str]] = None, + widget_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if dashboard_id is not None: - pulumi.set(__self__, "dashboard_id", dashboard_id) + _setter("dashboard_id", dashboard_id) if description is not None: - pulumi.set(__self__, "description", description) + _setter("description", description) if parameters is not None: - pulumi.set(__self__, "parameters", parameters) + _setter("parameters", parameters) if position is not None: - pulumi.set(__self__, "position", position) + _setter("position", position) if text is not None: - pulumi.set(__self__, "text", text) + _setter("text", text) if title is not None: - pulumi.set(__self__, "title", title) + _setter("title", title) if visualization_id is not None: - pulumi.set(__self__, "visualization_id", visualization_id) + _setter("visualization_id", visualization_id) if widget_id is not None: - pulumi.set(__self__, "widget_id", widget_id) + _setter("widget_id", widget_id) @property @pulumi.getter(name="dashboardId") @@ -353,6 +399,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + SqlWidgetArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -380,6 +430,11 @@ def _internal_init(__self__, __props__.__dict__["dashboard_id"] = dashboard_id __props__.__dict__["description"] = description __props__.__dict__["parameters"] = parameters + if position is not None and not isinstance(position, SqlWidgetPositionArgs): + position = position or {} + def _setter(key, value): + position[key] = value + SqlWidgetPositionArgs._configure(_setter, **position) __props__.__dict__["position"] = position __props__.__dict__["text"] = text __props__.__dict__["title"] = title diff --git a/sdk/python/pulumi_databricks/storage_credential.py b/sdk/python/pulumi_databricks/storage_credential.py index 5224bc6c..572d2fa4 100644 --- a/sdk/python/pulumi_databricks/storage_credential.py +++ b/sdk/python/pulumi_databricks/storage_credential.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -39,28 +39,57 @@ def __init__(__self__, *, `azure_service_principal` optional configuration block to use service principal as credential details for Azure (Legacy): """ + StorageCredentialArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + aws_iam_role=aws_iam_role, + azure_managed_identity=azure_managed_identity, + azure_service_principal=azure_service_principal, + comment=comment, + databricks_gcp_service_account=databricks_gcp_service_account, + force_destroy=force_destroy, + gcp_service_account_key=gcp_service_account_key, + metastore_id=metastore_id, + name=name, + owner=owner, + read_only=read_only, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + aws_iam_role: Optional[pulumi.Input['StorageCredentialAwsIamRoleArgs']] = None, + azure_managed_identity: Optional[pulumi.Input['StorageCredentialAzureManagedIdentityArgs']] = None, + azure_service_principal: Optional[pulumi.Input['StorageCredentialAzureServicePrincipalArgs']] = None, + comment: Optional[pulumi.Input[str]] = None, + databricks_gcp_service_account: Optional[pulumi.Input['StorageCredentialDatabricksGcpServiceAccountArgs']] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + gcp_service_account_key: Optional[pulumi.Input['StorageCredentialGcpServiceAccountKeyArgs']] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if aws_iam_role is not None: - pulumi.set(__self__, "aws_iam_role", aws_iam_role) + _setter("aws_iam_role", aws_iam_role) if azure_managed_identity is not None: - pulumi.set(__self__, "azure_managed_identity", azure_managed_identity) + _setter("azure_managed_identity", azure_managed_identity) if azure_service_principal is not None: - pulumi.set(__self__, "azure_service_principal", azure_service_principal) + _setter("azure_service_principal", azure_service_principal) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if databricks_gcp_service_account is not None: - pulumi.set(__self__, "databricks_gcp_service_account", databricks_gcp_service_account) + _setter("databricks_gcp_service_account", databricks_gcp_service_account) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if gcp_service_account_key is not None: - pulumi.set(__self__, "gcp_service_account_key", gcp_service_account_key) + _setter("gcp_service_account_key", gcp_service_account_key) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) @property @pulumi.getter(name="awsIamRole") @@ -207,28 +236,57 @@ def __init__(__self__, *, `azure_service_principal` optional configuration block to use service principal as credential details for Azure (Legacy): """ + _StorageCredentialState._configure( + lambda key, value: pulumi.set(__self__, key, value), + aws_iam_role=aws_iam_role, + azure_managed_identity=azure_managed_identity, + azure_service_principal=azure_service_principal, + comment=comment, + databricks_gcp_service_account=databricks_gcp_service_account, + force_destroy=force_destroy, + gcp_service_account_key=gcp_service_account_key, + metastore_id=metastore_id, + name=name, + owner=owner, + read_only=read_only, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + aws_iam_role: Optional[pulumi.Input['StorageCredentialAwsIamRoleArgs']] = None, + azure_managed_identity: Optional[pulumi.Input['StorageCredentialAzureManagedIdentityArgs']] = None, + azure_service_principal: Optional[pulumi.Input['StorageCredentialAzureServicePrincipalArgs']] = None, + comment: Optional[pulumi.Input[str]] = None, + databricks_gcp_service_account: Optional[pulumi.Input['StorageCredentialDatabricksGcpServiceAccountArgs']] = None, + force_destroy: Optional[pulumi.Input[bool]] = None, + gcp_service_account_key: Optional[pulumi.Input['StorageCredentialGcpServiceAccountKeyArgs']] = None, + metastore_id: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + read_only: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if aws_iam_role is not None: - pulumi.set(__self__, "aws_iam_role", aws_iam_role) + _setter("aws_iam_role", aws_iam_role) if azure_managed_identity is not None: - pulumi.set(__self__, "azure_managed_identity", azure_managed_identity) + _setter("azure_managed_identity", azure_managed_identity) if azure_service_principal is not None: - pulumi.set(__self__, "azure_service_principal", azure_service_principal) + _setter("azure_service_principal", azure_service_principal) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if databricks_gcp_service_account is not None: - pulumi.set(__self__, "databricks_gcp_service_account", databricks_gcp_service_account) + _setter("databricks_gcp_service_account", databricks_gcp_service_account) if force_destroy is not None: - pulumi.set(__self__, "force_destroy", force_destroy) + _setter("force_destroy", force_destroy) if gcp_service_account_key is not None: - pulumi.set(__self__, "gcp_service_account_key", gcp_service_account_key) + _setter("gcp_service_account_key", gcp_service_account_key) if metastore_id is not None: - pulumi.set(__self__, "metastore_id", metastore_id) + _setter("metastore_id", metastore_id) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if read_only is not None: - pulumi.set(__self__, "read_only", read_only) + _setter("read_only", read_only) @property @pulumi.getter(name="awsIamRole") @@ -532,6 +590,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + StorageCredentialArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, @@ -557,12 +619,37 @@ def _internal_init(__self__, raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = StorageCredentialArgs.__new__(StorageCredentialArgs) + if aws_iam_role is not None and not isinstance(aws_iam_role, StorageCredentialAwsIamRoleArgs): + aws_iam_role = aws_iam_role or {} + def _setter(key, value): + aws_iam_role[key] = value + StorageCredentialAwsIamRoleArgs._configure(_setter, **aws_iam_role) __props__.__dict__["aws_iam_role"] = aws_iam_role + if azure_managed_identity is not None and not isinstance(azure_managed_identity, StorageCredentialAzureManagedIdentityArgs): + azure_managed_identity = azure_managed_identity or {} + def _setter(key, value): + azure_managed_identity[key] = value + StorageCredentialAzureManagedIdentityArgs._configure(_setter, **azure_managed_identity) __props__.__dict__["azure_managed_identity"] = azure_managed_identity + if azure_service_principal is not None and not isinstance(azure_service_principal, StorageCredentialAzureServicePrincipalArgs): + azure_service_principal = azure_service_principal or {} + def _setter(key, value): + azure_service_principal[key] = value + StorageCredentialAzureServicePrincipalArgs._configure(_setter, **azure_service_principal) __props__.__dict__["azure_service_principal"] = azure_service_principal __props__.__dict__["comment"] = comment + if databricks_gcp_service_account is not None and not isinstance(databricks_gcp_service_account, StorageCredentialDatabricksGcpServiceAccountArgs): + databricks_gcp_service_account = databricks_gcp_service_account or {} + def _setter(key, value): + databricks_gcp_service_account[key] = value + StorageCredentialDatabricksGcpServiceAccountArgs._configure(_setter, **databricks_gcp_service_account) __props__.__dict__["databricks_gcp_service_account"] = databricks_gcp_service_account __props__.__dict__["force_destroy"] = force_destroy + if gcp_service_account_key is not None and not isinstance(gcp_service_account_key, StorageCredentialGcpServiceAccountKeyArgs): + gcp_service_account_key = gcp_service_account_key or {} + def _setter(key, value): + gcp_service_account_key[key] = value + StorageCredentialGcpServiceAccountKeyArgs._configure(_setter, **gcp_service_account_key) __props__.__dict__["gcp_service_account_key"] = gcp_service_account_key __props__.__dict__["metastore_id"] = metastore_id __props__.__dict__["name"] = name diff --git a/sdk/python/pulumi_databricks/table.py b/sdk/python/pulumi_databricks/table.py index 2425601a..a04b4806 100644 --- a/sdk/python/pulumi_databricks/table.py +++ b/sdk/python/pulumi_databricks/table.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities from . import outputs from ._inputs import * @@ -31,25 +31,56 @@ def __init__(__self__, *, """ The set of arguments for constructing a Table resource. """ - pulumi.set(__self__, "catalog_name", catalog_name) - pulumi.set(__self__, "columns", columns) - pulumi.set(__self__, "data_source_format", data_source_format) - pulumi.set(__self__, "schema_name", schema_name) - pulumi.set(__self__, "table_type", table_type) + TableArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + columns=columns, + data_source_format=data_source_format, + schema_name=schema_name, + table_type=table_type, + comment=comment, + name=name, + owner=owner, + properties=properties, + storage_credential_name=storage_credential_name, + storage_location=storage_location, + view_definition=view_definition, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: pulumi.Input[str], + columns: pulumi.Input[Sequence[pulumi.Input['TableColumnArgs']]], + data_source_format: pulumi.Input[str], + schema_name: pulumi.Input[str], + table_type: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + storage_credential_name: Optional[pulumi.Input[str]] = None, + storage_location: Optional[pulumi.Input[str]] = None, + view_definition: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("catalog_name", catalog_name) + _setter("columns", columns) + _setter("data_source_format", data_source_format) + _setter("schema_name", schema_name) + _setter("table_type", table_type) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if storage_credential_name is not None: - pulumi.set(__self__, "storage_credential_name", storage_credential_name) + _setter("storage_credential_name", storage_credential_name) if storage_location is not None: - pulumi.set(__self__, "storage_location", storage_location) + _setter("storage_location", storage_location) if view_definition is not None: - pulumi.set(__self__, "view_definition", view_definition) + _setter("view_definition", view_definition) @property @pulumi.getter(name="catalogName") @@ -178,30 +209,61 @@ def __init__(__self__, *, """ Input properties used for looking up and filtering Table resources. """ + _TableState._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + columns=columns, + comment=comment, + data_source_format=data_source_format, + name=name, + owner=owner, + properties=properties, + schema_name=schema_name, + storage_credential_name=storage_credential_name, + storage_location=storage_location, + table_type=table_type, + view_definition=view_definition, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: Optional[pulumi.Input[str]] = None, + columns: Optional[pulumi.Input[Sequence[pulumi.Input['TableColumnArgs']]]] = None, + comment: Optional[pulumi.Input[str]] = None, + data_source_format: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + properties: Optional[pulumi.Input[Mapping[str, Any]]] = None, + schema_name: Optional[pulumi.Input[str]] = None, + storage_credential_name: Optional[pulumi.Input[str]] = None, + storage_location: Optional[pulumi.Input[str]] = None, + table_type: Optional[pulumi.Input[str]] = None, + view_definition: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if catalog_name is not None: - pulumi.set(__self__, "catalog_name", catalog_name) + _setter("catalog_name", catalog_name) if columns is not None: - pulumi.set(__self__, "columns", columns) + _setter("columns", columns) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if data_source_format is not None: - pulumi.set(__self__, "data_source_format", data_source_format) + _setter("data_source_format", data_source_format) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if properties is not None: - pulumi.set(__self__, "properties", properties) + _setter("properties", properties) if schema_name is not None: - pulumi.set(__self__, "schema_name", schema_name) + _setter("schema_name", schema_name) if storage_credential_name is not None: - pulumi.set(__self__, "storage_credential_name", storage_credential_name) + _setter("storage_credential_name", storage_credential_name) if storage_location is not None: - pulumi.set(__self__, "storage_location", storage_location) + _setter("storage_location", storage_location) if table_type is not None: - pulumi.set(__self__, "table_type", table_type) + _setter("table_type", table_type) if view_definition is not None: - pulumi.set(__self__, "view_definition", view_definition) + _setter("view_definition", view_definition) @property @pulumi.getter(name="catalogName") @@ -353,6 +415,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + TableArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/token.py b/sdk/python/pulumi_databricks/token.py index 9b796092..e7b8962b 100644 --- a/sdk/python/pulumi_databricks/token.py +++ b/sdk/python/pulumi_databricks/token.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['TokenArgs', 'Token'] @@ -24,16 +24,33 @@ def __init__(__self__, *, :param pulumi.Input[str] comment: (String) Comment that will appear on the user’s settings page for this token. :param pulumi.Input[int] lifetime_seconds: (Integer) The lifetime of the token, in seconds. If no lifetime is specified, the token remains valid indefinitely. """ + TokenArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + comment=comment, + creation_time=creation_time, + expiry_time=expiry_time, + lifetime_seconds=lifetime_seconds, + token_id=token_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + comment: Optional[pulumi.Input[str]] = None, + creation_time: Optional[pulumi.Input[int]] = None, + expiry_time: Optional[pulumi.Input[int]] = None, + lifetime_seconds: Optional[pulumi.Input[int]] = None, + token_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if expiry_time is not None: - pulumi.set(__self__, "expiry_time", expiry_time) + _setter("expiry_time", expiry_time) if lifetime_seconds is not None: - pulumi.set(__self__, "lifetime_seconds", lifetime_seconds) + _setter("lifetime_seconds", lifetime_seconds) if token_id is not None: - pulumi.set(__self__, "token_id", token_id) + _setter("token_id", token_id) @property @pulumi.getter @@ -102,18 +119,37 @@ def __init__(__self__, *, :param pulumi.Input[int] lifetime_seconds: (Integer) The lifetime of the token, in seconds. If no lifetime is specified, the token remains valid indefinitely. :param pulumi.Input[str] token_value: **Sensitive** value of the newly-created token. """ + _TokenState._configure( + lambda key, value: pulumi.set(__self__, key, value), + comment=comment, + creation_time=creation_time, + expiry_time=expiry_time, + lifetime_seconds=lifetime_seconds, + token_id=token_id, + token_value=token_value, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + comment: Optional[pulumi.Input[str]] = None, + creation_time: Optional[pulumi.Input[int]] = None, + expiry_time: Optional[pulumi.Input[int]] = None, + lifetime_seconds: Optional[pulumi.Input[int]] = None, + token_id: Optional[pulumi.Input[str]] = None, + token_value: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if creation_time is not None: - pulumi.set(__self__, "creation_time", creation_time) + _setter("creation_time", creation_time) if expiry_time is not None: - pulumi.set(__self__, "expiry_time", expiry_time) + _setter("expiry_time", expiry_time) if lifetime_seconds is not None: - pulumi.set(__self__, "lifetime_seconds", lifetime_seconds) + _setter("lifetime_seconds", lifetime_seconds) if token_id is not None: - pulumi.set(__self__, "token_id", token_id) + _setter("token_id", token_id) if token_value is not None: - pulumi.set(__self__, "token_value", token_value) + _setter("token_value", token_value) @property @pulumi.getter @@ -283,6 +319,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + TokenArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/user.py b/sdk/python/pulumi_databricks/user.py index 01dfdfdf..ecfbc87d 100644 --- a/sdk/python/pulumi_databricks/user.py +++ b/sdk/python/pulumi_databricks/user.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['UserArgs', 'User'] @@ -45,35 +45,72 @@ def __init__(__self__, *, :param pulumi.Input[str] home: Home folder of the user, e.g. `/Users/mr.foo@example.com`. :param pulumi.Input[str] repos: Personal Repos location of the user, e.g. `/Repos/mr.foo@example.com`. """ - pulumi.set(__self__, "user_name", user_name) + UserArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + user_name=user_name, + acl_principal_id=acl_principal_id, + active=active, + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + databricks_sql_access=databricks_sql_access, + disable_as_user_deletion=disable_as_user_deletion, + display_name=display_name, + external_id=external_id, + force=force, + force_delete_home_dir=force_delete_home_dir, + force_delete_repos=force_delete_repos, + home=home, + repos=repos, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + user_name: pulumi.Input[str], + acl_principal_id: Optional[pulumi.Input[str]] = None, + active: Optional[pulumi.Input[bool]] = None, + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + disable_as_user_deletion: Optional[pulumi.Input[bool]] = None, + display_name: Optional[pulumi.Input[str]] = None, + external_id: Optional[pulumi.Input[str]] = None, + force: Optional[pulumi.Input[bool]] = None, + force_delete_home_dir: Optional[pulumi.Input[bool]] = None, + force_delete_repos: Optional[pulumi.Input[bool]] = None, + home: Optional[pulumi.Input[str]] = None, + repos: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("user_name", user_name) if acl_principal_id is not None: - pulumi.set(__self__, "acl_principal_id", acl_principal_id) + _setter("acl_principal_id", acl_principal_id) if active is not None: - pulumi.set(__self__, "active", active) + _setter("active", active) if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if disable_as_user_deletion is not None: - pulumi.set(__self__, "disable_as_user_deletion", disable_as_user_deletion) + _setter("disable_as_user_deletion", disable_as_user_deletion) if display_name is not None: - pulumi.set(__self__, "display_name", display_name) + _setter("display_name", display_name) if external_id is not None: - pulumi.set(__self__, "external_id", external_id) + _setter("external_id", external_id) if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if force_delete_home_dir is not None: - pulumi.set(__self__, "force_delete_home_dir", force_delete_home_dir) + _setter("force_delete_home_dir", force_delete_home_dir) if force_delete_repos is not None: - pulumi.set(__self__, "force_delete_repos", force_delete_repos) + _setter("force_delete_repos", force_delete_repos) if home is not None: - pulumi.set(__self__, "home", home) + _setter("home", home) if repos is not None: - pulumi.set(__self__, "repos", repos) + _setter("repos", repos) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="userName") @@ -284,36 +321,73 @@ def __init__(__self__, *, :param pulumi.Input[str] repos: Personal Repos location of the user, e.g. `/Repos/mr.foo@example.com`. :param pulumi.Input[str] user_name: This is the username of the given user and will be their form of access and identity. """ + _UserState._configure( + lambda key, value: pulumi.set(__self__, key, value), + acl_principal_id=acl_principal_id, + active=active, + allow_cluster_create=allow_cluster_create, + allow_instance_pool_create=allow_instance_pool_create, + databricks_sql_access=databricks_sql_access, + disable_as_user_deletion=disable_as_user_deletion, + display_name=display_name, + external_id=external_id, + force=force, + force_delete_home_dir=force_delete_home_dir, + force_delete_repos=force_delete_repos, + home=home, + repos=repos, + user_name=user_name, + workspace_access=workspace_access, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + acl_principal_id: Optional[pulumi.Input[str]] = None, + active: Optional[pulumi.Input[bool]] = None, + allow_cluster_create: Optional[pulumi.Input[bool]] = None, + allow_instance_pool_create: Optional[pulumi.Input[bool]] = None, + databricks_sql_access: Optional[pulumi.Input[bool]] = None, + disable_as_user_deletion: Optional[pulumi.Input[bool]] = None, + display_name: Optional[pulumi.Input[str]] = None, + external_id: Optional[pulumi.Input[str]] = None, + force: Optional[pulumi.Input[bool]] = None, + force_delete_home_dir: Optional[pulumi.Input[bool]] = None, + force_delete_repos: Optional[pulumi.Input[bool]] = None, + home: Optional[pulumi.Input[str]] = None, + repos: Optional[pulumi.Input[str]] = None, + user_name: Optional[pulumi.Input[str]] = None, + workspace_access: Optional[pulumi.Input[bool]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if acl_principal_id is not None: - pulumi.set(__self__, "acl_principal_id", acl_principal_id) + _setter("acl_principal_id", acl_principal_id) if active is not None: - pulumi.set(__self__, "active", active) + _setter("active", active) if allow_cluster_create is not None: - pulumi.set(__self__, "allow_cluster_create", allow_cluster_create) + _setter("allow_cluster_create", allow_cluster_create) if allow_instance_pool_create is not None: - pulumi.set(__self__, "allow_instance_pool_create", allow_instance_pool_create) + _setter("allow_instance_pool_create", allow_instance_pool_create) if databricks_sql_access is not None: - pulumi.set(__self__, "databricks_sql_access", databricks_sql_access) + _setter("databricks_sql_access", databricks_sql_access) if disable_as_user_deletion is not None: - pulumi.set(__self__, "disable_as_user_deletion", disable_as_user_deletion) + _setter("disable_as_user_deletion", disable_as_user_deletion) if display_name is not None: - pulumi.set(__self__, "display_name", display_name) + _setter("display_name", display_name) if external_id is not None: - pulumi.set(__self__, "external_id", external_id) + _setter("external_id", external_id) if force is not None: - pulumi.set(__self__, "force", force) + _setter("force", force) if force_delete_home_dir is not None: - pulumi.set(__self__, "force_delete_home_dir", force_delete_home_dir) + _setter("force_delete_home_dir", force_delete_home_dir) if force_delete_repos is not None: - pulumi.set(__self__, "force_delete_repos", force_delete_repos) + _setter("force_delete_repos", force_delete_repos) if home is not None: - pulumi.set(__self__, "home", home) + _setter("home", home) if repos is not None: - pulumi.set(__self__, "repos", repos) + _setter("repos", repos) if user_name is not None: - pulumi.set(__self__, "user_name", user_name) + _setter("user_name", user_name) if workspace_access is not None: - pulumi.set(__self__, "workspace_access", workspace_access) + _setter("workspace_access", workspace_access) @property @pulumi.getter(name="aclPrincipalId") @@ -737,6 +811,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + UserArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/user_instance_profile.py b/sdk/python/pulumi_databricks/user_instance_profile.py index ef4ae7ed..cef093d3 100644 --- a/sdk/python/pulumi_databricks/user_instance_profile.py +++ b/sdk/python/pulumi_databricks/user_instance_profile.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['UserInstanceProfileArgs', 'UserInstanceProfile'] @@ -21,8 +21,19 @@ def __init__(__self__, *, :param pulumi.Input[str] instance_profile_id: This is the id of the instance profile resource. :param pulumi.Input[str] user_id: This is the id of the user resource. """ - pulumi.set(__self__, "instance_profile_id", instance_profile_id) - pulumi.set(__self__, "user_id", user_id) + UserInstanceProfileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + instance_profile_id=instance_profile_id, + user_id=user_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + instance_profile_id: pulumi.Input[str], + user_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("instance_profile_id", instance_profile_id) + _setter("user_id", user_id) @property @pulumi.getter(name="instanceProfileId") @@ -59,10 +70,21 @@ def __init__(__self__, *, :param pulumi.Input[str] instance_profile_id: This is the id of the instance profile resource. :param pulumi.Input[str] user_id: This is the id of the user resource. """ + _UserInstanceProfileState._configure( + lambda key, value: pulumi.set(__self__, key, value), + instance_profile_id=instance_profile_id, + user_id=user_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + instance_profile_id: Optional[pulumi.Input[str]] = None, + user_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if instance_profile_id is not None: - pulumi.set(__self__, "instance_profile_id", instance_profile_id) + _setter("instance_profile_id", instance_profile_id) if user_id is not None: - pulumi.set(__self__, "user_id", user_id) + _setter("user_id", user_id) @property @pulumi.getter(name="instanceProfileId") @@ -182,6 +204,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + UserInstanceProfileArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/user_role.py b/sdk/python/pulumi_databricks/user_role.py index 60d44315..f9e6624f 100644 --- a/sdk/python/pulumi_databricks/user_role.py +++ b/sdk/python/pulumi_databricks/user_role.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['UserRoleArgs', 'UserRole'] @@ -21,8 +21,19 @@ def __init__(__self__, *, :param pulumi.Input[str] role: Either a role name or the ARN/ID of the instance profile resource. :param pulumi.Input[str] user_id: This is the id of the user resource. """ - pulumi.set(__self__, "role", role) - pulumi.set(__self__, "user_id", user_id) + UserRoleArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + role=role, + user_id=user_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role: pulumi.Input[str], + user_id: pulumi.Input[str], + opts: Optional[pulumi.ResourceOptions]=None): + _setter("role", role) + _setter("user_id", user_id) @property @pulumi.getter @@ -59,10 +70,21 @@ def __init__(__self__, *, :param pulumi.Input[str] role: Either a role name or the ARN/ID of the instance profile resource. :param pulumi.Input[str] user_id: This is the id of the user resource. """ + _UserRoleState._configure( + lambda key, value: pulumi.set(__self__, key, value), + role=role, + user_id=user_id, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + role: Optional[pulumi.Input[str]] = None, + user_id: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if role is not None: - pulumi.set(__self__, "role", role) + _setter("role", role) if user_id is not None: - pulumi.set(__self__, "user_id", user_id) + _setter("user_id", user_id) @property @pulumi.getter @@ -206,6 +228,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + UserRoleArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/volume.py b/sdk/python/pulumi_databricks/volume.py index 114c39e0..ee5d5106 100644 --- a/sdk/python/pulumi_databricks/volume.py +++ b/sdk/python/pulumi_databricks/volume.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['VolumeArgs', 'Volume'] @@ -31,17 +31,38 @@ def __init__(__self__, *, :param pulumi.Input[str] owner: Name of the volume owner. :param pulumi.Input[str] storage_location: Path inside an External Location. Only used for `EXTERNAL` Volumes. """ - pulumi.set(__self__, "catalog_name", catalog_name) - pulumi.set(__self__, "schema_name", schema_name) - pulumi.set(__self__, "volume_type", volume_type) + VolumeArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + schema_name=schema_name, + volume_type=volume_type, + comment=comment, + name=name, + owner=owner, + storage_location=storage_location, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: pulumi.Input[str], + schema_name: pulumi.Input[str], + volume_type: pulumi.Input[str], + comment: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + storage_location: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("catalog_name", catalog_name) + _setter("schema_name", schema_name) + _setter("volume_type", volume_type) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if storage_location is not None: - pulumi.set(__self__, "storage_location", storage_location) + _setter("storage_location", storage_location) @property @pulumi.getter(name="catalogName") @@ -148,20 +169,41 @@ def __init__(__self__, *, :param pulumi.Input[str] storage_location: Path inside an External Location. Only used for `EXTERNAL` Volumes. :param pulumi.Input[str] volume_type: Volume type. `EXTERNAL` or `MANAGED`. """ + _VolumeState._configure( + lambda key, value: pulumi.set(__self__, key, value), + catalog_name=catalog_name, + comment=comment, + name=name, + owner=owner, + schema_name=schema_name, + storage_location=storage_location, + volume_type=volume_type, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + catalog_name: Optional[pulumi.Input[str]] = None, + comment: Optional[pulumi.Input[str]] = None, + name: Optional[pulumi.Input[str]] = None, + owner: Optional[pulumi.Input[str]] = None, + schema_name: Optional[pulumi.Input[str]] = None, + storage_location: Optional[pulumi.Input[str]] = None, + volume_type: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if catalog_name is not None: - pulumi.set(__self__, "catalog_name", catalog_name) + _setter("catalog_name", catalog_name) if comment is not None: - pulumi.set(__self__, "comment", comment) + _setter("comment", comment) if name is not None: - pulumi.set(__self__, "name", name) + _setter("name", name) if owner is not None: - pulumi.set(__self__, "owner", owner) + _setter("owner", owner) if schema_name is not None: - pulumi.set(__self__, "schema_name", schema_name) + _setter("schema_name", schema_name) if storage_location is not None: - pulumi.set(__self__, "storage_location", storage_location) + _setter("storage_location", storage_location) if volume_type is not None: - pulumi.set(__self__, "volume_type", volume_type) + _setter("volume_type", volume_type) @property @pulumi.getter(name="catalogName") @@ -413,6 +455,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + VolumeArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/workspace_conf.py b/sdk/python/pulumi_databricks/workspace_conf.py index f34cbfba..c471eb26 100644 --- a/sdk/python/pulumi_databricks/workspace_conf.py +++ b/sdk/python/pulumi_databricks/workspace_conf.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['WorkspaceConfArgs', 'WorkspaceConf'] @@ -19,8 +19,17 @@ def __init__(__self__, *, The set of arguments for constructing a WorkspaceConf resource. :param pulumi.Input[Mapping[str, Any]] custom_config: Key-value map of strings that represent workspace configuration. Upon resource deletion, properties that start with `enable` or `enforce` will be reset to `false` value, regardless of initial default one. """ + WorkspaceConfArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + custom_config=custom_config, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + custom_config: Optional[pulumi.Input[Mapping[str, Any]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if custom_config is not None: - pulumi.set(__self__, "custom_config", custom_config) + _setter("custom_config", custom_config) @property @pulumi.getter(name="customConfig") @@ -43,8 +52,17 @@ def __init__(__self__, *, Input properties used for looking up and filtering WorkspaceConf resources. :param pulumi.Input[Mapping[str, Any]] custom_config: Key-value map of strings that represent workspace configuration. Upon resource deletion, properties that start with `enable` or `enforce` will be reset to `false` value, regardless of initial default one. """ + _WorkspaceConfState._configure( + lambda key, value: pulumi.set(__self__, key, value), + custom_config=custom_config, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + custom_config: Optional[pulumi.Input[Mapping[str, Any]]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if custom_config is not None: - pulumi.set(__self__, "custom_config", custom_config) + _setter("custom_config", custom_config) @property @pulumi.getter(name="customConfig") @@ -134,6 +152,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + WorkspaceConfArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, diff --git a/sdk/python/pulumi_databricks/workspace_file.py b/sdk/python/pulumi_databricks/workspace_file.py index d85216b5..d0da54b6 100644 --- a/sdk/python/pulumi_databricks/workspace_file.py +++ b/sdk/python/pulumi_databricks/workspace_file.py @@ -6,7 +6,7 @@ import warnings import pulumi import pulumi.runtime -from typing import Any, Mapping, Optional, Sequence, Union, overload +from typing import Any, Callable, Mapping, Optional, Sequence, Union, overload from . import _utilities __all__ = ['WorkspaceFileArgs', 'WorkspaceFile'] @@ -25,15 +25,32 @@ def __init__(__self__, *, :param pulumi.Input[int] object_id: Unique identifier for a workspace file :param pulumi.Input[str] source: Path to file on local filesystem. Conflicts with `content_base64`. """ - pulumi.set(__self__, "path", path) + WorkspaceFileArgs._configure( + lambda key, value: pulumi.set(__self__, key, value), + path=path, + content_base64=content_base64, + md5=md5, + object_id=object_id, + source=source, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + path: pulumi.Input[str], + content_base64: Optional[pulumi.Input[str]] = None, + md5: Optional[pulumi.Input[str]] = None, + object_id: Optional[pulumi.Input[int]] = None, + source: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): + _setter("path", path) if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if object_id is not None: - pulumi.set(__self__, "object_id", object_id) + _setter("object_id", object_id) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) @property @pulumi.getter @@ -106,18 +123,37 @@ def __init__(__self__, *, :param pulumi.Input[str] source: Path to file on local filesystem. Conflicts with `content_base64`. :param pulumi.Input[str] url: Routable URL of the workspace file """ + _WorkspaceFileState._configure( + lambda key, value: pulumi.set(__self__, key, value), + content_base64=content_base64, + md5=md5, + object_id=object_id, + path=path, + source=source, + url=url, + ) + @staticmethod + def _configure( + _setter: Callable[[Any, Any], None], + content_base64: Optional[pulumi.Input[str]] = None, + md5: Optional[pulumi.Input[str]] = None, + object_id: Optional[pulumi.Input[int]] = None, + path: Optional[pulumi.Input[str]] = None, + source: Optional[pulumi.Input[str]] = None, + url: Optional[pulumi.Input[str]] = None, + opts: Optional[pulumi.ResourceOptions]=None): if content_base64 is not None: - pulumi.set(__self__, "content_base64", content_base64) + _setter("content_base64", content_base64) if md5 is not None: - pulumi.set(__self__, "md5", md5) + _setter("md5", md5) if object_id is not None: - pulumi.set(__self__, "object_id", object_id) + _setter("object_id", object_id) if path is not None: - pulumi.set(__self__, "path", path) + _setter("path", path) if source is not None: - pulumi.set(__self__, "source", source) + _setter("source", source) if url is not None: - pulumi.set(__self__, "url", url) + _setter("url", url) @property @pulumi.getter(name="contentBase64") @@ -237,6 +273,10 @@ def __init__(__self__, resource_name: str, *args, **kwargs): if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: + kwargs = kwargs or {} + def _setter(key, value): + kwargs[key] = value + WorkspaceFileArgs._configure(_setter, **kwargs) __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__,