From 213a5232eeccdebfd6b780dba7314b00fccf644e Mon Sep 17 00:00:00 2001 From: Ian Wahbe Date: Mon, 27 Nov 2023 12:39:24 -0800 Subject: [PATCH] Upgrade terraform-provider-mongodbatlas to v1.13.1 (#335) This PR was generated via `$ upgrade-provider pulumi/pulumi-mongodbatlas`. --- - Upgrading terraform-provider-mongodbatlas from 1.13.0 to 1.13.1. Fixes #333 --- .../bridge-metadata.json | 21 + .../pulumi-resource-mongodbatlas/schema.json | 96 ++++- provider/go.mod | 32 +- provider/go.sum | 58 +-- sdk/dotnet/Cluster.cs | 4 +- sdk/dotnet/GetOnlineArchive.cs | 14 + sdk/dotnet/GetOnlineArchives.cs | 10 + .../OnlineArchiveDataProcessRegionArgs.cs | 32 ++ .../OnlineArchiveDataProcessRegionGetArgs.cs | 32 ++ sdk/dotnet/OnlineArchive.cs | 18 + ...GetOnlineArchiveDataProcessRegionResult.cs | 29 ++ ...neArchivesResultDataProcessRegionResult.cs | 29 ++ .../Outputs/GetOnlineArchivesResultResult.cs | 4 + .../Outputs/OnlineArchiveDataProcessRegion.cs | 35 ++ sdk/dotnet/SearchDeployment.cs | 3 +- sdk/go/mongodbatlas/cluster.go | 14 +- sdk/go/mongodbatlas/getOnlineArchive.go | 10 + sdk/go/mongodbatlas/getOnlineArchives.go | 5 + sdk/go/mongodbatlas/onlineArchive.go | 15 + sdk/go/mongodbatlas/pulumiTypes.go | 376 ++++++++++++++++++ sdk/go/mongodbatlas/searchDeployment.go | 3 +- .../java/com/pulumi/mongodbatlas/Cluster.java | 12 +- .../mongodbatlas/MongodbatlasFunctions.java | 40 ++ .../pulumi/mongodbatlas/OnlineArchive.java | 15 + .../mongodbatlas/OnlineArchiveArgs.java | 38 ++ .../pulumi/mongodbatlas/SearchDeployment.java | 3 +- .../OnlineArchiveDataProcessRegionArgs.java | 120 ++++++ .../inputs/OnlineArchiveState.java | 38 ++ .../GetOnlineArchiveDataProcessRegion.java | 58 +++ .../outputs/GetOnlineArchiveResult.java | 16 + .../outputs/GetOnlineArchivesResult.java | 16 + ...OnlineArchivesResultDataProcessRegion.java | 58 +++ .../OnlineArchiveDataProcessRegion.java | 76 ++++ sdk/nodejs/cluster.ts | 4 +- sdk/nodejs/getOnlineArchive.ts | 11 + sdk/nodejs/getOnlineArchives.ts | 10 + sdk/nodejs/onlineArchive.ts | 14 + sdk/nodejs/searchDeployment.ts | 3 +- sdk/nodejs/types/input.ts | 11 + sdk/nodejs/types/output.ts | 22 + sdk/python/pulumi_mongodbatlas/_inputs.py | 40 ++ sdk/python/pulumi_mongodbatlas/cluster.py | 4 +- .../pulumi_mongodbatlas/get_online_archive.py | 22 +- .../get_online_archives.py | 10 + .../pulumi_mongodbatlas/online_archive.py | 47 +++ sdk/python/pulumi_mongodbatlas/outputs.py | 96 +++++ .../pulumi_mongodbatlas/search_deployment.py | 6 +- 47 files changed, 1559 insertions(+), 71 deletions(-) create mode 100644 sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionArgs.cs create mode 100644 sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionGetArgs.cs create mode 100644 sdk/dotnet/Outputs/GetOnlineArchiveDataProcessRegionResult.cs create mode 100644 sdk/dotnet/Outputs/GetOnlineArchivesResultDataProcessRegionResult.cs create mode 100644 sdk/dotnet/Outputs/OnlineArchiveDataProcessRegion.cs create mode 100644 sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveDataProcessRegionArgs.java create mode 100644 sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveDataProcessRegion.java create mode 100644 sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResultDataProcessRegion.java create mode 100644 sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/OnlineArchiveDataProcessRegion.java diff --git a/provider/cmd/pulumi-resource-mongodbatlas/bridge-metadata.json b/provider/cmd/pulumi-resource-mongodbatlas/bridge-metadata.json index c1736476..bc704fea 100644 --- a/provider/cmd/pulumi-resource-mongodbatlas/bridge-metadata.json +++ b/provider/cmd/pulumi-resource-mongodbatlas/bridge-metadata.json @@ -641,6 +641,9 @@ "data_expiration_rule": { "maxItemsOne": true }, + "data_process_region": { + "maxItemsOne": true + }, "partition_fields": { "maxItemsOne": false }, @@ -2231,6 +2234,9 @@ "data_expiration_rule": { "maxItemsOne": false }, + "data_process_region": { + "maxItemsOne": false + }, "partition_fields": { "maxItemsOne": false }, @@ -2253,6 +2259,9 @@ "data_expiration_rule": { "maxItemsOne": false }, + "data_process_region": { + "maxItemsOne": false + }, "partition_fields": { "maxItemsOne": false }, @@ -3339,6 +3348,9 @@ "mongodbatlas:index/OnlineArchiveDataExpirationRule:OnlineArchiveDataExpirationRule": { "expireAfterDays": "expire_after_days" }, + "mongodbatlas:index/OnlineArchiveDataProcessRegion:OnlineArchiveDataProcessRegion": { + "cloudProvider": "cloud_provider" + }, "mongodbatlas:index/OnlineArchivePartitionField:OnlineArchivePartitionField": { "fieldName": "field_name", "fieldType": "field_type" @@ -5168,6 +5180,7 @@ "collectionType": "collection_type", "criterias": "criteria", "dataExpirationRules": "data_expiration_rule", + "dataProcessRegions": "data_process_region", "dbName": "db_name", "partitionFields": "partition_fields", "projectId": "project_id", @@ -5181,6 +5194,9 @@ "mongodbatlas:index/getOnlineArchiveDataExpirationRule:getOnlineArchiveDataExpirationRule": { "expireAfterDays": "expire_after_days" }, + "mongodbatlas:index/getOnlineArchiveDataProcessRegion:getOnlineArchiveDataProcessRegion": { + "cloudProvider": "cloud_provider" + }, "mongodbatlas:index/getOnlineArchivePartitionField:getOnlineArchivePartitionField": { "fieldName": "field_name", "fieldType": "field_type" @@ -5205,6 +5221,7 @@ "collectionType": "collection_type", "criterias": "criteria", "dataExpirationRules": "data_expiration_rule", + "dataProcessRegions": "data_process_region", "dbName": "db_name", "partitionFields": "partition_fields", "projectId": "project_id", @@ -5218,6 +5235,9 @@ "mongodbatlas:index/getOnlineArchivesResultDataExpirationRule:getOnlineArchivesResultDataExpirationRule": { "expireAfterDays": "expire_after_days" }, + "mongodbatlas:index/getOnlineArchivesResultDataProcessRegion:getOnlineArchivesResultDataProcessRegion": { + "cloudProvider": "cloud_provider" + }, "mongodbatlas:index/getOnlineArchivesResultPartitionField:getOnlineArchivesResultPartitionField": { "fieldName": "field_name", "fieldType": "field_type" @@ -5657,6 +5677,7 @@ "collName": "coll_name", "collectionType": "collection_type", "dataExpirationRule": "data_expiration_rule", + "dataProcessRegion": "data_process_region", "dbName": "db_name", "partitionFields": "partition_fields", "projectId": "project_id", diff --git a/provider/cmd/pulumi-resource-mongodbatlas/schema.json b/provider/cmd/pulumi-resource-mongodbatlas/schema.json index d5b8a749..4f9be0f5 100644 --- a/provider/cmd/pulumi-resource-mongodbatlas/schema.json +++ b/provider/cmd/pulumi-resource-mongodbatlas/schema.json @@ -2799,6 +2799,27 @@ "expireAfterDays" ] }, + "mongodbatlas:index/OnlineArchiveDataProcessRegion:OnlineArchiveDataProcessRegion": { + "properties": { + "cloudProvider": { + "type": "string", + "description": "Human-readable label that identifies the Cloud service provider where you wish to store your archived data.\n" + }, + "region": { + "type": "string", + "description": "Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive)\n" + } + }, + "type": "object", + "language": { + "nodejs": { + "requiredOutputs": [ + "cloudProvider", + "region" + ] + } + } + }, "mongodbatlas:index/OnlineArchivePartitionField:OnlineArchivePartitionField": { "properties": { "fieldName": { @@ -10304,6 +10325,26 @@ } } }, + "mongodbatlas:index/getOnlineArchiveDataProcessRegion:getOnlineArchiveDataProcessRegion": { + "properties": { + "cloudProvider": { + "type": "string" + }, + "region": { + "type": "string" + } + }, + "type": "object", + "required": [ + "cloudProvider", + "region" + ], + "language": { + "nodejs": { + "requiredInputs": [] + } + } + }, "mongodbatlas:index/getOnlineArchivePartitionField:getOnlineArchivePartitionField": { "properties": { "fieldName": { @@ -10393,6 +10434,12 @@ "$ref": "#/types/mongodbatlas:index/getOnlineArchivesResultDataExpirationRule:getOnlineArchivesResultDataExpirationRule" } }, + "dataProcessRegions": { + "type": "array", + "items": { + "$ref": "#/types/mongodbatlas:index/getOnlineArchivesResultDataProcessRegion:getOnlineArchivesResultDataProcessRegion" + } + }, "dbName": { "type": "string" }, @@ -10427,6 +10474,7 @@ "collectionType", "criterias", "dataExpirationRules", + "dataProcessRegions", "dbName", "partitionFields", "paused", @@ -10488,6 +10536,26 @@ } } }, + "mongodbatlas:index/getOnlineArchivesResultDataProcessRegion:getOnlineArchivesResultDataProcessRegion": { + "properties": { + "cloudProvider": { + "type": "string" + }, + "region": { + "type": "string" + } + }, + "type": "object", + "required": [ + "cloudProvider", + "region" + ], + "language": { + "nodejs": { + "requiredInputs": [] + } + } + }, "mongodbatlas:index/getOnlineArchivesResultPartitionField:getOnlineArchivesResultPartitionField": { "properties": { "fieldName": { @@ -13971,9 +14039,11 @@ } }, "required": [ + "acceptDataRisksAndForceReplicaSetReconfig", "advancedConfiguration", "autoScalingComputeEnabled", "autoScalingComputeScaleDownEnabled", + "autoScalingDiskGbEnabled", "backingProviderName", "backupEnabled", "biConnectorConfig", @@ -16919,6 +16989,10 @@ "$ref": "#/types/mongodbatlas:index/OnlineArchiveDataExpirationRule:OnlineArchiveDataExpirationRule", "description": "Rule for specifying when data should be deleted from the archive. See data expiration rule.\n" }, + "dataProcessRegion": { + "$ref": "#/types/mongodbatlas:index/OnlineArchiveDataProcessRegion:OnlineArchiveDataProcessRegion", + "description": "Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated.\n" + }, "dbName": { "type": "string", "description": "Name of the database that contains the collection.\n" @@ -16956,6 +17030,7 @@ "collName", "collectionType", "criteria", + "dataProcessRegion", "dbName", "partitionFields", "paused", @@ -16984,6 +17059,10 @@ "$ref": "#/types/mongodbatlas:index/OnlineArchiveDataExpirationRule:OnlineArchiveDataExpirationRule", "description": "Rule for specifying when data should be deleted from the archive. See data expiration rule.\n" }, + "dataProcessRegion": { + "$ref": "#/types/mongodbatlas:index/OnlineArchiveDataProcessRegion:OnlineArchiveDataProcessRegion", + "description": "Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated.\n" + }, "dbName": { "type": "string", "description": "Name of the database that contains the collection.\n" @@ -17047,6 +17126,10 @@ "$ref": "#/types/mongodbatlas:index/OnlineArchiveDataExpirationRule:OnlineArchiveDataExpirationRule", "description": "Rule for specifying when data should be deleted from the archive. See data expiration rule.\n" }, + "dataProcessRegion": { + "$ref": "#/types/mongodbatlas:index/OnlineArchiveDataProcessRegion:OnlineArchiveDataProcessRegion", + "description": "Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated.\n" + }, "dbName": { "type": "string", "description": "Name of the database that contains the collection.\n" @@ -18604,7 +18687,7 @@ } }, "mongodbatlas:index/searchDeployment:SearchDeployment": { - "description": "`mongodbatlas.SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster.\n\n\u003e **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation).\n\u003e **NOTE:** Only a single search deployment resource can be defined for each cluster.\n\n\n{{% examples %}}\n## Example Usage\n{{% example %}}\n\n```java\npackage generated_program;\n\nimport com.pulumi.Context;\nimport com.pulumi.Pulumi;\nimport com.pulumi.core.Output;\nimport com.pulumi.mongodbatlas.SearchDeployment;\nimport com.pulumi.mongodbatlas.SearchDeploymentArgs;\nimport com.pulumi.mongodbatlas.inputs.SearchDeploymentSpecArgs;\nimport java.util.List;\nimport java.util.ArrayList;\nimport java.util.Map;\nimport java.io.File;\nimport java.nio.file.Files;\nimport java.nio.file.Paths;\n\npublic class App {\n public static void main(String[] args) {\n Pulumi.run(App::stack);\n }\n\n public static void stack(Context ctx) {\n var test = new SearchDeployment(\"test\", SearchDeploymentArgs.builder() \n .clusterName(\"NAME OF CLUSTER\")\n .projectId(\"PROJECT ID\")\n .specs(SearchDeploymentSpecArgs.builder()\n .instance_size(\"S20_HIGHCPU_NVME\")\n .node_count(2)\n .build())\n .build());\n\n }\n}\n```\n```yaml\nresources:\n test:\n type: mongodbatlas:SearchDeployment\n properties:\n clusterName: NAME OF CLUSTER\n projectId: PROJECT ID\n specs:\n - instance_size: S20_HIGHCPU_NVME\n node_count: 2\n```\n{{% /example %}}\n{{% /examples %}}\n\n## Import\n\nSearch node resource can be imported using the project ID and cluster name, in the format `PROJECT_ID-CLUSTER_NAME`, e.g.\n\n```sh\n $ pulumi import mongodbatlas:index/searchDeployment:SearchDeployment test 650972848269185c55f40ca1-Cluster0\n```\n For more information see[MongoDB Atlas API - Search Node](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/#tag/Atlas-Search/operation/createAtlasSearchDeployment) Documentation.\n\n", + "description": "`mongodbatlas.SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster.\n\n\u003e **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation).\n\n\u003e **NOTE:** Only a single search deployment resource can be defined for each cluster.\n\n\n{{% examples %}}\n## Example Usage\n{{% example %}}\n\n```java\npackage generated_program;\n\nimport com.pulumi.Context;\nimport com.pulumi.Pulumi;\nimport com.pulumi.core.Output;\nimport com.pulumi.mongodbatlas.SearchDeployment;\nimport com.pulumi.mongodbatlas.SearchDeploymentArgs;\nimport com.pulumi.mongodbatlas.inputs.SearchDeploymentSpecArgs;\nimport java.util.List;\nimport java.util.ArrayList;\nimport java.util.Map;\nimport java.io.File;\nimport java.nio.file.Files;\nimport java.nio.file.Paths;\n\npublic class App {\n public static void main(String[] args) {\n Pulumi.run(App::stack);\n }\n\n public static void stack(Context ctx) {\n var test = new SearchDeployment(\"test\", SearchDeploymentArgs.builder() \n .clusterName(\"NAME OF CLUSTER\")\n .projectId(\"PROJECT ID\")\n .specs(SearchDeploymentSpecArgs.builder()\n .instance_size(\"S20_HIGHCPU_NVME\")\n .node_count(2)\n .build())\n .build());\n\n }\n}\n```\n```yaml\nresources:\n test:\n type: mongodbatlas:SearchDeployment\n properties:\n clusterName: NAME OF CLUSTER\n projectId: PROJECT ID\n specs:\n - instance_size: S20_HIGHCPU_NVME\n node_count: 2\n```\n{{% /example %}}\n{{% /examples %}}\n\n## Import\n\nSearch node resource can be imported using the project ID and cluster name, in the format `PROJECT_ID-CLUSTER_NAME`, e.g.\n\n```sh\n $ pulumi import mongodbatlas:index/searchDeployment:SearchDeployment test 650972848269185c55f40ca1-Cluster0\n```\n For more information see[MongoDB Atlas API - Search Node](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/#tag/Atlas-Search/operation/createAtlasSearchDeployment) Documentation.\n\n", "properties": { "clusterName": { "type": "string", @@ -24045,7 +24128,7 @@ } }, "mongodbatlas:index/getOnlineArchive:getOnlineArchive": { - "description": "`mongodbatlas.OnlineArchive` describes an Online Archive\n\n\u003e **NOTE:** Groups and projects are synonymous terms. You may find group_id in the official documentation.\n\n\n{{% examples %}}\n## Example Usage\n{{% example %}}\n\n```typescript\nimport * as pulumi from \"@pulumi/pulumi\";\nimport * as mongodbatlas from \"@pulumi/mongodbatlas\";\n\nconst test = mongodbatlas.getOnlineArchive({\n projectId: _var.project_id,\n clusterName: _var.cluster_name,\n archiveId: \"5ebad3c1fe9c0ab8d37d61e1\",\n});\n```\n```python\nimport pulumi\nimport pulumi_mongodbatlas as mongodbatlas\n\ntest = mongodbatlas.get_online_archive(project_id=var[\"project_id\"],\n cluster_name=var[\"cluster_name\"],\n archive_id=\"5ebad3c1fe9c0ab8d37d61e1\")\n```\n```csharp\nusing System.Collections.Generic;\nusing System.Linq;\nusing Pulumi;\nusing Mongodbatlas = Pulumi.Mongodbatlas;\n\nreturn await Deployment.RunAsync(() =\u003e \n{\n var test = Mongodbatlas.GetOnlineArchive.Invoke(new()\n {\n ProjectId = @var.Project_id,\n ClusterName = @var.Cluster_name,\n ArchiveId = \"5ebad3c1fe9c0ab8d37d61e1\",\n });\n\n});\n```\n```go\npackage main\n\nimport (\n\t\"github.com/pulumi/pulumi-mongodbatlas/sdk/v3/go/mongodbatlas\"\n\t\"github.com/pulumi/pulumi/sdk/v3/go/pulumi\"\n)\n\nfunc main() {\n\tpulumi.Run(func(ctx *pulumi.Context) error {\n\t\t_, err := mongodbatlas.LookupOnlineArchive(ctx, \u0026mongodbatlas.LookupOnlineArchiveArgs{\n\t\t\tProjectId: _var.Project_id,\n\t\t\tClusterName: _var.Cluster_name,\n\t\t\tArchiveId: \"5ebad3c1fe9c0ab8d37d61e1\",\n\t\t}, nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n}\n```\n```java\npackage generated_program;\n\nimport com.pulumi.Context;\nimport com.pulumi.Pulumi;\nimport com.pulumi.core.Output;\nimport com.pulumi.mongodbatlas.MongodbatlasFunctions;\nimport com.pulumi.mongodbatlas.inputs.GetOnlineArchiveArgs;\nimport java.util.List;\nimport java.util.ArrayList;\nimport java.util.Map;\nimport java.io.File;\nimport java.nio.file.Files;\nimport java.nio.file.Paths;\n\npublic class App {\n public static void main(String[] args) {\n Pulumi.run(App::stack);\n }\n\n public static void stack(Context ctx) {\n final var test = MongodbatlasFunctions.getOnlineArchive(GetOnlineArchiveArgs.builder()\n .projectId(var_.project_id())\n .clusterName(var_.cluster_name())\n .archiveId(\"5ebad3c1fe9c0ab8d37d61e1\")\n .build());\n\n }\n}\n```\n```yaml\nvariables:\n test:\n fn::invoke:\n Function: mongodbatlas:getOnlineArchive\n Arguments:\n projectId: ${var.project_id}\n clusterName: ${var.cluster_name}\n archiveId: 5ebad3c1fe9c0ab8d37d61e1\n```\n{{% /example %}}\n{{% /examples %}}\n## Attributes reference\n\n* `db_name` - Name of the database that contains the collection.\n* `coll_name` - Name of the collection.\n* `collection_type` - Type of MongoDB collection that you want to return. This value can be \"TIMESERIES\" or \"STANDARD\". Default is \"STANDARD\".\n* `criteria` - Criteria to use for archiving data. See criteria.\n* `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule.\n* `schedule` - Regular frequency and duration when archiving process occurs. See schedule.\n* `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields.\n* `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails.\n* `state` - Status of the online archive. Valid values are: Pending, Archiving, Idle, Pausing, Paused, Orphaned and Deleted\n\n### Criteria\n* `type` - Type of criteria (DATE, CUSTOM)\n* `date_field` - Indexed database parameter that stores the date that determines when data moves to the online archive. MongoDB Cloud archives the data when the current date exceeds the date in this database parameter plus the number of days specified through the expireAfterDays parameter. Set this parameter when `type` is `DATE`.\n* `date_format` - Syntax used to write the date after which data moves to the online archive. Date can be expressed as ISO 8601 or Epoch timestamps. The Epoch timestamp can be expressed as nanoseconds, milliseconds, or seconds. Set this parameter when `type` is `DATE`. You must set `type` to `DATE` if `collectionType` is `TIMESERIES`. Valid values: ISODATE (default), EPOCH_SECONDS, EPOCH_MILLIS, EPOCH_NANOSECONDS.\n* `expire_after_days` - Number of days after the value in the criteria.dateField when MongoDB Cloud archives data in the specified cluster. Set this parameter when `type` is `DATE`.\n* `query` - JSON query to use to select documents for archiving. Atlas uses the specified query with the db.collection.find(query) command. The empty document {} to return all documents is not supported. Set this parameter when `type` is `CUSTOM`.\n\n### Data Expiration Rule\n* `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215.\n\n\n### Schedule\n\n* `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`.\n* `start_hour` - Hour of the day when the when the scheduled window to run one online archive starts. \n* `end_hour` - Hour of the day when the scheduled window to run one online archive ends.\n* `start_minute` - Minute of the hour when the scheduled window to run one online archive starts.\n* `end_minute` - Minute of the hour when the scheduled window to run one online archive ends.\n* `day_of_month` - Day of the month when the scheduled archive starts. Set this parameter when `type` is `MONTHLY`.\n* `day_of_week` - Day of the week when the scheduled archive starts. The week starts with Monday (1) and ends with Sunday (7).Set this parameter when `type` is `WEEKLY`.\n\n### Partition\n* `field_name` - Human-readable label that identifies the parameter that MongoDB Cloud uses to partition data. To specify a nested parameter, use the dot notation.\n* `order` - Sequence in which MongoDB Cloud slices the collection data to create partitions. The resource expresses this sequence starting with zero. The value of the `criteria.dateField` parameter defaults as the first item in the partition sequence.\n* `field_type` - Data type of the parameter that that MongoDB Cloud uses to partition data. Partition parameters of type UUID must be of binary subtype 4. MongoDB Cloud skips partition parameters of type UUID with subtype 3. Valid values: `date`, `int`, `long`, `objectId`, `string`, `uuid`.\n\nSee [MongoDB Atlas API](https://docs.atlas.mongodb.com/reference/api/online-archive-get-one/) Documentation for more information.\n", + "description": "`mongodbatlas.OnlineArchive` describes an Online Archive\n\n\u003e **NOTE:** Groups and projects are synonymous terms. You may find group_id in the official documentation.\n\n\n{{% examples %}}\n## Example Usage\n{{% example %}}\n\n```typescript\nimport * as pulumi from \"@pulumi/pulumi\";\nimport * as mongodbatlas from \"@pulumi/mongodbatlas\";\n\nconst test = mongodbatlas.getOnlineArchive({\n projectId: _var.project_id,\n clusterName: _var.cluster_name,\n archiveId: \"5ebad3c1fe9c0ab8d37d61e1\",\n});\n```\n```python\nimport pulumi\nimport pulumi_mongodbatlas as mongodbatlas\n\ntest = mongodbatlas.get_online_archive(project_id=var[\"project_id\"],\n cluster_name=var[\"cluster_name\"],\n archive_id=\"5ebad3c1fe9c0ab8d37d61e1\")\n```\n```csharp\nusing System.Collections.Generic;\nusing System.Linq;\nusing Pulumi;\nusing Mongodbatlas = Pulumi.Mongodbatlas;\n\nreturn await Deployment.RunAsync(() =\u003e \n{\n var test = Mongodbatlas.GetOnlineArchive.Invoke(new()\n {\n ProjectId = @var.Project_id,\n ClusterName = @var.Cluster_name,\n ArchiveId = \"5ebad3c1fe9c0ab8d37d61e1\",\n });\n\n});\n```\n```go\npackage main\n\nimport (\n\t\"github.com/pulumi/pulumi-mongodbatlas/sdk/v3/go/mongodbatlas\"\n\t\"github.com/pulumi/pulumi/sdk/v3/go/pulumi\"\n)\n\nfunc main() {\n\tpulumi.Run(func(ctx *pulumi.Context) error {\n\t\t_, err := mongodbatlas.LookupOnlineArchive(ctx, \u0026mongodbatlas.LookupOnlineArchiveArgs{\n\t\t\tProjectId: _var.Project_id,\n\t\t\tClusterName: _var.Cluster_name,\n\t\t\tArchiveId: \"5ebad3c1fe9c0ab8d37d61e1\",\n\t\t}, nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n}\n```\n```java\npackage generated_program;\n\nimport com.pulumi.Context;\nimport com.pulumi.Pulumi;\nimport com.pulumi.core.Output;\nimport com.pulumi.mongodbatlas.MongodbatlasFunctions;\nimport com.pulumi.mongodbatlas.inputs.GetOnlineArchiveArgs;\nimport java.util.List;\nimport java.util.ArrayList;\nimport java.util.Map;\nimport java.io.File;\nimport java.nio.file.Files;\nimport java.nio.file.Paths;\n\npublic class App {\n public static void main(String[] args) {\n Pulumi.run(App::stack);\n }\n\n public static void stack(Context ctx) {\n final var test = MongodbatlasFunctions.getOnlineArchive(GetOnlineArchiveArgs.builder()\n .projectId(var_.project_id())\n .clusterName(var_.cluster_name())\n .archiveId(\"5ebad3c1fe9c0ab8d37d61e1\")\n .build());\n\n }\n}\n```\n```yaml\nvariables:\n test:\n fn::invoke:\n Function: mongodbatlas:getOnlineArchive\n Arguments:\n projectId: ${var.project_id}\n clusterName: ${var.cluster_name}\n archiveId: 5ebad3c1fe9c0ab8d37d61e1\n```\n{{% /example %}}\n{{% /examples %}}\n## Attributes reference\n\n* `db_name` - Name of the database that contains the collection.\n* `coll_name` - Name of the collection.\n* `collection_type` - Type of MongoDB collection that you want to return. This value can be \"TIMESERIES\" or \"STANDARD\". Default is \"STANDARD\".\n* `criteria` - Criteria to use for archiving data. See criteria.\n* `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule.\n* `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region.\n* `schedule` - Regular frequency and duration when archiving process occurs. See schedule.\n* `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields.\n* `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails.\n* `state` - Status of the online archive. Valid values are: Pending, Archiving, Idle, Pausing, Paused, Orphaned and Deleted\n\n### Criteria\n* `type` - Type of criteria (DATE, CUSTOM)\n* `date_field` - Indexed database parameter that stores the date that determines when data moves to the online archive. MongoDB Cloud archives the data when the current date exceeds the date in this database parameter plus the number of days specified through the expireAfterDays parameter. Set this parameter when `type` is `DATE`.\n* `date_format` - Syntax used to write the date after which data moves to the online archive. Date can be expressed as ISO 8601 or Epoch timestamps. The Epoch timestamp can be expressed as nanoseconds, milliseconds, or seconds. Set this parameter when `type` is `DATE`. You must set `type` to `DATE` if `collectionType` is `TIMESERIES`. Valid values: ISODATE (default), EPOCH_SECONDS, EPOCH_MILLIS, EPOCH_NANOSECONDS.\n* `expire_after_days` - Number of days after the value in the criteria.dateField when MongoDB Cloud archives data in the specified cluster. Set this parameter when `type` is `DATE`.\n* `query` - JSON query to use to select documents for archiving. Atlas uses the specified query with the db.collection.find(query) command. The empty document {} to return all documents is not supported. Set this parameter when `type` is `CUSTOM`.\n\n### Data Expiration Rule\n* `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215.\n\n### Data Process Region\n* `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data.\n* `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive)\n\n\n### Schedule\n\n* `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`.\n* `start_hour` - Hour of the day when the when the scheduled window to run one online archive starts. \n* `end_hour` - Hour of the day when the scheduled window to run one online archive ends.\n* `start_minute` - Minute of the hour when the scheduled window to run one online archive starts.\n* `end_minute` - Minute of the hour when the scheduled window to run one online archive ends.\n* `day_of_month` - Day of the month when the scheduled archive starts. Set this parameter when `type` is `MONTHLY`.\n* `day_of_week` - Day of the week when the scheduled archive starts. The week starts with Monday (1) and ends with Sunday (7).Set this parameter when `type` is `WEEKLY`.\n\n### Partition\n* `field_name` - Human-readable label that identifies the parameter that MongoDB Cloud uses to partition data. To specify a nested parameter, use the dot notation.\n* `order` - Sequence in which MongoDB Cloud slices the collection data to create partitions. The resource expresses this sequence starting with zero. The value of the `criteria.dateField` parameter defaults as the first item in the partition sequence.\n* `field_type` - Data type of the parameter that that MongoDB Cloud uses to partition data. Partition parameters of type UUID must be of binary subtype 4. MongoDB Cloud skips partition parameters of type UUID with subtype 3. Valid values: `date`, `int`, `long`, `objectId`, `string`, `uuid`.\n\nSee [MongoDB Atlas API](https://docs.atlas.mongodb.com/reference/api/online-archive-get-one/) Documentation for more information.\n", "inputs": { "description": "A collection of arguments for invoking getOnlineArchive.\n", "properties": { @@ -24096,6 +24179,12 @@ "$ref": "#/types/mongodbatlas:index/getOnlineArchiveDataExpirationRule:getOnlineArchiveDataExpirationRule" } }, + "dataProcessRegions": { + "type": "array", + "items": { + "$ref": "#/types/mongodbatlas:index/getOnlineArchiveDataProcessRegion:getOnlineArchiveDataProcessRegion" + } + }, "dbName": { "type": "string" }, @@ -24133,6 +24222,7 @@ "collectionType", "criterias", "dataExpirationRules", + "dataProcessRegions", "dbName", "partitionFields", "paused", @@ -24144,7 +24234,7 @@ } }, "mongodbatlas:index/getOnlineArchives:getOnlineArchives": { - "description": "`mongodbatlas.OnlineArchive` Describes the list of all the online archives for a cluster\n\n\u003e **NOTE:** Groups and projects are synonymous terms. You may find group_id in the official documentation.\n\n\n{{% examples %}}\n## Example Usage\n{{% example %}}\n\n```typescript\nimport * as pulumi from \"@pulumi/pulumi\";\nimport * as mongodbatlas from \"@pulumi/mongodbatlas\";\n\nconst test = mongodbatlas.getOnlineArchives({\n projectId: _var.project_id,\n clusterName: _var.cluster_name,\n});\n```\n```python\nimport pulumi\nimport pulumi_mongodbatlas as mongodbatlas\n\ntest = mongodbatlas.get_online_archives(project_id=var[\"project_id\"],\n cluster_name=var[\"cluster_name\"])\n```\n```csharp\nusing System.Collections.Generic;\nusing System.Linq;\nusing Pulumi;\nusing Mongodbatlas = Pulumi.Mongodbatlas;\n\nreturn await Deployment.RunAsync(() =\u003e \n{\n var test = Mongodbatlas.GetOnlineArchives.Invoke(new()\n {\n ProjectId = @var.Project_id,\n ClusterName = @var.Cluster_name,\n });\n\n});\n```\n```go\npackage main\n\nimport (\n\t\"github.com/pulumi/pulumi-mongodbatlas/sdk/v3/go/mongodbatlas\"\n\t\"github.com/pulumi/pulumi/sdk/v3/go/pulumi\"\n)\n\nfunc main() {\n\tpulumi.Run(func(ctx *pulumi.Context) error {\n\t\t_, err := mongodbatlas.LookupOnlineArchives(ctx, \u0026mongodbatlas.LookupOnlineArchivesArgs{\n\t\t\tProjectId: _var.Project_id,\n\t\t\tClusterName: _var.Cluster_name,\n\t\t}, nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n}\n```\n```java\npackage generated_program;\n\nimport com.pulumi.Context;\nimport com.pulumi.Pulumi;\nimport com.pulumi.core.Output;\nimport com.pulumi.mongodbatlas.MongodbatlasFunctions;\nimport com.pulumi.mongodbatlas.inputs.GetOnlineArchivesArgs;\nimport java.util.List;\nimport java.util.ArrayList;\nimport java.util.Map;\nimport java.io.File;\nimport java.nio.file.Files;\nimport java.nio.file.Paths;\n\npublic class App {\n public static void main(String[] args) {\n Pulumi.run(App::stack);\n }\n\n public static void stack(Context ctx) {\n final var test = MongodbatlasFunctions.getOnlineArchives(GetOnlineArchivesArgs.builder()\n .projectId(var_.project_id())\n .clusterName(var_.cluster_name())\n .build());\n\n }\n}\n```\n```yaml\nvariables:\n test:\n fn::invoke:\n Function: mongodbatlas:getOnlineArchives\n Arguments:\n projectId: ${var.project_id}\n clusterName: ${var.cluster_name}\n```\n{{% /example %}}\n{{% /examples %}}\n## Attributes reference\n\n* `db_name` - Name of the database that contains the collection.\n* `coll_name` - Name of the collection.\n* `collection_type` - Type of MongoDB collection that you want to return. This value can be \"TIMESERIES\" or \"STANDARD\". Default is \"STANDARD\". \n* `criteria` - Criteria to use for archiving data. See criteria.\n* `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule.\n* `schedule` - Regular frequency and duration when archiving process occurs. See schedule.\n* `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields.\n* `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails.\n* `state` - Status of the online archive. Valid values are: Pending, Archiving, Idle, Pausing, Paused, Orphaned and Deleted\n\n### Criteria\n* `type` - Type of criteria (DATE, CUSTOM)\n* `date_field` - Indexed database parameter that stores the date that determines when data moves to the online archive. MongoDB Cloud archives the data when the current date exceeds the date in this database parameter plus the number of days specified through the expireAfterDays parameter. Set this parameter when `type` is `DATE`.\n* `date_format` - Syntax used to write the date after which data moves to the online archive. Date can be expressed as ISO 8601 or Epoch timestamps. The Epoch timestamp can be expressed as nanoseconds, milliseconds, or seconds. Set this parameter when `type` is `DATE`. You must set `type` to `DATE` if `collectionType` is `TIMESERIES`. Valid values: ISODATE (default), EPOCH_SECONDS, EPOCH_MILLIS, EPOCH_NANOSECONDS.\n* `expire_after_days` - Number of days after the value in the criteria.dateField when MongoDB Cloud archives data in the specified cluster. Set this parameter when `type` is `DATE`.\n* `query` - JSON query to use to select documents for archiving. Atlas uses the specified query with the db.collection.find(query) command. The empty document {} to return all documents is not supported. Set this parameter when `type` is `CUSTOM`.\n\n### Data Expiration Rule\n* `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215.\n\n### Schedule\n\n* `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`).\n* `start_hour` - Hour of the day when the when the scheduled window to run one online archive starts. \n* `end_hour` - Hour of the day when the scheduled window to run one online archive ends.\n* `start_minute` - Minute of the hour when the scheduled window to run one online archive starts.\n* `end_minute` - Minute of the hour when the scheduled window to run one online archive ends.\n* `day_of_month` - Day of the month when the scheduled archive starts.\n* `day_of_week` - Day of the week when the scheduled archive starts. The week starts with Monday (1) and ends with Sunday (7).\n\n### Partition\n* `field_name` - Human-readable label that identifies the parameter that MongoDB Cloud uses to partition data. To specify a nested parameter, use the dot notation.\n* `order` - Sequence in which MongoDB Cloud slices the collection data to create partitions. The resource expresses this sequence starting with zero. The value of the `criteria.dateField` parameter defaults as the first item in the partition sequence.\n* `field_type` - Data type of the parameter that that MongoDB Cloud uses to partition data. Partition parameters of type UUID must be of binary subtype 4. MongoDB Cloud skips partition parameters of type UUID with subtype 3. Valid values: `date`, `int`, `long`, `objectId`, `string`, `uuid`.\n", + "description": "`mongodbatlas.OnlineArchive` Describes the list of all the online archives for a cluster\n\n\u003e **NOTE:** Groups and projects are synonymous terms. You may find group_id in the official documentation.\n\n\n{{% examples %}}\n## Example Usage\n{{% example %}}\n\n```typescript\nimport * as pulumi from \"@pulumi/pulumi\";\nimport * as mongodbatlas from \"@pulumi/mongodbatlas\";\n\nconst test = mongodbatlas.getOnlineArchives({\n projectId: _var.project_id,\n clusterName: _var.cluster_name,\n});\n```\n```python\nimport pulumi\nimport pulumi_mongodbatlas as mongodbatlas\n\ntest = mongodbatlas.get_online_archives(project_id=var[\"project_id\"],\n cluster_name=var[\"cluster_name\"])\n```\n```csharp\nusing System.Collections.Generic;\nusing System.Linq;\nusing Pulumi;\nusing Mongodbatlas = Pulumi.Mongodbatlas;\n\nreturn await Deployment.RunAsync(() =\u003e \n{\n var test = Mongodbatlas.GetOnlineArchives.Invoke(new()\n {\n ProjectId = @var.Project_id,\n ClusterName = @var.Cluster_name,\n });\n\n});\n```\n```go\npackage main\n\nimport (\n\t\"github.com/pulumi/pulumi-mongodbatlas/sdk/v3/go/mongodbatlas\"\n\t\"github.com/pulumi/pulumi/sdk/v3/go/pulumi\"\n)\n\nfunc main() {\n\tpulumi.Run(func(ctx *pulumi.Context) error {\n\t\t_, err := mongodbatlas.LookupOnlineArchives(ctx, \u0026mongodbatlas.LookupOnlineArchivesArgs{\n\t\t\tProjectId: _var.Project_id,\n\t\t\tClusterName: _var.Cluster_name,\n\t\t}, nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n}\n```\n```java\npackage generated_program;\n\nimport com.pulumi.Context;\nimport com.pulumi.Pulumi;\nimport com.pulumi.core.Output;\nimport com.pulumi.mongodbatlas.MongodbatlasFunctions;\nimport com.pulumi.mongodbatlas.inputs.GetOnlineArchivesArgs;\nimport java.util.List;\nimport java.util.ArrayList;\nimport java.util.Map;\nimport java.io.File;\nimport java.nio.file.Files;\nimport java.nio.file.Paths;\n\npublic class App {\n public static void main(String[] args) {\n Pulumi.run(App::stack);\n }\n\n public static void stack(Context ctx) {\n final var test = MongodbatlasFunctions.getOnlineArchives(GetOnlineArchivesArgs.builder()\n .projectId(var_.project_id())\n .clusterName(var_.cluster_name())\n .build());\n\n }\n}\n```\n```yaml\nvariables:\n test:\n fn::invoke:\n Function: mongodbatlas:getOnlineArchives\n Arguments:\n projectId: ${var.project_id}\n clusterName: ${var.cluster_name}\n```\n{{% /example %}}\n{{% /examples %}}\n## Attributes reference\n\n* `db_name` - Name of the database that contains the collection.\n* `coll_name` - Name of the collection.\n* `collection_type` - Type of MongoDB collection that you want to return. This value can be \"TIMESERIES\" or \"STANDARD\". Default is \"STANDARD\". \n* `criteria` - Criteria to use for archiving data. See criteria.\n* `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule.\n* `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region.\n* `schedule` - Regular frequency and duration when archiving process occurs. See schedule.\n* `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields.\n* `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails.\n* `state` - Status of the online archive. Valid values are: Pending, Archiving, Idle, Pausing, Paused, Orphaned and Deleted\n\n### Criteria\n* `type` - Type of criteria (DATE, CUSTOM)\n* `date_field` - Indexed database parameter that stores the date that determines when data moves to the online archive. MongoDB Cloud archives the data when the current date exceeds the date in this database parameter plus the number of days specified through the expireAfterDays parameter. Set this parameter when `type` is `DATE`.\n* `date_format` - Syntax used to write the date after which data moves to the online archive. Date can be expressed as ISO 8601 or Epoch timestamps. The Epoch timestamp can be expressed as nanoseconds, milliseconds, or seconds. Set this parameter when `type` is `DATE`. You must set `type` to `DATE` if `collectionType` is `TIMESERIES`. Valid values: ISODATE (default), EPOCH_SECONDS, EPOCH_MILLIS, EPOCH_NANOSECONDS.\n* `expire_after_days` - Number of days after the value in the criteria.dateField when MongoDB Cloud archives data in the specified cluster. Set this parameter when `type` is `DATE`.\n* `query` - JSON query to use to select documents for archiving. Atlas uses the specified query with the db.collection.find(query) command. The empty document {} to return all documents is not supported. Set this parameter when `type` is `CUSTOM`.\n\n### Data Expiration Rule\n* `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215.\n\n### Data Process Region\n* `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data.\n* `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive)\n\n### Schedule\n\n* `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`).\n* `start_hour` - Hour of the day when the when the scheduled window to run one online archive starts. \n* `end_hour` - Hour of the day when the scheduled window to run one online archive ends.\n* `start_minute` - Minute of the hour when the scheduled window to run one online archive starts.\n* `end_minute` - Minute of the hour when the scheduled window to run one online archive ends.\n* `day_of_month` - Day of the month when the scheduled archive starts.\n* `day_of_week` - Day of the week when the scheduled archive starts. The week starts with Monday (1) and ends with Sunday (7).\n\n### Partition\n* `field_name` - Human-readable label that identifies the parameter that MongoDB Cloud uses to partition data. To specify a nested parameter, use the dot notation.\n* `order` - Sequence in which MongoDB Cloud slices the collection data to create partitions. The resource expresses this sequence starting with zero. The value of the `criteria.dateField` parameter defaults as the first item in the partition sequence.\n* `field_type` - Data type of the parameter that that MongoDB Cloud uses to partition data. Partition parameters of type UUID must be of binary subtype 4. MongoDB Cloud skips partition parameters of type UUID with subtype 3. Valid values: `date`, `int`, `long`, `objectId`, `string`, `uuid`.\n", "inputs": { "description": "A collection of arguments for invoking getOnlineArchives.\n", "properties": { diff --git a/provider/go.mod b/provider/go.mod index 218edf3d..25c9d4dd 100644 --- a/provider/go.mod +++ b/provider/go.mod @@ -5,18 +5,18 @@ go 1.21 replace github.com/hashicorp/terraform-plugin-sdk/v2 => github.com/pulumi/terraform-plugin-sdk/v2 v2.0.0-20230912190043-e6d96b3b8f7e require ( - github.com/mongodb/terraform-provider-mongodbatlas v1.13.0 + github.com/mongodb/terraform-provider-mongodbatlas v1.13.1 github.com/pulumi/pulumi-terraform-bridge/pf v0.20.0 github.com/pulumi/pulumi-terraform-bridge/v3 v3.66.0 github.com/pulumi/pulumi/sdk/v3 v3.94.2 ) require ( - cloud.google.com/go v0.110.4 // indirect - cloud.google.com/go/compute v1.20.1 // indirect + cloud.google.com/go v0.110.7 // indirect + cloud.google.com/go/compute v1.23.0 // indirect cloud.google.com/go/compute/metadata v0.2.3 // indirect cloud.google.com/go/iam v1.1.1 // indirect - cloud.google.com/go/kms v1.12.1 // indirect + cloud.google.com/go/kms v1.15.0 // indirect cloud.google.com/go/logging v1.7.0 // indirect cloud.google.com/go/longrunning v0.5.1 // indirect cloud.google.com/go/storage v1.30.1 // indirect @@ -47,7 +47,7 @@ require ( github.com/armon/go-metrics v0.4.0 // indirect github.com/armon/go-radix v1.0.0 // indirect github.com/atotto/clipboard v0.1.4 // indirect - github.com/aws/aws-sdk-go v1.47.11 // indirect + github.com/aws/aws-sdk-go v1.48.2 // indirect github.com/aws/aws-sdk-go-v2 v1.17.3 // indirect github.com/aws/aws-sdk-go-v2/config v1.15.15 // indirect github.com/aws/aws-sdk-go-v2/credentials v1.12.10 // indirect @@ -87,7 +87,7 @@ require ( github.com/gofrs/uuid v4.2.0+incompatible // indirect github.com/gogo/protobuf v1.3.2 // indirect github.com/golang-jwt/jwt/v4 v4.4.2 // indirect - github.com/golang/glog v1.1.0 // indirect + github.com/golang/glog v1.1.2 // indirect github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da // indirect github.com/golang/protobuf v1.5.3 // indirect github.com/golang/snappy v0.0.4 // indirect @@ -95,7 +95,7 @@ require ( github.com/google/go-querystring v1.1.0 // indirect github.com/google/s2a-go v0.1.4 // indirect github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 // indirect - github.com/google/uuid v1.3.0 // indirect + github.com/google/uuid v1.3.1 // indirect github.com/google/wire v0.5.0 // indirect github.com/googleapis/enterprise-certificate-proxy v0.2.3 // indirect github.com/googleapis/gax-go/v2 v2.11.0 // indirect @@ -108,7 +108,7 @@ require ( github.com/hashicorp/go-hclog v1.5.0 // indirect github.com/hashicorp/go-immutable-radix v1.3.1 // indirect github.com/hashicorp/go-multierror v1.1.1 // indirect - github.com/hashicorp/go-plugin v1.5.1 // indirect + github.com/hashicorp/go-plugin v1.5.2 // indirect github.com/hashicorp/go-retryablehttp v0.7.1 // indirect github.com/hashicorp/go-rootcerts v1.0.2 // indirect github.com/hashicorp/go-safetemp v1.0.0 // indirect @@ -126,12 +126,12 @@ require ( github.com/hashicorp/terraform-plugin-framework v1.4.2 // indirect github.com/hashicorp/terraform-plugin-framework-timeouts v0.4.1 // indirect github.com/hashicorp/terraform-plugin-framework-validators v0.12.0 // indirect - github.com/hashicorp/terraform-plugin-go v0.19.0 // indirect + github.com/hashicorp/terraform-plugin-go v0.19.1 // indirect github.com/hashicorp/terraform-plugin-log v0.9.0 // indirect github.com/hashicorp/terraform-plugin-mux v0.12.0 // indirect github.com/hashicorp/terraform-plugin-sdk v1.17.2 // indirect github.com/hashicorp/terraform-plugin-sdk/v2 v2.30.0 // indirect - github.com/hashicorp/terraform-registry-address v0.2.2 // indirect + github.com/hashicorp/terraform-registry-address v0.2.3 // indirect github.com/hashicorp/terraform-svchost v0.1.1 // indirect github.com/hashicorp/vault/api v1.8.2 // indirect github.com/hashicorp/vault/sdk v0.6.1 // indirect @@ -207,7 +207,7 @@ require ( github.com/uber/jaeger-lib v2.4.1+incompatible // indirect github.com/ulikunitz/xz v0.5.11 // indirect github.com/vmihailenco/msgpack v4.0.4+incompatible // indirect - github.com/vmihailenco/msgpack/v5 v5.3.5 // indirect + github.com/vmihailenco/msgpack/v5 v5.4.1 // indirect github.com/vmihailenco/tagparser/v2 v2.0.0 // indirect github.com/xanzy/ssh-agent v0.3.3 // indirect github.com/xeipuuv/gojsonpointer v0.0.0-20180127040702-4e3ac2762d5f // indirect @@ -226,7 +226,7 @@ require ( golang.org/x/exp v0.0.0-20231006140011-7918f672742d // indirect golang.org/x/mod v0.13.0 // indirect golang.org/x/net v0.17.0 // indirect - golang.org/x/oauth2 v0.8.0 // indirect + golang.org/x/oauth2 v0.11.0 // indirect golang.org/x/sync v0.4.0 // indirect golang.org/x/sys v0.14.0 // indirect golang.org/x/term v0.14.0 // indirect @@ -236,10 +236,10 @@ require ( golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2 // indirect google.golang.org/api v0.126.0 // indirect google.golang.org/appengine v1.6.7 // indirect - google.golang.org/genproto v0.0.0-20230726155614-23370e0ffb3e // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20230706204954-ccb25ca9f130 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20230706204954-ccb25ca9f130 // indirect - google.golang.org/grpc v1.57.1 // indirect + google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20230822172742-b8732ec3820d // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20230822172742-b8732ec3820d // indirect + google.golang.org/grpc v1.59.0 // indirect google.golang.org/protobuf v1.31.0 // indirect gopkg.in/square/go-jose.v2 v2.6.0 // indirect gopkg.in/warnings.v0 v0.1.2 // indirect diff --git a/provider/go.sum b/provider/go.sum index 02072be2..8b2c99fe 100644 --- a/provider/go.sum +++ b/provider/go.sum @@ -40,8 +40,8 @@ cloud.google.com/go v0.104.0/go.mod h1:OO6xxXdJyvuJPcEPBLN9BJPD+jep5G1+2U5B5gkRY cloud.google.com/go v0.105.0/go.mod h1:PrLgOJNe5nfE9UMxKxgXj4mD3voiP+YQ6gdt6KMFOKM= cloud.google.com/go v0.107.0/go.mod h1:wpc2eNrD7hXUTy8EKS10jkxpZBjASrORK7goS+3YX2I= cloud.google.com/go v0.110.0/go.mod h1:SJnCLqQ0FCFGSZMUNUf84MV3Aia54kn7pi8st7tMzaY= -cloud.google.com/go v0.110.4 h1:1JYyxKMN9hd5dR2MYTPWkGUgcoxVVhg0LKNKEo0qvmk= -cloud.google.com/go v0.110.4/go.mod h1:+EYjdK8e5RME/VY/qLCAtuyALQ9q67dvuum8i+H5xsI= +cloud.google.com/go v0.110.7 h1:rJyC7nWRg2jWGZ4wSJ5nY65GTdYJkg0cd/uXb+ACI6o= +cloud.google.com/go v0.110.7/go.mod h1:+EYjdK8e5RME/VY/qLCAtuyALQ9q67dvuum8i+H5xsI= cloud.google.com/go/accessapproval v1.4.0/go.mod h1:zybIuC3KpDOvotz59lFe5qxRZx6C75OtwbisN56xYB4= cloud.google.com/go/accessapproval v1.5.0/go.mod h1:HFy3tuiGvMdcd/u+Cu5b9NkO1pEICJ46IR82PoUdplw= cloud.google.com/go/accessapproval v1.6.0/go.mod h1:R0EiYnwV5fsRFiKZkPHr6mwyk2wxUJ30nL4j2pcFY2E= @@ -178,8 +178,8 @@ cloud.google.com/go/compute v1.15.1/go.mod h1:bjjoF/NtFUrkD/urWfdHaKuOPDR5nWIs63 cloud.google.com/go/compute v1.18.0/go.mod h1:1X7yHxec2Ga+Ss6jPyjxRxpu2uu7PLgsOVXvgU0yacs= cloud.google.com/go/compute v1.19.0/go.mod h1:rikpw2y+UMidAe9tISo04EHNOIf42RLYF/q8Bs93scU= cloud.google.com/go/compute v1.19.1/go.mod h1:6ylj3a05WF8leseCdIf77NK0g1ey+nj5IKd5/kvShxE= -cloud.google.com/go/compute v1.20.1 h1:6aKEtlUiwEpJzM001l0yFkpXmUVXaN8W+fbkb2AZNbg= -cloud.google.com/go/compute v1.20.1/go.mod h1:4tCnrn48xsqlwSAiLf1HXMQk8CONslYbdiEZc9FEIbM= +cloud.google.com/go/compute v1.23.0 h1:tP41Zoavr8ptEqaW6j+LQOnyBBhO7OkOMAGrgLopTwY= +cloud.google.com/go/compute v1.23.0/go.mod h1:4tCnrn48xsqlwSAiLf1HXMQk8CONslYbdiEZc9FEIbM= cloud.google.com/go/compute/metadata v0.1.0/go.mod h1:Z1VN+bulIf6bt4P/C37K4DyZYZEXYonfTBHHFPO/4UU= cloud.google.com/go/compute/metadata v0.2.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= cloud.google.com/go/compute/metadata v0.2.1/go.mod h1:jgHgmJd2RKBGzXqF5LR2EZMGxBkeanZ9wwa75XHJgOM= @@ -341,8 +341,8 @@ cloud.google.com/go/kms v1.8.0/go.mod h1:4xFEhYFqvW+4VMELtZyxomGSYtSQKzM178ylFW4 cloud.google.com/go/kms v1.9.0/go.mod h1:qb1tPTgfF9RQP8e1wq4cLFErVuTJv7UsSC915J8dh3w= cloud.google.com/go/kms v1.10.0/go.mod h1:ng3KTUtQQU9bPX3+QGLsflZIHlkbn8amFAMY63m8d24= cloud.google.com/go/kms v1.10.1/go.mod h1:rIWk/TryCkR59GMC3YtHtXeLzd634lBbKenvyySAyYI= -cloud.google.com/go/kms v1.12.1 h1:xZmZuwy2cwzsocmKDOPu4BL7umg8QXagQx6fKVmf45U= -cloud.google.com/go/kms v1.12.1/go.mod h1:c9J991h5DTl+kg7gi3MYomh12YEENGrf48ee/N/2CDM= +cloud.google.com/go/kms v1.15.0 h1:xYl5WEaSekKYN5gGRyhjvZKM22GVBBCzegGNVPy+aIs= +cloud.google.com/go/kms v1.15.0/go.mod h1:c9J991h5DTl+kg7gi3MYomh12YEENGrf48ee/N/2CDM= cloud.google.com/go/language v1.4.0/go.mod h1:F9dRpNFQmJbkaop6g0JhSBXCNlO90e1KWx5iDdxbWic= cloud.google.com/go/language v1.6.0/go.mod h1:6dJ8t3B+lUYfStgls25GusK04NLh3eDLQnWM3mdEbhI= cloud.google.com/go/language v1.7.0/go.mod h1:DJ6dYN/W+SQOjF8e1hLQXMF21AkH2w9wiPzPCJa2MIE= @@ -823,8 +823,8 @@ github.com/aws/aws-sdk-go v1.43.31/go.mod h1:y4AeaBuwd2Lk+GepC1E9v0qOiTws0MIWAX4 github.com/aws/aws-sdk-go v1.44.45/go.mod h1:y4AeaBuwd2Lk+GepC1E9v0qOiTws0MIWAX4oIKwKHZo= github.com/aws/aws-sdk-go v1.44.68/go.mod h1:y4AeaBuwd2Lk+GepC1E9v0qOiTws0MIWAX4oIKwKHZo= github.com/aws/aws-sdk-go v1.44.122/go.mod h1:y4AeaBuwd2Lk+GepC1E9v0qOiTws0MIWAX4oIKwKHZo= -github.com/aws/aws-sdk-go v1.47.11 h1:Dol+MA+hQblbnXUI3Vk9qvoekU6O1uDEuAItezjiWNQ= -github.com/aws/aws-sdk-go v1.47.11/go.mod h1:LF8svs817+Nz+DmiMQKTO3ubZ/6IaTpq3TjupRn3Eqk= +github.com/aws/aws-sdk-go v1.48.2 h1:Lf7+Y4WmHB0AQLRQZA46diSwDa+LWbwY6IGaYoCVtTc= +github.com/aws/aws-sdk-go v1.48.2/go.mod h1:LF8svs817+Nz+DmiMQKTO3ubZ/6IaTpq3TjupRn3Eqk= github.com/aws/aws-sdk-go-v2 v0.18.0/go.mod h1:JWVYvqSMppoMJC0x5wdwiImzgXTI9FuZwxzkQq9wy+g= github.com/aws/aws-sdk-go-v2 v1.16.8/go.mod h1:6CpKuLXg2w7If3ABZCl/qZ6rEgwtjZTn4eAf4RcEyuw= github.com/aws/aws-sdk-go-v2 v1.17.3 h1:shN7NlnVzvDUgPQ+1rLMSxY8OWRNDRYtiqe0p/PgrhY= @@ -1375,8 +1375,9 @@ github.com/golang-sql/sqlexp v0.1.0/go.mod h1:J4ad9Vo8ZCWQ2GMrC4UCQy1JpCbwU9m3EO github.com/golang/freetype v0.0.0-20170609003504-e2365dfdc4a0/go.mod h1:E/TSTwGwJL78qG/PmXZO1EjYhfJinVAhrmmHX6Z8B9k= github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfUlaOkMMuAQtPIUF8ecWP5IEl/CR7VP2Q= github.com/golang/glog v1.0.0/go.mod h1:EWib/APOK0SL3dFbYqvxE3UYd8E6s1ouQ7iEp/0LWV4= -github.com/golang/glog v1.1.0 h1:/d3pCKDPWNnvIWe0vVUpNP32qc8U3PDVxySP/y360qE= github.com/golang/glog v1.1.0/go.mod h1:pfYeQZ3JWZoXTV5sFc986z3HTpwQs9At6P4ImfuP3NQ= +github.com/golang/glog v1.1.2 h1:DVjP2PbBOzHyzA+dn3WhHIq4NdVu3Q+pvivFICf/7fo= +github.com/golang/glog v1.1.2/go.mod h1:zR+okUeTbrL6EL3xHUDxZuEtGv04p5shwip1+mL/rLQ= github.com/golang/groupcache v0.0.0-20160516000752-02826c3e7903/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20190129154638-5b532d6fd5ef/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20190702054246-869f871628b6/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= @@ -1487,8 +1488,9 @@ github.com/google/uuid v1.0.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+ github.com/google/uuid v1.1.1/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/google/uuid v1.1.2/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/google/uuid v1.2.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= -github.com/google/uuid v1.3.0 h1:t6JiXgmwXMjEs8VusXIJk2BXHsn+wx8BZdTaoZ5fu7I= github.com/google/uuid v1.3.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= +github.com/google/uuid v1.3.1 h1:KjJaJ9iWZ3jOFZIf1Lqf4laDRCasjl0BCmnEGxkdLb4= +github.com/google/uuid v1.3.1/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/google/wire v0.5.0 h1:I7ELFeVBr3yfPIcc8+MWvrjk+3VjbcSzoXm3JVa+jD8= github.com/google/wire v0.5.0/go.mod h1:ngWDr9Qvq3yZA10YrxfyGELY/AFWGVpy9c1LTRi1EoU= github.com/googleapis/enterprise-certificate-proxy v0.0.0-20220520183353-fd19c99a87aa/go.mod h1:17drOmN3MwGY7t0e+Ei9b45FFGA3fBs3x36SsCg1hq8= @@ -1593,8 +1595,9 @@ github.com/hashicorp/go-plugin v1.0.1/go.mod h1:++UyYGoz3o5w9ZzAdZxtQKrWWP+iqPBn github.com/hashicorp/go-plugin v1.3.0/go.mod h1:F9eH4LrE/ZsRdbwhfjs9k9HoDUwAHnYtXdgmf1AVNs0= github.com/hashicorp/go-plugin v1.4.3/go.mod h1:5fGEH17QVwTTcR0zV7yhDPLLmFX9YSZ38b18Udy6vYQ= github.com/hashicorp/go-plugin v1.4.4/go.mod h1:viDMjcLJuDui6pXb8U4HVfb8AamCWhHGUjr2IrTF67s= -github.com/hashicorp/go-plugin v1.5.1 h1:oGm7cWBaYIp3lJpx1RUEfLWophprE2EV/KUeqBYo+6k= github.com/hashicorp/go-plugin v1.5.1/go.mod h1:w1sAEES3g3PuV/RzUrgow20W2uErMly84hhD3um1WL4= +github.com/hashicorp/go-plugin v1.5.2 h1:aWv8eimFqWlsEiMrYZdPYl+FdHaBJSN4AWwGWfT1G2Y= +github.com/hashicorp/go-plugin v1.5.2/go.mod h1:w1sAEES3g3PuV/RzUrgow20W2uErMly84hhD3um1WL4= github.com/hashicorp/go-retryablehttp v0.5.3/go.mod h1:9B5zBasrRhHXnJnui7y6sL7es7NDiJgTc6Er0maI1Xs= github.com/hashicorp/go-retryablehttp v0.6.6/go.mod h1:vAew36LZh98gCBJNLH42IQ1ER/9wtLZZ8meHqQvEYWY= github.com/hashicorp/go-retryablehttp v0.7.1 h1:sUiuQAnLlbvmExtFQs72iFW/HXeUn8Z1aJLQ4LJJbTQ= @@ -1674,8 +1677,9 @@ github.com/hashicorp/terraform-plugin-framework-timeouts v0.4.1 h1:gm5b1kHgFFhaK github.com/hashicorp/terraform-plugin-framework-timeouts v0.4.1/go.mod h1:MsjL1sQ9L7wGwzJ5RjcI6FzEMdyoBnw+XK8ZnOvQOLY= github.com/hashicorp/terraform-plugin-framework-validators v0.12.0 h1:HOjBuMbOEzl7snOdOoUfE2Jgeto6JOjLVQ39Ls2nksc= github.com/hashicorp/terraform-plugin-framework-validators v0.12.0/go.mod h1:jfHGE/gzjxYz6XoUwi/aYiiKrJDeutQNUtGQXkaHklg= -github.com/hashicorp/terraform-plugin-go v0.19.0 h1:BuZx/6Cp+lkmiG0cOBk6Zps0Cb2tmqQpDM3iAtnhDQU= github.com/hashicorp/terraform-plugin-go v0.19.0/go.mod h1:EhRSkEPNoylLQntYsk5KrDHTZJh9HQoumZXbOGOXmec= +github.com/hashicorp/terraform-plugin-go v0.19.1 h1:lf/jTGTeELcz5IIbn/94mJdmnTjRYm6S6ct/JqCSr50= +github.com/hashicorp/terraform-plugin-go v0.19.1/go.mod h1:5NMIS+DXkfacX6o5HCpswda5yjkSYfKzn1Nfl9l+qRs= github.com/hashicorp/terraform-plugin-log v0.9.0 h1:i7hOA+vdAItN1/7UrfBqBwvYPQ9TFvymaRGZED3FCV0= github.com/hashicorp/terraform-plugin-log v0.9.0/go.mod h1:rKL8egZQ/eXSyDqzLUuwUYLVdlYeamldAHSxjUFADow= github.com/hashicorp/terraform-plugin-mux v0.12.0 h1:TJlmeslQ11WlQtIFAfth0vXx+gSNgvMEng2Rn9z3WZY= @@ -1687,8 +1691,9 @@ github.com/hashicorp/terraform-plugin-test v1.2.0/go.mod h1:QIJHYz8j+xJtdtLrFTlz github.com/hashicorp/terraform-plugin-test/v2 v2.2.1/go.mod h1:eZ9JL3O69Cb71Skn6OhHyj17sLmHRb+H6VrDcJjKrYU= github.com/hashicorp/terraform-plugin-testing v1.5.1 h1:T4aQh9JAhmWo4+t1A7x+rnxAJHCDIYW9kXyo4sVO92c= github.com/hashicorp/terraform-plugin-testing v1.5.1/go.mod h1:dg8clO6K59rZ8w9EshBmDp1CxTIPu3yA4iaDpX1h5u0= -github.com/hashicorp/terraform-registry-address v0.2.2 h1:lPQBg403El8PPicg/qONZJDC6YlgCVbWDtNmmZKtBno= github.com/hashicorp/terraform-registry-address v0.2.2/go.mod h1:LtwNbCihUoUZ3RYriyS2wF/lGPB6gF9ICLRtuDk7hSo= +github.com/hashicorp/terraform-registry-address v0.2.3 h1:2TAiKJ1A3MAkZlH1YI/aTVcLZRu7JseiXNRHbOAyoTI= +github.com/hashicorp/terraform-registry-address v0.2.3/go.mod h1:lFHA76T8jfQteVfT7caREqguFrW3c4MFSPhZB7HHgUM= github.com/hashicorp/terraform-svchost v0.0.0-20191011084731-65d371908596/go.mod h1:kNDNcF7sN4DocDLBkQYz73HGKwN1ANB1blq4lIYLYvg= github.com/hashicorp/terraform-svchost v0.0.0-20200729002733-f050f53b9734/go.mod h1:kNDNcF7sN4DocDLBkQYz73HGKwN1ANB1blq4lIYLYvg= github.com/hashicorp/terraform-svchost v0.1.1 h1:EZZimZ1GxdqFRinZ1tpJwVxxt49xc/S52uzrw4x0jKQ= @@ -1995,8 +2000,8 @@ github.com/modern-go/reflect2 v1.0.2/go.mod h1:yWuevngMOJpCy52FWWMvUC8ws7m/LJsjY github.com/modocache/gover v0.0.0-20171022184752-b58185e213c5/go.mod h1:caMODM3PzxT8aQXRPkAt8xlV/e7d7w8GM5g0fa5F0D8= github.com/mongodb-forks/digest v1.0.5 h1:EJu3wtLZcA0HCvsZpX5yuD193/sW9tHiNvrEM5apXMk= github.com/mongodb-forks/digest v1.0.5/go.mod h1:rb+EX8zotClD5Dj4NdgxnJXG9nwrlx3NWKJ8xttz1Dg= -github.com/mongodb/terraform-provider-mongodbatlas v1.13.0 h1:DxKthBy8YqV1s6BZ7lTiUPEEyn2oSVC9CLo+dFOnWGU= -github.com/mongodb/terraform-provider-mongodbatlas v1.13.0/go.mod h1:qMDkn7lbmkuhg8TS4V97TI4/yobGtgqG0X6kfuk6gBI= +github.com/mongodb/terraform-provider-mongodbatlas v1.13.1 h1:d8F1mKSCzUmfExfHvY8AyeDm42v3xg3HyKerzpHttp8= +github.com/mongodb/terraform-provider-mongodbatlas v1.13.1/go.mod h1:yag33qwEynvJRbsLiq8v5wKjbmxP0/tQJsDEDh4pPXY= github.com/montanaflynn/stats v0.0.0-20171201202039-1bf9dbcd8cbe/go.mod h1:wL8QJuTMNUDYhXwkmfOly8iTdp5TEcJFWZD2D7SIkUc= github.com/montanaflynn/stats v0.6.6/go.mod h1:etXPPgVO6n31NxCd9KQUMvCM+ve0ruNzt6R8Bnaayow= github.com/morikuni/aec v1.0.0/go.mod h1:BbKIizmSmc5MMPqRYbxO4ZU0S0+P200+tUnFx7PXmsc= @@ -2410,8 +2415,9 @@ github.com/vmihailenco/msgpack v4.0.1+incompatible/go.mod h1:fy3FlTQTDXWkZ7Bh6Ac github.com/vmihailenco/msgpack v4.0.4+incompatible h1:dSLoQfGFAo3F6OoNhwUmLwVgaUXK79GlxNBwueZn0xI= github.com/vmihailenco/msgpack v4.0.4+incompatible/go.mod h1:fy3FlTQTDXWkZ7Bh6AcGMlsjHatGryHQYUTf1ShIgkk= github.com/vmihailenco/msgpack/v4 v4.3.12/go.mod h1:gborTTJjAo/GWTqqRjrLCn9pgNN+NXzzngzBKDPIqw4= -github.com/vmihailenco/msgpack/v5 v5.3.5 h1:5gO0H1iULLWGhs2H5tbAHIZTV8/cYafcFOr9znI5mJU= github.com/vmihailenco/msgpack/v5 v5.3.5/go.mod h1:7xyJ9e+0+9SaZT0Wt1RGleJXzli6Q/V5KbhBonMG9jc= +github.com/vmihailenco/msgpack/v5 v5.4.1 h1:cQriyiUvjTwOHg8QZaPihLWeRAAVoCpE00IUPn0Bjt8= +github.com/vmihailenco/msgpack/v5 v5.4.1/go.mod h1:GaZTsDaehaPpQVyxrf5mtQlH+pc21PIudVV/E3rRQok= github.com/vmihailenco/tagparser v0.1.1/go.mod h1:OeAg3pn3UbLjkWt+rN9oFYB6u/cQgqMEUPoW2WPyhdI= github.com/vmihailenco/tagparser/v2 v2.0.0 h1:y09buUbR+b5aycVFQs/g70pqKVZNBmxwAhO7/IwNM9g= github.com/vmihailenco/tagparser/v2 v2.0.0/go.mod h1:Wri+At7QHww0WTrCBeu4J6bNtoV6mEfg5OIWRZA9qds= @@ -2819,8 +2825,8 @@ golang.org/x/oauth2 v0.4.0/go.mod h1:RznEsdpjGAINPTOF0UH/t+xJ75L18YO3Ho6Pyn+uRec golang.org/x/oauth2 v0.5.0/go.mod h1:9/XBHVqLaWO3/BRHs5jbpYCnOZVjj5V0ndyaAM7KB4I= golang.org/x/oauth2 v0.6.0/go.mod h1:ycmewcwgD4Rpr3eZJLSB4Kyyljb3qDh40vJ8STE5HKw= golang.org/x/oauth2 v0.7.0/go.mod h1:hPLQkd9LyjfXTiRohC/41GhcFqxisoUQ99sCUOHO9x4= -golang.org/x/oauth2 v0.8.0 h1:6dkIjl3j3LtZ/O3sTgZTMsLKSftL/B8Zgq4huOIIUu8= -golang.org/x/oauth2 v0.8.0/go.mod h1:yr7u4HXZRm1R1kBWqr/xKNqewf0plRYoB7sla+BCIXE= +golang.org/x/oauth2 v0.11.0 h1:vPL4xzxBM4niKCW6g9whtaWVXTJf1U5e4aZxxFx/gbU= +golang.org/x/oauth2 v0.11.0/go.mod h1:LdF7O/8bLR/qWK9DrpXmbHLTouvRHK0SgJl0GmDBchk= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181221193216-37e7f081c4d4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -3398,16 +3404,16 @@ google.golang.org/genproto v0.0.0-20230330154414-c0448cd141ea/go.mod h1:UUQDJDOl google.golang.org/genproto v0.0.0-20230331144136-dcfb400f0633/go.mod h1:UUQDJDOlWu4KYeJZffbWgBkS1YFobzKbLVfK69pe0Ak= google.golang.org/genproto v0.0.0-20230525234025-438c736192d0/go.mod h1:9ExIQyXL5hZrHzQceCwuSYwZZ5QZBazOcprJ5rgs3lY= google.golang.org/genproto v0.0.0-20230526161137-0005af68ea54/go.mod h1:zqTuNwFlFRsw5zIts5VnzLQxSRqh+CGOTVMlYbY0Eyk= -google.golang.org/genproto v0.0.0-20230726155614-23370e0ffb3e h1:xIXmWJ303kJCuogpj0bHq+dcjcZHU+XFyc1I0Yl9cRg= -google.golang.org/genproto v0.0.0-20230726155614-23370e0ffb3e/go.mod h1:0ggbjUrZYpy1q+ANUS30SEoGZ53cdfwtbuG7Ptgy108= +google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= +google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= google.golang.org/genproto/googleapis/api v0.0.0-20230525234020-1aefcd67740a/go.mod h1:ts19tUU+Z0ZShN1y3aPyq2+O3d5FUNNgT6FtOzmrNn8= google.golang.org/genproto/googleapis/api v0.0.0-20230525234035-dd9d682886f9/go.mod h1:vHYtlOoi6TsQ3Uk2yxR7NI5z8uoV+3pZtR4jmHIkRig= -google.golang.org/genproto/googleapis/api v0.0.0-20230706204954-ccb25ca9f130 h1:XVeBY8d/FaK4848myy41HBqnDwvxeV3zMZhwN1TvAMU= -google.golang.org/genproto/googleapis/api v0.0.0-20230706204954-ccb25ca9f130/go.mod h1:mPBs5jNgx2GuQGvFwUvVKqtn6HsUw9nP64BedgvqEsQ= +google.golang.org/genproto/googleapis/api v0.0.0-20230822172742-b8732ec3820d h1:DoPTO70H+bcDXcd39vOqb2viZxgqeBeSGtZ55yZU4/Q= +google.golang.org/genproto/googleapis/api v0.0.0-20230822172742-b8732ec3820d/go.mod h1:KjSP20unUpOx5kyQUFa7k4OJg0qeJ7DEZflGDu2p6Bk= google.golang.org/genproto/googleapis/rpc v0.0.0-20230525234015-3fc162c6f38a/go.mod h1:xURIpW9ES5+/GZhnV6beoEtxQrnkRGIfP5VQG2tCBLc= google.golang.org/genproto/googleapis/rpc v0.0.0-20230525234030-28d5490b6b19/go.mod h1:66JfowdXAEgad5O9NnYcsNPLCPZJD++2L9X0PCMODrA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20230706204954-ccb25ca9f130 h1:2FZP5XuJY9zQyGM5N0rtovnoXjiMUEIUMvw0m9wlpLc= -google.golang.org/genproto/googleapis/rpc v0.0.0-20230706204954-ccb25ca9f130/go.mod h1:8mL13HKkDa+IuJ8yruA3ci0q+0vsUz4m//+ottjwS5o= +google.golang.org/genproto/googleapis/rpc v0.0.0-20230822172742-b8732ec3820d h1:uvYuEyMHKNt+lT4K3bN6fGswmK8qSvcreM3BwjDh+y4= +google.golang.org/genproto/googleapis/rpc v0.0.0-20230822172742-b8732ec3820d/go.mod h1:+Bk1OCOj40wS2hwAMA+aCW9ypzm63QTBBHp6lQ3p+9M= google.golang.org/grpc v0.0.0-20160317175043-d3ddb4469d5a/go.mod h1:yo6s7OP7yaDglbqo1J04qKzAhqBH6lvTonzMVmEdcZw= google.golang.org/grpc v1.8.0/go.mod h1:yo6s7OP7yaDglbqo1J04qKzAhqBH6lvTonzMVmEdcZw= google.golang.org/grpc v1.14.0/go.mod h1:yo6s7OP7yaDglbqo1J04qKzAhqBH6lvTonzMVmEdcZw= @@ -3461,8 +3467,8 @@ google.golang.org/grpc v1.52.3/go.mod h1:pu6fVzoFb+NBYNAvQL08ic+lvB2IojljRYuun5v google.golang.org/grpc v1.53.0/go.mod h1:OnIrk0ipVdj4N5d9IUoFUx72/VlD7+jUsHwZgwSMQpw= google.golang.org/grpc v1.54.0/go.mod h1:PUSEXI6iWghWaB6lXM4knEgpJNu2qUcKfDtNci3EC2g= google.golang.org/grpc v1.57.0/go.mod h1:Sd+9RMTACXwmub0zcNY2c4arhtrbBYD1AUHI/dt16Mo= -google.golang.org/grpc v1.57.1 h1:upNTNqv0ES+2ZOOqACwVtS3Il8M12/+Hz41RCPzAjQg= -google.golang.org/grpc v1.57.1/go.mod h1:Sd+9RMTACXwmub0zcNY2c4arhtrbBYD1AUHI/dt16Mo= +google.golang.org/grpc v1.59.0 h1:Z5Iec2pjwb+LEOqzpB2MR12/eKFhDPhuqW91O+4bwUk= +google.golang.org/grpc v1.59.0/go.mod h1:aUPDwccQo6OTjy7Hct4AfBPD1GptF4fyUjIkQ9YtF98= google.golang.org/grpc/cmd/protoc-gen-go-grpc v1.1.0/go.mod h1:6Kw0yEErY5E/yWrBtf03jp27GLLJujG4z/JK95pnjjw= google.golang.org/protobuf v0.0.0-20200109180630-ec00e32a8dfd/go.mod h1:DFci5gLYBciE7Vtevhsrf46CRTquxDuWsQurQQe4oz8= google.golang.org/protobuf v0.0.0-20200221191635-4d8936d0db64/go.mod h1:kwYJMbMJ01Woi6D6+Kah6886xMZcty6N08ah7+eCXa0= diff --git a/sdk/dotnet/Cluster.cs b/sdk/dotnet/Cluster.cs index 2dabb844..e64bed3c 100644 --- a/sdk/dotnet/Cluster.cs +++ b/sdk/dotnet/Cluster.cs @@ -26,7 +26,7 @@ public partial class Cluster : global::Pulumi.CustomResource /// If reconfiguration is necessary to regain a primary due to a regional outage, submit this field alongside your topology reconfiguration to request a new regional outage resistant topology. Forced reconfigurations during an outage of the majority of electable nodes carry a risk of data loss if replicated writes (even majority committed writes) have not been replicated to the new primary node. MongoDB Atlas docs contain more information. To proceed with an operation which carries that risk, set `accept_data_risks_and_force_replica_set_reconfig` to the current date. Learn more about Reconfiguring a Replica Set during a regional outage [here](https://dochub.mongodb.org/core/regional-outage-reconfigure-replica-set). /// [Output("acceptDataRisksAndForceReplicaSetReconfig")] - public Output AcceptDataRisksAndForceReplicaSetReconfig { get; private set; } = null!; + public Output AcceptDataRisksAndForceReplicaSetReconfig { get; private set; } = null!; [Output("advancedConfiguration")] public Output AdvancedConfiguration { get; private set; } = null!; @@ -42,7 +42,7 @@ public partial class Cluster : global::Pulumi.CustomResource public Output AutoScalingComputeScaleDownEnabled { get; private set; } = null!; [Output("autoScalingDiskGbEnabled")] - public Output AutoScalingDiskGbEnabled { get; private set; } = null!; + public Output AutoScalingDiskGbEnabled { get; private set; } = null!; /// /// Cloud service provider on which the server for a multi-tenant cluster is provisioned. diff --git a/sdk/dotnet/GetOnlineArchive.cs b/sdk/dotnet/GetOnlineArchive.cs index 3e8585aa..ef95ab7d 100644 --- a/sdk/dotnet/GetOnlineArchive.cs +++ b/sdk/dotnet/GetOnlineArchive.cs @@ -47,6 +47,7 @@ public static class GetOnlineArchive /// * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". /// * `criteria` - Criteria to use for archiving data. See criteria. /// * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + /// * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. /// * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. /// * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. /// * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -62,6 +63,10 @@ public static class GetOnlineArchive /// ### Data Expiration Rule /// * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. /// + /// ### Data Process Region + /// * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + /// * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + /// /// /// ### Schedule /// @@ -119,6 +124,7 @@ public static Task InvokeAsync(GetOnlineArchiveArgs args /// * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". /// * `criteria` - Criteria to use for archiving data. See criteria. /// * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + /// * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. /// * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. /// * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. /// * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -134,6 +140,10 @@ public static Task InvokeAsync(GetOnlineArchiveArgs args /// ### Data Expiration Rule /// * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. /// + /// ### Data Process Region + /// * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + /// * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + /// /// /// ### Schedule /// @@ -219,6 +229,7 @@ public sealed class GetOnlineArchiveResult public readonly string CollectionType; public readonly ImmutableArray Criterias; public readonly ImmutableArray DataExpirationRules; + public readonly ImmutableArray DataProcessRegions; public readonly string DbName; /// /// The provider-assigned unique ID for this managed resource. @@ -244,6 +255,8 @@ private GetOnlineArchiveResult( ImmutableArray dataExpirationRules, + ImmutableArray dataProcessRegions, + string dbName, string id, @@ -264,6 +277,7 @@ private GetOnlineArchiveResult( CollectionType = collectionType; Criterias = criterias; DataExpirationRules = dataExpirationRules; + DataProcessRegions = dataProcessRegions; DbName = dbName; Id = id; PartitionFields = partitionFields; diff --git a/sdk/dotnet/GetOnlineArchives.cs b/sdk/dotnet/GetOnlineArchives.cs index 609a2b63..a398afbe 100644 --- a/sdk/dotnet/GetOnlineArchives.cs +++ b/sdk/dotnet/GetOnlineArchives.cs @@ -46,6 +46,7 @@ public static class GetOnlineArchives /// * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". /// * `criteria` - Criteria to use for archiving data. See criteria. /// * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + /// * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. /// * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. /// * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. /// * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -61,6 +62,10 @@ public static class GetOnlineArchives /// ### Data Expiration Rule /// * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. /// + /// ### Data Process Region + /// * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + /// * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + /// /// ### Schedule /// /// * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). @@ -114,6 +119,7 @@ public static Task InvokeAsync(GetOnlineArchivesArgs ar /// * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". /// * `criteria` - Criteria to use for archiving data. See criteria. /// * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + /// * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. /// * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. /// * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. /// * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -129,6 +135,10 @@ public static Task InvokeAsync(GetOnlineArchivesArgs ar /// ### Data Expiration Rule /// * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. /// + /// ### Data Process Region + /// * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + /// * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + /// /// ### Schedule /// /// * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). diff --git a/sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionArgs.cs b/sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionArgs.cs new file mode 100644 index 00000000..0a3080eb --- /dev/null +++ b/sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionArgs.cs @@ -0,0 +1,32 @@ +// *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +using System; +using System.Collections.Generic; +using System.Collections.Immutable; +using System.Threading.Tasks; +using Pulumi.Serialization; + +namespace Pulumi.Mongodbatlas.Inputs +{ + + public sealed class OnlineArchiveDataProcessRegionArgs : global::Pulumi.ResourceArgs + { + /// + /// Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + /// + [Input("cloudProvider")] + public Input? CloudProvider { get; set; } + + /// + /// Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + /// + [Input("region")] + public Input? Region { get; set; } + + public OnlineArchiveDataProcessRegionArgs() + { + } + public static new OnlineArchiveDataProcessRegionArgs Empty => new OnlineArchiveDataProcessRegionArgs(); + } +} diff --git a/sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionGetArgs.cs b/sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionGetArgs.cs new file mode 100644 index 00000000..c64a13b8 --- /dev/null +++ b/sdk/dotnet/Inputs/OnlineArchiveDataProcessRegionGetArgs.cs @@ -0,0 +1,32 @@ +// *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +using System; +using System.Collections.Generic; +using System.Collections.Immutable; +using System.Threading.Tasks; +using Pulumi.Serialization; + +namespace Pulumi.Mongodbatlas.Inputs +{ + + public sealed class OnlineArchiveDataProcessRegionGetArgs : global::Pulumi.ResourceArgs + { + /// + /// Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + /// + [Input("cloudProvider")] + public Input? CloudProvider { get; set; } + + /// + /// Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + /// + [Input("region")] + public Input? Region { get; set; } + + public OnlineArchiveDataProcessRegionGetArgs() + { + } + public static new OnlineArchiveDataProcessRegionGetArgs Empty => new OnlineArchiveDataProcessRegionGetArgs(); + } +} diff --git a/sdk/dotnet/OnlineArchive.cs b/sdk/dotnet/OnlineArchive.cs index 6fb1d5bf..56dd8f61 100644 --- a/sdk/dotnet/OnlineArchive.cs +++ b/sdk/dotnet/OnlineArchive.cs @@ -144,6 +144,12 @@ public partial class OnlineArchive : global::Pulumi.CustomResource [Output("dataExpirationRule")] public Output DataExpirationRule { get; private set; } = null!; + /// + /// Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + /// + [Output("dataProcessRegion")] + public Output DataProcessRegion { get; private set; } = null!; + /// /// Name of the database that contains the collection. /// @@ -259,6 +265,12 @@ public sealed class OnlineArchiveArgs : global::Pulumi.ResourceArgs [Input("dataExpirationRule")] public Input? DataExpirationRule { get; set; } + /// + /// Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + /// + [Input("dataProcessRegion")] + public Input? DataProcessRegion { get; set; } + /// /// Name of the database that contains the collection. /// @@ -342,6 +354,12 @@ public sealed class OnlineArchiveState : global::Pulumi.ResourceArgs [Input("dataExpirationRule")] public Input? DataExpirationRule { get; set; } + /// + /// Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + /// + [Input("dataProcessRegion")] + public Input? DataProcessRegion { get; set; } + /// /// Name of the database that contains the collection. /// diff --git a/sdk/dotnet/Outputs/GetOnlineArchiveDataProcessRegionResult.cs b/sdk/dotnet/Outputs/GetOnlineArchiveDataProcessRegionResult.cs new file mode 100644 index 00000000..23084c2d --- /dev/null +++ b/sdk/dotnet/Outputs/GetOnlineArchiveDataProcessRegionResult.cs @@ -0,0 +1,29 @@ +// *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +using System; +using System.Collections.Generic; +using System.Collections.Immutable; +using System.Threading.Tasks; +using Pulumi.Serialization; + +namespace Pulumi.Mongodbatlas.Outputs +{ + + [OutputType] + public sealed class GetOnlineArchiveDataProcessRegionResult + { + public readonly string CloudProvider; + public readonly string Region; + + [OutputConstructor] + private GetOnlineArchiveDataProcessRegionResult( + string cloudProvider, + + string region) + { + CloudProvider = cloudProvider; + Region = region; + } + } +} diff --git a/sdk/dotnet/Outputs/GetOnlineArchivesResultDataProcessRegionResult.cs b/sdk/dotnet/Outputs/GetOnlineArchivesResultDataProcessRegionResult.cs new file mode 100644 index 00000000..faeb61ba --- /dev/null +++ b/sdk/dotnet/Outputs/GetOnlineArchivesResultDataProcessRegionResult.cs @@ -0,0 +1,29 @@ +// *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +using System; +using System.Collections.Generic; +using System.Collections.Immutable; +using System.Threading.Tasks; +using Pulumi.Serialization; + +namespace Pulumi.Mongodbatlas.Outputs +{ + + [OutputType] + public sealed class GetOnlineArchivesResultDataProcessRegionResult + { + public readonly string CloudProvider; + public readonly string Region; + + [OutputConstructor] + private GetOnlineArchivesResultDataProcessRegionResult( + string cloudProvider, + + string region) + { + CloudProvider = cloudProvider; + Region = region; + } + } +} diff --git a/sdk/dotnet/Outputs/GetOnlineArchivesResultResult.cs b/sdk/dotnet/Outputs/GetOnlineArchivesResultResult.cs index b6502259..ee331e9e 100644 --- a/sdk/dotnet/Outputs/GetOnlineArchivesResultResult.cs +++ b/sdk/dotnet/Outputs/GetOnlineArchivesResultResult.cs @@ -26,6 +26,7 @@ public sealed class GetOnlineArchivesResultResult public readonly string CollectionType; public readonly ImmutableArray Criterias; public readonly ImmutableArray DataExpirationRules; + public readonly ImmutableArray DataProcessRegions; public readonly string DbName; public readonly ImmutableArray PartitionFields; public readonly bool Paused; @@ -50,6 +51,8 @@ private GetOnlineArchivesResultResult( ImmutableArray dataExpirationRules, + ImmutableArray dataProcessRegions, + string dbName, ImmutableArray partitionFields, @@ -68,6 +71,7 @@ private GetOnlineArchivesResultResult( CollectionType = collectionType; Criterias = criterias; DataExpirationRules = dataExpirationRules; + DataProcessRegions = dataProcessRegions; DbName = dbName; PartitionFields = partitionFields; Paused = paused; diff --git a/sdk/dotnet/Outputs/OnlineArchiveDataProcessRegion.cs b/sdk/dotnet/Outputs/OnlineArchiveDataProcessRegion.cs new file mode 100644 index 00000000..efd592f7 --- /dev/null +++ b/sdk/dotnet/Outputs/OnlineArchiveDataProcessRegion.cs @@ -0,0 +1,35 @@ +// *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +using System; +using System.Collections.Generic; +using System.Collections.Immutable; +using System.Threading.Tasks; +using Pulumi.Serialization; + +namespace Pulumi.Mongodbatlas.Outputs +{ + + [OutputType] + public sealed class OnlineArchiveDataProcessRegion + { + /// + /// Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + /// + public readonly string? CloudProvider; + /// + /// Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + /// + public readonly string? Region; + + [OutputConstructor] + private OnlineArchiveDataProcessRegion( + string? cloudProvider, + + string? region) + { + CloudProvider = cloudProvider; + Region = region; + } + } +} diff --git a/sdk/dotnet/SearchDeployment.cs b/sdk/dotnet/SearchDeployment.cs index 47e7fd1a..fa79758f 100644 --- a/sdk/dotnet/SearchDeployment.cs +++ b/sdk/dotnet/SearchDeployment.cs @@ -13,7 +13,8 @@ namespace Pulumi.Mongodbatlas /// `mongodbatlas.SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster. /// /// > **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation). - /// **NOTE:** Only a single search deployment resource can be defined for each cluster. + /// + /// > **NOTE:** Only a single search deployment resource can be defined for each cluster. /// /// ## Import /// diff --git a/sdk/go/mongodbatlas/cluster.go b/sdk/go/mongodbatlas/cluster.go index 61e279d2..e5edafdf 100644 --- a/sdk/go/mongodbatlas/cluster.go +++ b/sdk/go/mongodbatlas/cluster.go @@ -27,13 +27,13 @@ type Cluster struct { pulumi.CustomResourceState // If reconfiguration is necessary to regain a primary due to a regional outage, submit this field alongside your topology reconfiguration to request a new regional outage resistant topology. Forced reconfigurations during an outage of the majority of electable nodes carry a risk of data loss if replicated writes (even majority committed writes) have not been replicated to the new primary node. MongoDB Atlas docs contain more information. To proceed with an operation which carries that risk, set `acceptDataRisksAndForceReplicaSetReconfig` to the current date. Learn more about Reconfiguring a Replica Set during a regional outage [here](https://dochub.mongodb.org/core/regional-outage-reconfigure-replica-set). - AcceptDataRisksAndForceReplicaSetReconfig pulumi.StringPtrOutput `pulumi:"acceptDataRisksAndForceReplicaSetReconfig"` + AcceptDataRisksAndForceReplicaSetReconfig pulumi.StringOutput `pulumi:"acceptDataRisksAndForceReplicaSetReconfig"` AdvancedConfiguration ClusterAdvancedConfigurationOutput `pulumi:"advancedConfiguration"` AutoScalingComputeEnabled pulumi.BoolOutput `pulumi:"autoScalingComputeEnabled"` // Set to `true` to enable the cluster tier to scale down. This option is only available if `autoScaling.compute.enabled` is `true`. // - If this option is enabled, you must specify a value for `providerSettings.autoScaling.compute.minInstanceSize` - AutoScalingComputeScaleDownEnabled pulumi.BoolOutput `pulumi:"autoScalingComputeScaleDownEnabled"` - AutoScalingDiskGbEnabled pulumi.BoolPtrOutput `pulumi:"autoScalingDiskGbEnabled"` + AutoScalingComputeScaleDownEnabled pulumi.BoolOutput `pulumi:"autoScalingComputeScaleDownEnabled"` + AutoScalingDiskGbEnabled pulumi.BoolOutput `pulumi:"autoScalingDiskGbEnabled"` // Cloud service provider on which the server for a multi-tenant cluster is provisioned. // // This setting is only valid when providerSetting.providerName is TENANT and providerSetting.instanceSizeName is M2 or M5. @@ -787,8 +787,8 @@ func (o ClusterOutput) ToClusterOutputWithContext(ctx context.Context) ClusterOu } // If reconfiguration is necessary to regain a primary due to a regional outage, submit this field alongside your topology reconfiguration to request a new regional outage resistant topology. Forced reconfigurations during an outage of the majority of electable nodes carry a risk of data loss if replicated writes (even majority committed writes) have not been replicated to the new primary node. MongoDB Atlas docs contain more information. To proceed with an operation which carries that risk, set `acceptDataRisksAndForceReplicaSetReconfig` to the current date. Learn more about Reconfiguring a Replica Set during a regional outage [here](https://dochub.mongodb.org/core/regional-outage-reconfigure-replica-set). -func (o ClusterOutput) AcceptDataRisksAndForceReplicaSetReconfig() pulumi.StringPtrOutput { - return o.ApplyT(func(v *Cluster) pulumi.StringPtrOutput { return v.AcceptDataRisksAndForceReplicaSetReconfig }).(pulumi.StringPtrOutput) +func (o ClusterOutput) AcceptDataRisksAndForceReplicaSetReconfig() pulumi.StringOutput { + return o.ApplyT(func(v *Cluster) pulumi.StringOutput { return v.AcceptDataRisksAndForceReplicaSetReconfig }).(pulumi.StringOutput) } func (o ClusterOutput) AdvancedConfiguration() ClusterAdvancedConfigurationOutput { @@ -805,8 +805,8 @@ func (o ClusterOutput) AutoScalingComputeScaleDownEnabled() pulumi.BoolOutput { return o.ApplyT(func(v *Cluster) pulumi.BoolOutput { return v.AutoScalingComputeScaleDownEnabled }).(pulumi.BoolOutput) } -func (o ClusterOutput) AutoScalingDiskGbEnabled() pulumi.BoolPtrOutput { - return o.ApplyT(func(v *Cluster) pulumi.BoolPtrOutput { return v.AutoScalingDiskGbEnabled }).(pulumi.BoolPtrOutput) +func (o ClusterOutput) AutoScalingDiskGbEnabled() pulumi.BoolOutput { + return o.ApplyT(func(v *Cluster) pulumi.BoolOutput { return v.AutoScalingDiskGbEnabled }).(pulumi.BoolOutput) } // Cloud service provider on which the server for a multi-tenant cluster is provisioned. diff --git a/sdk/go/mongodbatlas/getOnlineArchive.go b/sdk/go/mongodbatlas/getOnlineArchive.go index a93fddf8..4824ed3c 100644 --- a/sdk/go/mongodbatlas/getOnlineArchive.go +++ b/sdk/go/mongodbatlas/getOnlineArchive.go @@ -49,6 +49,7 @@ import ( // * `collectionType` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". // * `criteria` - Criteria to use for archiving data. See criteria. // * `dataExpirationRule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. +// * `dataProcessRegion` - Settings to configure the region where you wish to store your archived data. See data process region. // * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. // * `partitionFields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. // * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -64,6 +65,10 @@ import ( // ### Data Expiration Rule // * `expireAfterDays` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. // +// ### Data Process Region +// * `cloudProvider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. +// * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) +// // ### Schedule // // * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. @@ -108,6 +113,7 @@ type LookupOnlineArchiveResult struct { CollectionType string `pulumi:"collectionType"` Criterias []GetOnlineArchiveCriteria `pulumi:"criterias"` DataExpirationRules []GetOnlineArchiveDataExpirationRule `pulumi:"dataExpirationRules"` + DataProcessRegions []GetOnlineArchiveDataProcessRegion `pulumi:"dataProcessRegions"` DbName string `pulumi:"dbName"` // The provider-assigned unique ID for this managed resource. Id string `pulumi:"id"` @@ -184,6 +190,10 @@ func (o LookupOnlineArchiveResultOutput) DataExpirationRules() GetOnlineArchiveD return o.ApplyT(func(v LookupOnlineArchiveResult) []GetOnlineArchiveDataExpirationRule { return v.DataExpirationRules }).(GetOnlineArchiveDataExpirationRuleArrayOutput) } +func (o LookupOnlineArchiveResultOutput) DataProcessRegions() GetOnlineArchiveDataProcessRegionArrayOutput { + return o.ApplyT(func(v LookupOnlineArchiveResult) []GetOnlineArchiveDataProcessRegion { return v.DataProcessRegions }).(GetOnlineArchiveDataProcessRegionArrayOutput) +} + func (o LookupOnlineArchiveResultOutput) DbName() pulumi.StringOutput { return o.ApplyT(func(v LookupOnlineArchiveResult) string { return v.DbName }).(pulumi.StringOutput) } diff --git a/sdk/go/mongodbatlas/getOnlineArchives.go b/sdk/go/mongodbatlas/getOnlineArchives.go index cf985242..cf99b4f0 100644 --- a/sdk/go/mongodbatlas/getOnlineArchives.go +++ b/sdk/go/mongodbatlas/getOnlineArchives.go @@ -48,6 +48,7 @@ import ( // * `collectionType` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". // * `criteria` - Criteria to use for archiving data. See criteria. // * `dataExpirationRule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. +// * `dataProcessRegion` - Settings to configure the region where you wish to store your archived data. See data process region. // * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. // * `partitionFields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. // * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -63,6 +64,10 @@ import ( // ### Data Expiration Rule // * `expireAfterDays` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. // +// ### Data Process Region +// * `cloudProvider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. +// * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) +// // ### Schedule // // * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). diff --git a/sdk/go/mongodbatlas/onlineArchive.go b/sdk/go/mongodbatlas/onlineArchive.go index 8e2ff818..57009a51 100644 --- a/sdk/go/mongodbatlas/onlineArchive.go +++ b/sdk/go/mongodbatlas/onlineArchive.go @@ -128,6 +128,8 @@ type OnlineArchive struct { Criteria OnlineArchiveCriteriaOutput `pulumi:"criteria"` // Rule for specifying when data should be deleted from the archive. See data expiration rule. DataExpirationRule OnlineArchiveDataExpirationRulePtrOutput `pulumi:"dataExpirationRule"` + // Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + DataProcessRegion OnlineArchiveDataProcessRegionOutput `pulumi:"dataProcessRegion"` // Name of the database that contains the collection. DbName pulumi.StringOutput `pulumi:"dbName"` // Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. @@ -200,6 +202,8 @@ type onlineArchiveState struct { Criteria *OnlineArchiveCriteria `pulumi:"criteria"` // Rule for specifying when data should be deleted from the archive. See data expiration rule. DataExpirationRule *OnlineArchiveDataExpirationRule `pulumi:"dataExpirationRule"` + // Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + DataProcessRegion *OnlineArchiveDataProcessRegion `pulumi:"dataProcessRegion"` // Name of the database that contains the collection. DbName *string `pulumi:"dbName"` // Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. @@ -228,6 +232,8 @@ type OnlineArchiveState struct { Criteria OnlineArchiveCriteriaPtrInput // Rule for specifying when data should be deleted from the archive. See data expiration rule. DataExpirationRule OnlineArchiveDataExpirationRulePtrInput + // Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + DataProcessRegion OnlineArchiveDataProcessRegionPtrInput // Name of the database that contains the collection. DbName pulumi.StringPtrInput // Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. @@ -258,6 +264,8 @@ type onlineArchiveArgs struct { Criteria OnlineArchiveCriteria `pulumi:"criteria"` // Rule for specifying when data should be deleted from the archive. See data expiration rule. DataExpirationRule *OnlineArchiveDataExpirationRule `pulumi:"dataExpirationRule"` + // Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + DataProcessRegion *OnlineArchiveDataProcessRegion `pulumi:"dataProcessRegion"` // Name of the database that contains the collection. DbName string `pulumi:"dbName"` // Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. @@ -283,6 +291,8 @@ type OnlineArchiveArgs struct { Criteria OnlineArchiveCriteriaInput // Rule for specifying when data should be deleted from the archive. See data expiration rule. DataExpirationRule OnlineArchiveDataExpirationRulePtrInput + // Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + DataProcessRegion OnlineArchiveDataProcessRegionPtrInput // Name of the database that contains the collection. DbName pulumi.StringInput // Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. @@ -413,6 +423,11 @@ func (o OnlineArchiveOutput) DataExpirationRule() OnlineArchiveDataExpirationRul return o.ApplyT(func(v *OnlineArchive) OnlineArchiveDataExpirationRulePtrOutput { return v.DataExpirationRule }).(OnlineArchiveDataExpirationRulePtrOutput) } +// Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. +func (o OnlineArchiveOutput) DataProcessRegion() OnlineArchiveDataProcessRegionOutput { + return o.ApplyT(func(v *OnlineArchive) OnlineArchiveDataProcessRegionOutput { return v.DataProcessRegion }).(OnlineArchiveDataProcessRegionOutput) +} + // Name of the database that contains the collection. func (o OnlineArchiveOutput) DbName() pulumi.StringOutput { return o.ApplyT(func(v *OnlineArchive) pulumi.StringOutput { return v.DbName }).(pulumi.StringOutput) diff --git a/sdk/go/mongodbatlas/pulumiTypes.go b/sdk/go/mongodbatlas/pulumiTypes.go index 7387cdb3..48f8117d 100644 --- a/sdk/go/mongodbatlas/pulumiTypes.go +++ b/sdk/go/mongodbatlas/pulumiTypes.go @@ -13625,6 +13625,162 @@ func (o OnlineArchiveDataExpirationRulePtrOutput) ExpireAfterDays() pulumi.IntPt }).(pulumi.IntPtrOutput) } +type OnlineArchiveDataProcessRegion struct { + // Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + CloudProvider *string `pulumi:"cloudProvider"` + // Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + Region *string `pulumi:"region"` +} + +// OnlineArchiveDataProcessRegionInput is an input type that accepts OnlineArchiveDataProcessRegionArgs and OnlineArchiveDataProcessRegionOutput values. +// You can construct a concrete instance of `OnlineArchiveDataProcessRegionInput` via: +// +// OnlineArchiveDataProcessRegionArgs{...} +type OnlineArchiveDataProcessRegionInput interface { + pulumi.Input + + ToOnlineArchiveDataProcessRegionOutput() OnlineArchiveDataProcessRegionOutput + ToOnlineArchiveDataProcessRegionOutputWithContext(context.Context) OnlineArchiveDataProcessRegionOutput +} + +type OnlineArchiveDataProcessRegionArgs struct { + // Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + CloudProvider pulumi.StringPtrInput `pulumi:"cloudProvider"` + // Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + Region pulumi.StringPtrInput `pulumi:"region"` +} + +func (OnlineArchiveDataProcessRegionArgs) ElementType() reflect.Type { + return reflect.TypeOf((*OnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (i OnlineArchiveDataProcessRegionArgs) ToOnlineArchiveDataProcessRegionOutput() OnlineArchiveDataProcessRegionOutput { + return i.ToOnlineArchiveDataProcessRegionOutputWithContext(context.Background()) +} + +func (i OnlineArchiveDataProcessRegionArgs) ToOnlineArchiveDataProcessRegionOutputWithContext(ctx context.Context) OnlineArchiveDataProcessRegionOutput { + return pulumi.ToOutputWithContext(ctx, i).(OnlineArchiveDataProcessRegionOutput) +} + +func (i OnlineArchiveDataProcessRegionArgs) ToOnlineArchiveDataProcessRegionPtrOutput() OnlineArchiveDataProcessRegionPtrOutput { + return i.ToOnlineArchiveDataProcessRegionPtrOutputWithContext(context.Background()) +} + +func (i OnlineArchiveDataProcessRegionArgs) ToOnlineArchiveDataProcessRegionPtrOutputWithContext(ctx context.Context) OnlineArchiveDataProcessRegionPtrOutput { + return pulumi.ToOutputWithContext(ctx, i).(OnlineArchiveDataProcessRegionOutput).ToOnlineArchiveDataProcessRegionPtrOutputWithContext(ctx) +} + +// OnlineArchiveDataProcessRegionPtrInput is an input type that accepts OnlineArchiveDataProcessRegionArgs, OnlineArchiveDataProcessRegionPtr and OnlineArchiveDataProcessRegionPtrOutput values. +// You can construct a concrete instance of `OnlineArchiveDataProcessRegionPtrInput` via: +// +// OnlineArchiveDataProcessRegionArgs{...} +// +// or: +// +// nil +type OnlineArchiveDataProcessRegionPtrInput interface { + pulumi.Input + + ToOnlineArchiveDataProcessRegionPtrOutput() OnlineArchiveDataProcessRegionPtrOutput + ToOnlineArchiveDataProcessRegionPtrOutputWithContext(context.Context) OnlineArchiveDataProcessRegionPtrOutput +} + +type onlineArchiveDataProcessRegionPtrType OnlineArchiveDataProcessRegionArgs + +func OnlineArchiveDataProcessRegionPtr(v *OnlineArchiveDataProcessRegionArgs) OnlineArchiveDataProcessRegionPtrInput { + return (*onlineArchiveDataProcessRegionPtrType)(v) +} + +func (*onlineArchiveDataProcessRegionPtrType) ElementType() reflect.Type { + return reflect.TypeOf((**OnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (i *onlineArchiveDataProcessRegionPtrType) ToOnlineArchiveDataProcessRegionPtrOutput() OnlineArchiveDataProcessRegionPtrOutput { + return i.ToOnlineArchiveDataProcessRegionPtrOutputWithContext(context.Background()) +} + +func (i *onlineArchiveDataProcessRegionPtrType) ToOnlineArchiveDataProcessRegionPtrOutputWithContext(ctx context.Context) OnlineArchiveDataProcessRegionPtrOutput { + return pulumi.ToOutputWithContext(ctx, i).(OnlineArchiveDataProcessRegionPtrOutput) +} + +type OnlineArchiveDataProcessRegionOutput struct{ *pulumi.OutputState } + +func (OnlineArchiveDataProcessRegionOutput) ElementType() reflect.Type { + return reflect.TypeOf((*OnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (o OnlineArchiveDataProcessRegionOutput) ToOnlineArchiveDataProcessRegionOutput() OnlineArchiveDataProcessRegionOutput { + return o +} + +func (o OnlineArchiveDataProcessRegionOutput) ToOnlineArchiveDataProcessRegionOutputWithContext(ctx context.Context) OnlineArchiveDataProcessRegionOutput { + return o +} + +func (o OnlineArchiveDataProcessRegionOutput) ToOnlineArchiveDataProcessRegionPtrOutput() OnlineArchiveDataProcessRegionPtrOutput { + return o.ToOnlineArchiveDataProcessRegionPtrOutputWithContext(context.Background()) +} + +func (o OnlineArchiveDataProcessRegionOutput) ToOnlineArchiveDataProcessRegionPtrOutputWithContext(ctx context.Context) OnlineArchiveDataProcessRegionPtrOutput { + return o.ApplyTWithContext(ctx, func(_ context.Context, v OnlineArchiveDataProcessRegion) *OnlineArchiveDataProcessRegion { + return &v + }).(OnlineArchiveDataProcessRegionPtrOutput) +} + +// Human-readable label that identifies the Cloud service provider where you wish to store your archived data. +func (o OnlineArchiveDataProcessRegionOutput) CloudProvider() pulumi.StringPtrOutput { + return o.ApplyT(func(v OnlineArchiveDataProcessRegion) *string { return v.CloudProvider }).(pulumi.StringPtrOutput) +} + +// Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) +func (o OnlineArchiveDataProcessRegionOutput) Region() pulumi.StringPtrOutput { + return o.ApplyT(func(v OnlineArchiveDataProcessRegion) *string { return v.Region }).(pulumi.StringPtrOutput) +} + +type OnlineArchiveDataProcessRegionPtrOutput struct{ *pulumi.OutputState } + +func (OnlineArchiveDataProcessRegionPtrOutput) ElementType() reflect.Type { + return reflect.TypeOf((**OnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (o OnlineArchiveDataProcessRegionPtrOutput) ToOnlineArchiveDataProcessRegionPtrOutput() OnlineArchiveDataProcessRegionPtrOutput { + return o +} + +func (o OnlineArchiveDataProcessRegionPtrOutput) ToOnlineArchiveDataProcessRegionPtrOutputWithContext(ctx context.Context) OnlineArchiveDataProcessRegionPtrOutput { + return o +} + +func (o OnlineArchiveDataProcessRegionPtrOutput) Elem() OnlineArchiveDataProcessRegionOutput { + return o.ApplyT(func(v *OnlineArchiveDataProcessRegion) OnlineArchiveDataProcessRegion { + if v != nil { + return *v + } + var ret OnlineArchiveDataProcessRegion + return ret + }).(OnlineArchiveDataProcessRegionOutput) +} + +// Human-readable label that identifies the Cloud service provider where you wish to store your archived data. +func (o OnlineArchiveDataProcessRegionPtrOutput) CloudProvider() pulumi.StringPtrOutput { + return o.ApplyT(func(v *OnlineArchiveDataProcessRegion) *string { + if v == nil { + return nil + } + return v.CloudProvider + }).(pulumi.StringPtrOutput) +} + +// Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) +func (o OnlineArchiveDataProcessRegionPtrOutput) Region() pulumi.StringPtrOutput { + return o.ApplyT(func(v *OnlineArchiveDataProcessRegion) *string { + if v == nil { + return nil + } + return v.Region + }).(pulumi.StringPtrOutput) +} + type OnlineArchivePartitionField struct { // Human-readable label that identifies the parameter that MongoDB Cloud uses to partition data. To specify a nested parameter, use the dot notation. FieldName string `pulumi:"fieldName"` @@ -41177,6 +41333,106 @@ func (o GetOnlineArchiveDataExpirationRuleArrayOutput) Index(i pulumi.IntInput) }).(GetOnlineArchiveDataExpirationRuleOutput) } +type GetOnlineArchiveDataProcessRegion struct { + CloudProvider string `pulumi:"cloudProvider"` + Region string `pulumi:"region"` +} + +// GetOnlineArchiveDataProcessRegionInput is an input type that accepts GetOnlineArchiveDataProcessRegionArgs and GetOnlineArchiveDataProcessRegionOutput values. +// You can construct a concrete instance of `GetOnlineArchiveDataProcessRegionInput` via: +// +// GetOnlineArchiveDataProcessRegionArgs{...} +type GetOnlineArchiveDataProcessRegionInput interface { + pulumi.Input + + ToGetOnlineArchiveDataProcessRegionOutput() GetOnlineArchiveDataProcessRegionOutput + ToGetOnlineArchiveDataProcessRegionOutputWithContext(context.Context) GetOnlineArchiveDataProcessRegionOutput +} + +type GetOnlineArchiveDataProcessRegionArgs struct { + CloudProvider pulumi.StringInput `pulumi:"cloudProvider"` + Region pulumi.StringInput `pulumi:"region"` +} + +func (GetOnlineArchiveDataProcessRegionArgs) ElementType() reflect.Type { + return reflect.TypeOf((*GetOnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (i GetOnlineArchiveDataProcessRegionArgs) ToGetOnlineArchiveDataProcessRegionOutput() GetOnlineArchiveDataProcessRegionOutput { + return i.ToGetOnlineArchiveDataProcessRegionOutputWithContext(context.Background()) +} + +func (i GetOnlineArchiveDataProcessRegionArgs) ToGetOnlineArchiveDataProcessRegionOutputWithContext(ctx context.Context) GetOnlineArchiveDataProcessRegionOutput { + return pulumi.ToOutputWithContext(ctx, i).(GetOnlineArchiveDataProcessRegionOutput) +} + +// GetOnlineArchiveDataProcessRegionArrayInput is an input type that accepts GetOnlineArchiveDataProcessRegionArray and GetOnlineArchiveDataProcessRegionArrayOutput values. +// You can construct a concrete instance of `GetOnlineArchiveDataProcessRegionArrayInput` via: +// +// GetOnlineArchiveDataProcessRegionArray{ GetOnlineArchiveDataProcessRegionArgs{...} } +type GetOnlineArchiveDataProcessRegionArrayInput interface { + pulumi.Input + + ToGetOnlineArchiveDataProcessRegionArrayOutput() GetOnlineArchiveDataProcessRegionArrayOutput + ToGetOnlineArchiveDataProcessRegionArrayOutputWithContext(context.Context) GetOnlineArchiveDataProcessRegionArrayOutput +} + +type GetOnlineArchiveDataProcessRegionArray []GetOnlineArchiveDataProcessRegionInput + +func (GetOnlineArchiveDataProcessRegionArray) ElementType() reflect.Type { + return reflect.TypeOf((*[]GetOnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (i GetOnlineArchiveDataProcessRegionArray) ToGetOnlineArchiveDataProcessRegionArrayOutput() GetOnlineArchiveDataProcessRegionArrayOutput { + return i.ToGetOnlineArchiveDataProcessRegionArrayOutputWithContext(context.Background()) +} + +func (i GetOnlineArchiveDataProcessRegionArray) ToGetOnlineArchiveDataProcessRegionArrayOutputWithContext(ctx context.Context) GetOnlineArchiveDataProcessRegionArrayOutput { + return pulumi.ToOutputWithContext(ctx, i).(GetOnlineArchiveDataProcessRegionArrayOutput) +} + +type GetOnlineArchiveDataProcessRegionOutput struct{ *pulumi.OutputState } + +func (GetOnlineArchiveDataProcessRegionOutput) ElementType() reflect.Type { + return reflect.TypeOf((*GetOnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (o GetOnlineArchiveDataProcessRegionOutput) ToGetOnlineArchiveDataProcessRegionOutput() GetOnlineArchiveDataProcessRegionOutput { + return o +} + +func (o GetOnlineArchiveDataProcessRegionOutput) ToGetOnlineArchiveDataProcessRegionOutputWithContext(ctx context.Context) GetOnlineArchiveDataProcessRegionOutput { + return o +} + +func (o GetOnlineArchiveDataProcessRegionOutput) CloudProvider() pulumi.StringOutput { + return o.ApplyT(func(v GetOnlineArchiveDataProcessRegion) string { return v.CloudProvider }).(pulumi.StringOutput) +} + +func (o GetOnlineArchiveDataProcessRegionOutput) Region() pulumi.StringOutput { + return o.ApplyT(func(v GetOnlineArchiveDataProcessRegion) string { return v.Region }).(pulumi.StringOutput) +} + +type GetOnlineArchiveDataProcessRegionArrayOutput struct{ *pulumi.OutputState } + +func (GetOnlineArchiveDataProcessRegionArrayOutput) ElementType() reflect.Type { + return reflect.TypeOf((*[]GetOnlineArchiveDataProcessRegion)(nil)).Elem() +} + +func (o GetOnlineArchiveDataProcessRegionArrayOutput) ToGetOnlineArchiveDataProcessRegionArrayOutput() GetOnlineArchiveDataProcessRegionArrayOutput { + return o +} + +func (o GetOnlineArchiveDataProcessRegionArrayOutput) ToGetOnlineArchiveDataProcessRegionArrayOutputWithContext(ctx context.Context) GetOnlineArchiveDataProcessRegionArrayOutput { + return o +} + +func (o GetOnlineArchiveDataProcessRegionArrayOutput) Index(i pulumi.IntInput) GetOnlineArchiveDataProcessRegionOutput { + return pulumi.All(o, i).ApplyT(func(vs []interface{}) GetOnlineArchiveDataProcessRegion { + return vs[0].([]GetOnlineArchiveDataProcessRegion)[vs[1].(int)] + }).(GetOnlineArchiveDataProcessRegionOutput) +} + type GetOnlineArchivePartitionField struct { FieldName string `pulumi:"fieldName"` FieldType string `pulumi:"fieldType"` @@ -41425,6 +41681,7 @@ type GetOnlineArchivesResult struct { CollectionType string `pulumi:"collectionType"` Criterias []GetOnlineArchivesResultCriteria `pulumi:"criterias"` DataExpirationRules []GetOnlineArchivesResultDataExpirationRule `pulumi:"dataExpirationRules"` + DataProcessRegions []GetOnlineArchivesResultDataProcessRegion `pulumi:"dataProcessRegions"` DbName string `pulumi:"dbName"` PartitionFields []GetOnlineArchivesResultPartitionField `pulumi:"partitionFields"` Paused bool `pulumi:"paused"` @@ -41457,6 +41714,7 @@ type GetOnlineArchivesResultArgs struct { CollectionType pulumi.StringInput `pulumi:"collectionType"` Criterias GetOnlineArchivesResultCriteriaArrayInput `pulumi:"criterias"` DataExpirationRules GetOnlineArchivesResultDataExpirationRuleArrayInput `pulumi:"dataExpirationRules"` + DataProcessRegions GetOnlineArchivesResultDataProcessRegionArrayInput `pulumi:"dataProcessRegions"` DbName pulumi.StringInput `pulumi:"dbName"` PartitionFields GetOnlineArchivesResultPartitionFieldArrayInput `pulumi:"partitionFields"` Paused pulumi.BoolInput `pulumi:"paused"` @@ -41548,6 +41806,12 @@ func (o GetOnlineArchivesResultOutput) DataExpirationRules() GetOnlineArchivesRe }).(GetOnlineArchivesResultDataExpirationRuleArrayOutput) } +func (o GetOnlineArchivesResultOutput) DataProcessRegions() GetOnlineArchivesResultDataProcessRegionArrayOutput { + return o.ApplyT(func(v GetOnlineArchivesResult) []GetOnlineArchivesResultDataProcessRegion { + return v.DataProcessRegions + }).(GetOnlineArchivesResultDataProcessRegionArrayOutput) +} + func (o GetOnlineArchivesResultOutput) DbName() pulumi.StringOutput { return o.ApplyT(func(v GetOnlineArchivesResult) string { return v.DbName }).(pulumi.StringOutput) } @@ -41805,6 +42069,106 @@ func (o GetOnlineArchivesResultDataExpirationRuleArrayOutput) Index(i pulumi.Int }).(GetOnlineArchivesResultDataExpirationRuleOutput) } +type GetOnlineArchivesResultDataProcessRegion struct { + CloudProvider string `pulumi:"cloudProvider"` + Region string `pulumi:"region"` +} + +// GetOnlineArchivesResultDataProcessRegionInput is an input type that accepts GetOnlineArchivesResultDataProcessRegionArgs and GetOnlineArchivesResultDataProcessRegionOutput values. +// You can construct a concrete instance of `GetOnlineArchivesResultDataProcessRegionInput` via: +// +// GetOnlineArchivesResultDataProcessRegionArgs{...} +type GetOnlineArchivesResultDataProcessRegionInput interface { + pulumi.Input + + ToGetOnlineArchivesResultDataProcessRegionOutput() GetOnlineArchivesResultDataProcessRegionOutput + ToGetOnlineArchivesResultDataProcessRegionOutputWithContext(context.Context) GetOnlineArchivesResultDataProcessRegionOutput +} + +type GetOnlineArchivesResultDataProcessRegionArgs struct { + CloudProvider pulumi.StringInput `pulumi:"cloudProvider"` + Region pulumi.StringInput `pulumi:"region"` +} + +func (GetOnlineArchivesResultDataProcessRegionArgs) ElementType() reflect.Type { + return reflect.TypeOf((*GetOnlineArchivesResultDataProcessRegion)(nil)).Elem() +} + +func (i GetOnlineArchivesResultDataProcessRegionArgs) ToGetOnlineArchivesResultDataProcessRegionOutput() GetOnlineArchivesResultDataProcessRegionOutput { + return i.ToGetOnlineArchivesResultDataProcessRegionOutputWithContext(context.Background()) +} + +func (i GetOnlineArchivesResultDataProcessRegionArgs) ToGetOnlineArchivesResultDataProcessRegionOutputWithContext(ctx context.Context) GetOnlineArchivesResultDataProcessRegionOutput { + return pulumi.ToOutputWithContext(ctx, i).(GetOnlineArchivesResultDataProcessRegionOutput) +} + +// GetOnlineArchivesResultDataProcessRegionArrayInput is an input type that accepts GetOnlineArchivesResultDataProcessRegionArray and GetOnlineArchivesResultDataProcessRegionArrayOutput values. +// You can construct a concrete instance of `GetOnlineArchivesResultDataProcessRegionArrayInput` via: +// +// GetOnlineArchivesResultDataProcessRegionArray{ GetOnlineArchivesResultDataProcessRegionArgs{...} } +type GetOnlineArchivesResultDataProcessRegionArrayInput interface { + pulumi.Input + + ToGetOnlineArchivesResultDataProcessRegionArrayOutput() GetOnlineArchivesResultDataProcessRegionArrayOutput + ToGetOnlineArchivesResultDataProcessRegionArrayOutputWithContext(context.Context) GetOnlineArchivesResultDataProcessRegionArrayOutput +} + +type GetOnlineArchivesResultDataProcessRegionArray []GetOnlineArchivesResultDataProcessRegionInput + +func (GetOnlineArchivesResultDataProcessRegionArray) ElementType() reflect.Type { + return reflect.TypeOf((*[]GetOnlineArchivesResultDataProcessRegion)(nil)).Elem() +} + +func (i GetOnlineArchivesResultDataProcessRegionArray) ToGetOnlineArchivesResultDataProcessRegionArrayOutput() GetOnlineArchivesResultDataProcessRegionArrayOutput { + return i.ToGetOnlineArchivesResultDataProcessRegionArrayOutputWithContext(context.Background()) +} + +func (i GetOnlineArchivesResultDataProcessRegionArray) ToGetOnlineArchivesResultDataProcessRegionArrayOutputWithContext(ctx context.Context) GetOnlineArchivesResultDataProcessRegionArrayOutput { + return pulumi.ToOutputWithContext(ctx, i).(GetOnlineArchivesResultDataProcessRegionArrayOutput) +} + +type GetOnlineArchivesResultDataProcessRegionOutput struct{ *pulumi.OutputState } + +func (GetOnlineArchivesResultDataProcessRegionOutput) ElementType() reflect.Type { + return reflect.TypeOf((*GetOnlineArchivesResultDataProcessRegion)(nil)).Elem() +} + +func (o GetOnlineArchivesResultDataProcessRegionOutput) ToGetOnlineArchivesResultDataProcessRegionOutput() GetOnlineArchivesResultDataProcessRegionOutput { + return o +} + +func (o GetOnlineArchivesResultDataProcessRegionOutput) ToGetOnlineArchivesResultDataProcessRegionOutputWithContext(ctx context.Context) GetOnlineArchivesResultDataProcessRegionOutput { + return o +} + +func (o GetOnlineArchivesResultDataProcessRegionOutput) CloudProvider() pulumi.StringOutput { + return o.ApplyT(func(v GetOnlineArchivesResultDataProcessRegion) string { return v.CloudProvider }).(pulumi.StringOutput) +} + +func (o GetOnlineArchivesResultDataProcessRegionOutput) Region() pulumi.StringOutput { + return o.ApplyT(func(v GetOnlineArchivesResultDataProcessRegion) string { return v.Region }).(pulumi.StringOutput) +} + +type GetOnlineArchivesResultDataProcessRegionArrayOutput struct{ *pulumi.OutputState } + +func (GetOnlineArchivesResultDataProcessRegionArrayOutput) ElementType() reflect.Type { + return reflect.TypeOf((*[]GetOnlineArchivesResultDataProcessRegion)(nil)).Elem() +} + +func (o GetOnlineArchivesResultDataProcessRegionArrayOutput) ToGetOnlineArchivesResultDataProcessRegionArrayOutput() GetOnlineArchivesResultDataProcessRegionArrayOutput { + return o +} + +func (o GetOnlineArchivesResultDataProcessRegionArrayOutput) ToGetOnlineArchivesResultDataProcessRegionArrayOutputWithContext(ctx context.Context) GetOnlineArchivesResultDataProcessRegionArrayOutput { + return o +} + +func (o GetOnlineArchivesResultDataProcessRegionArrayOutput) Index(i pulumi.IntInput) GetOnlineArchivesResultDataProcessRegionOutput { + return pulumi.All(o, i).ApplyT(func(vs []interface{}) GetOnlineArchivesResultDataProcessRegion { + return vs[0].([]GetOnlineArchivesResultDataProcessRegion)[vs[1].(int)] + }).(GetOnlineArchivesResultDataProcessRegionOutput) +} + type GetOnlineArchivesResultPartitionField struct { FieldName string `pulumi:"fieldName"` FieldType string `pulumi:"fieldType"` @@ -45790,6 +46154,8 @@ func init() { pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchiveCriteriaPtrInput)(nil)).Elem(), OnlineArchiveCriteriaArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchiveDataExpirationRuleInput)(nil)).Elem(), OnlineArchiveDataExpirationRuleArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchiveDataExpirationRulePtrInput)(nil)).Elem(), OnlineArchiveDataExpirationRuleArgs{}) + pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchiveDataProcessRegionInput)(nil)).Elem(), OnlineArchiveDataProcessRegionArgs{}) + pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchiveDataProcessRegionPtrInput)(nil)).Elem(), OnlineArchiveDataProcessRegionArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchivePartitionFieldInput)(nil)).Elem(), OnlineArchivePartitionFieldArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchivePartitionFieldArrayInput)(nil)).Elem(), OnlineArchivePartitionFieldArray{}) pulumi.RegisterInputType(reflect.TypeOf((*OnlineArchiveScheduleInput)(nil)).Elem(), OnlineArchiveScheduleArgs{}) @@ -46192,6 +46558,8 @@ func init() { pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchiveCriteriaArrayInput)(nil)).Elem(), GetOnlineArchiveCriteriaArray{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchiveDataExpirationRuleInput)(nil)).Elem(), GetOnlineArchiveDataExpirationRuleArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchiveDataExpirationRuleArrayInput)(nil)).Elem(), GetOnlineArchiveDataExpirationRuleArray{}) + pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchiveDataProcessRegionInput)(nil)).Elem(), GetOnlineArchiveDataProcessRegionArgs{}) + pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchiveDataProcessRegionArrayInput)(nil)).Elem(), GetOnlineArchiveDataProcessRegionArray{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivePartitionFieldInput)(nil)).Elem(), GetOnlineArchivePartitionFieldArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivePartitionFieldArrayInput)(nil)).Elem(), GetOnlineArchivePartitionFieldArray{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchiveScheduleInput)(nil)).Elem(), GetOnlineArchiveScheduleArgs{}) @@ -46202,6 +46570,8 @@ func init() { pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultCriteriaArrayInput)(nil)).Elem(), GetOnlineArchivesResultCriteriaArray{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultDataExpirationRuleInput)(nil)).Elem(), GetOnlineArchivesResultDataExpirationRuleArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultDataExpirationRuleArrayInput)(nil)).Elem(), GetOnlineArchivesResultDataExpirationRuleArray{}) + pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultDataProcessRegionInput)(nil)).Elem(), GetOnlineArchivesResultDataProcessRegionArgs{}) + pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultDataProcessRegionArrayInput)(nil)).Elem(), GetOnlineArchivesResultDataProcessRegionArray{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultPartitionFieldInput)(nil)).Elem(), GetOnlineArchivesResultPartitionFieldArgs{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultPartitionFieldArrayInput)(nil)).Elem(), GetOnlineArchivesResultPartitionFieldArray{}) pulumi.RegisterInputType(reflect.TypeOf((*GetOnlineArchivesResultScheduleInput)(nil)).Elem(), GetOnlineArchivesResultScheduleArgs{}) @@ -46436,6 +46806,8 @@ func init() { pulumi.RegisterOutputType(OnlineArchiveCriteriaPtrOutput{}) pulumi.RegisterOutputType(OnlineArchiveDataExpirationRuleOutput{}) pulumi.RegisterOutputType(OnlineArchiveDataExpirationRulePtrOutput{}) + pulumi.RegisterOutputType(OnlineArchiveDataProcessRegionOutput{}) + pulumi.RegisterOutputType(OnlineArchiveDataProcessRegionPtrOutput{}) pulumi.RegisterOutputType(OnlineArchivePartitionFieldOutput{}) pulumi.RegisterOutputType(OnlineArchivePartitionFieldArrayOutput{}) pulumi.RegisterOutputType(OnlineArchiveScheduleOutput{}) @@ -46838,6 +47210,8 @@ func init() { pulumi.RegisterOutputType(GetOnlineArchiveCriteriaArrayOutput{}) pulumi.RegisterOutputType(GetOnlineArchiveDataExpirationRuleOutput{}) pulumi.RegisterOutputType(GetOnlineArchiveDataExpirationRuleArrayOutput{}) + pulumi.RegisterOutputType(GetOnlineArchiveDataProcessRegionOutput{}) + pulumi.RegisterOutputType(GetOnlineArchiveDataProcessRegionArrayOutput{}) pulumi.RegisterOutputType(GetOnlineArchivePartitionFieldOutput{}) pulumi.RegisterOutputType(GetOnlineArchivePartitionFieldArrayOutput{}) pulumi.RegisterOutputType(GetOnlineArchiveScheduleOutput{}) @@ -46848,6 +47222,8 @@ func init() { pulumi.RegisterOutputType(GetOnlineArchivesResultCriteriaArrayOutput{}) pulumi.RegisterOutputType(GetOnlineArchivesResultDataExpirationRuleOutput{}) pulumi.RegisterOutputType(GetOnlineArchivesResultDataExpirationRuleArrayOutput{}) + pulumi.RegisterOutputType(GetOnlineArchivesResultDataProcessRegionOutput{}) + pulumi.RegisterOutputType(GetOnlineArchivesResultDataProcessRegionArrayOutput{}) pulumi.RegisterOutputType(GetOnlineArchivesResultPartitionFieldOutput{}) pulumi.RegisterOutputType(GetOnlineArchivesResultPartitionFieldArrayOutput{}) pulumi.RegisterOutputType(GetOnlineArchivesResultScheduleOutput{}) diff --git a/sdk/go/mongodbatlas/searchDeployment.go b/sdk/go/mongodbatlas/searchDeployment.go index d9b0f759..2f1d780c 100644 --- a/sdk/go/mongodbatlas/searchDeployment.go +++ b/sdk/go/mongodbatlas/searchDeployment.go @@ -15,7 +15,8 @@ import ( // `SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster. // // > **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation). -// **NOTE:** Only a single search deployment resource can be defined for each cluster. +// +// > **NOTE:** Only a single search deployment resource can be defined for each cluster. // // ## Import // diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/Cluster.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/Cluster.java index 87ac4b97..74e99448 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/Cluster.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/Cluster.java @@ -43,14 +43,14 @@ public class Cluster extends com.pulumi.resources.CustomResource { * */ @Export(name="acceptDataRisksAndForceReplicaSetReconfig", refs={String.class}, tree="[0]") - private Output acceptDataRisksAndForceReplicaSetReconfig; + private Output acceptDataRisksAndForceReplicaSetReconfig; /** * @return If reconfiguration is necessary to regain a primary due to a regional outage, submit this field alongside your topology reconfiguration to request a new regional outage resistant topology. Forced reconfigurations during an outage of the majority of electable nodes carry a risk of data loss if replicated writes (even majority committed writes) have not been replicated to the new primary node. MongoDB Atlas docs contain more information. To proceed with an operation which carries that risk, set `accept_data_risks_and_force_replica_set_reconfig` to the current date. Learn more about Reconfiguring a Replica Set during a regional outage [here](https://dochub.mongodb.org/core/regional-outage-reconfigure-replica-set). * */ - public Output> acceptDataRisksAndForceReplicaSetReconfig() { - return Codegen.optional(this.acceptDataRisksAndForceReplicaSetReconfig); + public Output acceptDataRisksAndForceReplicaSetReconfig() { + return this.acceptDataRisksAndForceReplicaSetReconfig; } @Export(name="advancedConfiguration", refs={ClusterAdvancedConfiguration.class}, tree="[0]") private Output advancedConfiguration; @@ -81,10 +81,10 @@ public Output autoScalingComputeScaleDownEnabled() { return this.autoScalingComputeScaleDownEnabled; } @Export(name="autoScalingDiskGbEnabled", refs={Boolean.class}, tree="[0]") - private Output autoScalingDiskGbEnabled; + private Output autoScalingDiskGbEnabled; - public Output> autoScalingDiskGbEnabled() { - return Codegen.optional(this.autoScalingDiskGbEnabled); + public Output autoScalingDiskGbEnabled() { + return this.autoScalingDiskGbEnabled; } /** * Cloud service provider on which the server for a multi-tenant cluster is provisioned. diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/MongodbatlasFunctions.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/MongodbatlasFunctions.java index 284bc4a8..c9196138 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/MongodbatlasFunctions.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/MongodbatlasFunctions.java @@ -8543,6 +8543,7 @@ public static CompletableFuture getNetworkPeerin * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -8558,6 +8559,10 @@ public static CompletableFuture getNetworkPeerin * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. @@ -8622,6 +8627,7 @@ public static Output getOnlineArchive(GetOnlineArchiveAr * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -8637,6 +8643,10 @@ public static Output getOnlineArchive(GetOnlineArchiveAr * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. @@ -8701,6 +8711,7 @@ public static CompletableFuture getOnlineArchivePlain(Ge * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -8716,6 +8727,10 @@ public static CompletableFuture getOnlineArchivePlain(Ge * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. @@ -8780,6 +8795,7 @@ public static Output getOnlineArchive(GetOnlineArchiveAr * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -8795,6 +8811,10 @@ public static Output getOnlineArchive(GetOnlineArchiveAr * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. @@ -8858,6 +8878,7 @@ public static CompletableFuture getOnlineArchivePlain(Ge * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -8873,6 +8894,10 @@ public static CompletableFuture getOnlineArchivePlain(Ge * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). @@ -8934,6 +8959,7 @@ public static Output getOnlineArchives(GetOnlineA * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -8949,6 +8975,10 @@ public static Output getOnlineArchives(GetOnlineA * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). @@ -9010,6 +9040,7 @@ public static CompletableFuture getOnlineArchives * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -9025,6 +9056,10 @@ public static CompletableFuture getOnlineArchives * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). @@ -9086,6 +9121,7 @@ public static Output getOnlineArchives(GetOnlineA * * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -9101,6 +9137,10 @@ public static Output getOnlineArchives(GetOnlineA * ### Data Expiration Rule * * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchive.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchive.java index 9a52e542..d83b5d8e 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchive.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchive.java @@ -12,6 +12,7 @@ import com.pulumi.mongodbatlas.inputs.OnlineArchiveState; import com.pulumi.mongodbatlas.outputs.OnlineArchiveCriteria; import com.pulumi.mongodbatlas.outputs.OnlineArchiveDataExpirationRule; +import com.pulumi.mongodbatlas.outputs.OnlineArchiveDataProcessRegion; import com.pulumi.mongodbatlas.outputs.OnlineArchivePartitionField; import com.pulumi.mongodbatlas.outputs.OnlineArchiveSchedule; import java.lang.Boolean; @@ -222,6 +223,20 @@ public Output criteria() { public Output> dataExpirationRule() { return Codegen.optional(this.dataExpirationRule); } + /** + * Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + */ + @Export(name="dataProcessRegion", refs={OnlineArchiveDataProcessRegion.class}, tree="[0]") + private Output dataProcessRegion; + + /** + * @return Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + */ + public Output dataProcessRegion() { + return this.dataProcessRegion; + } /** * Name of the database that contains the collection. * diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchiveArgs.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchiveArgs.java index 6e141916..ae6809af 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchiveArgs.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/OnlineArchiveArgs.java @@ -7,6 +7,7 @@ import com.pulumi.core.annotations.Import; import com.pulumi.mongodbatlas.inputs.OnlineArchiveCriteriaArgs; import com.pulumi.mongodbatlas.inputs.OnlineArchiveDataExpirationRuleArgs; +import com.pulumi.mongodbatlas.inputs.OnlineArchiveDataProcessRegionArgs; import com.pulumi.mongodbatlas.inputs.OnlineArchivePartitionFieldArgs; import com.pulumi.mongodbatlas.inputs.OnlineArchiveScheduleArgs; import java.lang.Boolean; @@ -96,6 +97,21 @@ public Optional> dataExpirationRule( return Optional.ofNullable(this.dataExpirationRule); } + /** + * Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + */ + @Import(name="dataProcessRegion") + private @Nullable Output dataProcessRegion; + + /** + * @return Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + */ + public Optional> dataProcessRegion() { + return Optional.ofNullable(this.dataProcessRegion); + } + /** * Name of the database that contains the collection. * @@ -186,6 +202,7 @@ private OnlineArchiveArgs(OnlineArchiveArgs $) { this.collectionType = $.collectionType; this.criteria = $.criteria; this.dataExpirationRule = $.dataExpirationRule; + this.dataProcessRegion = $.dataProcessRegion; this.dbName = $.dbName; this.partitionFields = $.partitionFields; this.paused = $.paused; @@ -317,6 +334,27 @@ public Builder dataExpirationRule(OnlineArchiveDataExpirationRuleArgs dataExpira return dataExpirationRule(Output.of(dataExpirationRule)); } + /** + * @param dataProcessRegion Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + * @return builder + * + */ + public Builder dataProcessRegion(@Nullable Output dataProcessRegion) { + $.dataProcessRegion = dataProcessRegion; + return this; + } + + /** + * @param dataProcessRegion Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + * @return builder + * + */ + public Builder dataProcessRegion(OnlineArchiveDataProcessRegionArgs dataProcessRegion) { + return dataProcessRegion(Output.of(dataProcessRegion)); + } + /** * @param dbName Name of the database that contains the collection. * diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/SearchDeployment.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/SearchDeployment.java index 392cde67..b1682b2f 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/SearchDeployment.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/SearchDeployment.java @@ -21,7 +21,8 @@ * `mongodbatlas.SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster. * * > **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation). - * **NOTE:** Only a single search deployment resource can be defined for each cluster. + * + * > **NOTE:** Only a single search deployment resource can be defined for each cluster. * * ## Example Usage * diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveDataProcessRegionArgs.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveDataProcessRegionArgs.java new file mode 100644 index 00000000..791c732b --- /dev/null +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveDataProcessRegionArgs.java @@ -0,0 +1,120 @@ +// *** WARNING: this file was generated by pulumi-java-gen. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +package com.pulumi.mongodbatlas.inputs; + +import com.pulumi.core.Output; +import com.pulumi.core.annotations.Import; +import java.lang.String; +import java.util.Objects; +import java.util.Optional; +import javax.annotation.Nullable; + + +public final class OnlineArchiveDataProcessRegionArgs extends com.pulumi.resources.ResourceArgs { + + public static final OnlineArchiveDataProcessRegionArgs Empty = new OnlineArchiveDataProcessRegionArgs(); + + /** + * Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * + */ + @Import(name="cloudProvider") + private @Nullable Output cloudProvider; + + /** + * @return Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * + */ + public Optional> cloudProvider() { + return Optional.ofNullable(this.cloudProvider); + } + + /** + * Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * + */ + @Import(name="region") + private @Nullable Output region; + + /** + * @return Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * + */ + public Optional> region() { + return Optional.ofNullable(this.region); + } + + private OnlineArchiveDataProcessRegionArgs() {} + + private OnlineArchiveDataProcessRegionArgs(OnlineArchiveDataProcessRegionArgs $) { + this.cloudProvider = $.cloudProvider; + this.region = $.region; + } + + public static Builder builder() { + return new Builder(); + } + public static Builder builder(OnlineArchiveDataProcessRegionArgs defaults) { + return new Builder(defaults); + } + + public static final class Builder { + private OnlineArchiveDataProcessRegionArgs $; + + public Builder() { + $ = new OnlineArchiveDataProcessRegionArgs(); + } + + public Builder(OnlineArchiveDataProcessRegionArgs defaults) { + $ = new OnlineArchiveDataProcessRegionArgs(Objects.requireNonNull(defaults)); + } + + /** + * @param cloudProvider Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * + * @return builder + * + */ + public Builder cloudProvider(@Nullable Output cloudProvider) { + $.cloudProvider = cloudProvider; + return this; + } + + /** + * @param cloudProvider Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * + * @return builder + * + */ + public Builder cloudProvider(String cloudProvider) { + return cloudProvider(Output.of(cloudProvider)); + } + + /** + * @param region Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * + * @return builder + * + */ + public Builder region(@Nullable Output region) { + $.region = region; + return this; + } + + /** + * @param region Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * + * @return builder + * + */ + public Builder region(String region) { + return region(Output.of(region)); + } + + public OnlineArchiveDataProcessRegionArgs build() { + return $; + } + } + +} diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveState.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveState.java index ee16c9b9..cc885bf2 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveState.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/inputs/OnlineArchiveState.java @@ -7,6 +7,7 @@ import com.pulumi.core.annotations.Import; import com.pulumi.mongodbatlas.inputs.OnlineArchiveCriteriaArgs; import com.pulumi.mongodbatlas.inputs.OnlineArchiveDataExpirationRuleArgs; +import com.pulumi.mongodbatlas.inputs.OnlineArchiveDataProcessRegionArgs; import com.pulumi.mongodbatlas.inputs.OnlineArchivePartitionFieldArgs; import com.pulumi.mongodbatlas.inputs.OnlineArchiveScheduleArgs; import java.lang.Boolean; @@ -111,6 +112,21 @@ public Optional> dataExpirationRule( return Optional.ofNullable(this.dataExpirationRule); } + /** + * Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + */ + @Import(name="dataProcessRegion") + private @Nullable Output dataProcessRegion; + + /** + * @return Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + */ + public Optional> dataProcessRegion() { + return Optional.ofNullable(this.dataProcessRegion); + } + /** * Name of the database that contains the collection. * @@ -217,6 +233,7 @@ private OnlineArchiveState(OnlineArchiveState $) { this.collectionType = $.collectionType; this.criteria = $.criteria; this.dataExpirationRule = $.dataExpirationRule; + this.dataProcessRegion = $.dataProcessRegion; this.dbName = $.dbName; this.partitionFields = $.partitionFields; this.paused = $.paused; @@ -370,6 +387,27 @@ public Builder dataExpirationRule(OnlineArchiveDataExpirationRuleArgs dataExpira return dataExpirationRule(Output.of(dataExpirationRule)); } + /** + * @param dataProcessRegion Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + * @return builder + * + */ + public Builder dataProcessRegion(@Nullable Output dataProcessRegion) { + $.dataProcessRegion = dataProcessRegion; + return this; + } + + /** + * @param dataProcessRegion Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + * + * @return builder + * + */ + public Builder dataProcessRegion(OnlineArchiveDataProcessRegionArgs dataProcessRegion) { + return dataProcessRegion(Output.of(dataProcessRegion)); + } + /** * @param dbName Name of the database that contains the collection. * diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveDataProcessRegion.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveDataProcessRegion.java new file mode 100644 index 00000000..4f131242 --- /dev/null +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveDataProcessRegion.java @@ -0,0 +1,58 @@ +// *** WARNING: this file was generated by pulumi-java-gen. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +package com.pulumi.mongodbatlas.outputs; + +import com.pulumi.core.annotations.CustomType; +import java.lang.String; +import java.util.Objects; + +@CustomType +public final class GetOnlineArchiveDataProcessRegion { + private String cloudProvider; + private String region; + + private GetOnlineArchiveDataProcessRegion() {} + public String cloudProvider() { + return this.cloudProvider; + } + public String region() { + return this.region; + } + + public static Builder builder() { + return new Builder(); + } + + public static Builder builder(GetOnlineArchiveDataProcessRegion defaults) { + return new Builder(defaults); + } + @CustomType.Builder + public static final class Builder { + private String cloudProvider; + private String region; + public Builder() {} + public Builder(GetOnlineArchiveDataProcessRegion defaults) { + Objects.requireNonNull(defaults); + this.cloudProvider = defaults.cloudProvider; + this.region = defaults.region; + } + + @CustomType.Setter + public Builder cloudProvider(String cloudProvider) { + this.cloudProvider = Objects.requireNonNull(cloudProvider); + return this; + } + @CustomType.Setter + public Builder region(String region) { + this.region = Objects.requireNonNull(region); + return this; + } + public GetOnlineArchiveDataProcessRegion build() { + final var o = new GetOnlineArchiveDataProcessRegion(); + o.cloudProvider = cloudProvider; + o.region = region; + return o; + } + } +} diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveResult.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveResult.java index c34603b2..5d79a6a7 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveResult.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchiveResult.java @@ -6,6 +6,7 @@ import com.pulumi.core.annotations.CustomType; import com.pulumi.mongodbatlas.outputs.GetOnlineArchiveCriteria; import com.pulumi.mongodbatlas.outputs.GetOnlineArchiveDataExpirationRule; +import com.pulumi.mongodbatlas.outputs.GetOnlineArchiveDataProcessRegion; import com.pulumi.mongodbatlas.outputs.GetOnlineArchivePartitionField; import com.pulumi.mongodbatlas.outputs.GetOnlineArchiveSchedule; import java.lang.Boolean; @@ -21,6 +22,7 @@ public final class GetOnlineArchiveResult { private String collectionType; private List criterias; private List dataExpirationRules; + private List dataProcessRegions; private String dbName; /** * @return The provider-assigned unique ID for this managed resource. @@ -52,6 +54,9 @@ public List criterias() { public List dataExpirationRules() { return this.dataExpirationRules; } + public List dataProcessRegions() { + return this.dataProcessRegions; + } public String dbName() { return this.dbName; } @@ -93,6 +98,7 @@ public static final class Builder { private String collectionType; private List criterias; private List dataExpirationRules; + private List dataProcessRegions; private String dbName; private String id; private List partitionFields; @@ -109,6 +115,7 @@ public Builder(GetOnlineArchiveResult defaults) { this.collectionType = defaults.collectionType; this.criterias = defaults.criterias; this.dataExpirationRules = defaults.dataExpirationRules; + this.dataProcessRegions = defaults.dataProcessRegions; this.dbName = defaults.dbName; this.id = defaults.id; this.partitionFields = defaults.partitionFields; @@ -155,6 +162,14 @@ public Builder dataExpirationRules(GetOnlineArchiveDataExpirationRule... dataExp return dataExpirationRules(List.of(dataExpirationRules)); } @CustomType.Setter + public Builder dataProcessRegions(List dataProcessRegions) { + this.dataProcessRegions = Objects.requireNonNull(dataProcessRegions); + return this; + } + public Builder dataProcessRegions(GetOnlineArchiveDataProcessRegion... dataProcessRegions) { + return dataProcessRegions(List.of(dataProcessRegions)); + } + @CustomType.Setter public Builder dbName(String dbName) { this.dbName = Objects.requireNonNull(dbName); return this; @@ -203,6 +218,7 @@ public GetOnlineArchiveResult build() { o.collectionType = collectionType; o.criterias = criterias; o.dataExpirationRules = dataExpirationRules; + o.dataProcessRegions = dataProcessRegions; o.dbName = dbName; o.id = id; o.partitionFields = partitionFields; diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResult.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResult.java index 703820a2..0a1c3161 100644 --- a/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResult.java +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResult.java @@ -6,6 +6,7 @@ import com.pulumi.core.annotations.CustomType; import com.pulumi.mongodbatlas.outputs.GetOnlineArchivesResultCriteria; import com.pulumi.mongodbatlas.outputs.GetOnlineArchivesResultDataExpirationRule; +import com.pulumi.mongodbatlas.outputs.GetOnlineArchivesResultDataProcessRegion; import com.pulumi.mongodbatlas.outputs.GetOnlineArchivesResultPartitionField; import com.pulumi.mongodbatlas.outputs.GetOnlineArchivesResultSchedule; import java.lang.Boolean; @@ -29,6 +30,7 @@ public final class GetOnlineArchivesResult { private String collectionType; private List criterias; private List dataExpirationRules; + private List dataProcessRegions; private String dbName; private List partitionFields; private Boolean paused; @@ -67,6 +69,9 @@ public List criterias() { public List dataExpirationRules() { return this.dataExpirationRules; } + public List dataProcessRegions() { + return this.dataProcessRegions; + } public String dbName() { return this.dbName; } @@ -105,6 +110,7 @@ public static final class Builder { private String collectionType; private List criterias; private List dataExpirationRules; + private List dataProcessRegions; private String dbName; private List partitionFields; private Boolean paused; @@ -120,6 +126,7 @@ public Builder(GetOnlineArchivesResult defaults) { this.collectionType = defaults.collectionType; this.criterias = defaults.criterias; this.dataExpirationRules = defaults.dataExpirationRules; + this.dataProcessRegions = defaults.dataProcessRegions; this.dbName = defaults.dbName; this.partitionFields = defaults.partitionFields; this.paused = defaults.paused; @@ -165,6 +172,14 @@ public Builder dataExpirationRules(GetOnlineArchivesResultDataExpirationRule... return dataExpirationRules(List.of(dataExpirationRules)); } @CustomType.Setter + public Builder dataProcessRegions(List dataProcessRegions) { + this.dataProcessRegions = Objects.requireNonNull(dataProcessRegions); + return this; + } + public Builder dataProcessRegions(GetOnlineArchivesResultDataProcessRegion... dataProcessRegions) { + return dataProcessRegions(List.of(dataProcessRegions)); + } + @CustomType.Setter public Builder dbName(String dbName) { this.dbName = Objects.requireNonNull(dbName); return this; @@ -208,6 +223,7 @@ public GetOnlineArchivesResult build() { o.collectionType = collectionType; o.criterias = criterias; o.dataExpirationRules = dataExpirationRules; + o.dataProcessRegions = dataProcessRegions; o.dbName = dbName; o.partitionFields = partitionFields; o.paused = paused; diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResultDataProcessRegion.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResultDataProcessRegion.java new file mode 100644 index 00000000..d78e5a3c --- /dev/null +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/GetOnlineArchivesResultDataProcessRegion.java @@ -0,0 +1,58 @@ +// *** WARNING: this file was generated by pulumi-java-gen. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +package com.pulumi.mongodbatlas.outputs; + +import com.pulumi.core.annotations.CustomType; +import java.lang.String; +import java.util.Objects; + +@CustomType +public final class GetOnlineArchivesResultDataProcessRegion { + private String cloudProvider; + private String region; + + private GetOnlineArchivesResultDataProcessRegion() {} + public String cloudProvider() { + return this.cloudProvider; + } + public String region() { + return this.region; + } + + public static Builder builder() { + return new Builder(); + } + + public static Builder builder(GetOnlineArchivesResultDataProcessRegion defaults) { + return new Builder(defaults); + } + @CustomType.Builder + public static final class Builder { + private String cloudProvider; + private String region; + public Builder() {} + public Builder(GetOnlineArchivesResultDataProcessRegion defaults) { + Objects.requireNonNull(defaults); + this.cloudProvider = defaults.cloudProvider; + this.region = defaults.region; + } + + @CustomType.Setter + public Builder cloudProvider(String cloudProvider) { + this.cloudProvider = Objects.requireNonNull(cloudProvider); + return this; + } + @CustomType.Setter + public Builder region(String region) { + this.region = Objects.requireNonNull(region); + return this; + } + public GetOnlineArchivesResultDataProcessRegion build() { + final var o = new GetOnlineArchivesResultDataProcessRegion(); + o.cloudProvider = cloudProvider; + o.region = region; + return o; + } + } +} diff --git a/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/OnlineArchiveDataProcessRegion.java b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/OnlineArchiveDataProcessRegion.java new file mode 100644 index 00000000..9ef0b591 --- /dev/null +++ b/sdk/java/src/main/java/com/pulumi/mongodbatlas/outputs/OnlineArchiveDataProcessRegion.java @@ -0,0 +1,76 @@ +// *** WARNING: this file was generated by pulumi-java-gen. *** +// *** Do not edit by hand unless you're certain you know what you are doing! *** + +package com.pulumi.mongodbatlas.outputs; + +import com.pulumi.core.annotations.CustomType; +import java.lang.String; +import java.util.Objects; +import java.util.Optional; +import javax.annotation.Nullable; + +@CustomType +public final class OnlineArchiveDataProcessRegion { + /** + * @return Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * + */ + private @Nullable String cloudProvider; + /** + * @return Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * + */ + private @Nullable String region; + + private OnlineArchiveDataProcessRegion() {} + /** + * @return Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * + */ + public Optional cloudProvider() { + return Optional.ofNullable(this.cloudProvider); + } + /** + * @return Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * + */ + public Optional region() { + return Optional.ofNullable(this.region); + } + + public static Builder builder() { + return new Builder(); + } + + public static Builder builder(OnlineArchiveDataProcessRegion defaults) { + return new Builder(defaults); + } + @CustomType.Builder + public static final class Builder { + private @Nullable String cloudProvider; + private @Nullable String region; + public Builder() {} + public Builder(OnlineArchiveDataProcessRegion defaults) { + Objects.requireNonNull(defaults); + this.cloudProvider = defaults.cloudProvider; + this.region = defaults.region; + } + + @CustomType.Setter + public Builder cloudProvider(@Nullable String cloudProvider) { + this.cloudProvider = cloudProvider; + return this; + } + @CustomType.Setter + public Builder region(@Nullable String region) { + this.region = region; + return this; + } + public OnlineArchiveDataProcessRegion build() { + final var o = new OnlineArchiveDataProcessRegion(); + o.cloudProvider = cloudProvider; + o.region = region; + return o; + } + } +} diff --git a/sdk/nodejs/cluster.ts b/sdk/nodejs/cluster.ts index 806f7c1a..c9038829 100644 --- a/sdk/nodejs/cluster.ts +++ b/sdk/nodejs/cluster.ts @@ -47,7 +47,7 @@ export class Cluster extends pulumi.CustomResource { /** * If reconfiguration is necessary to regain a primary due to a regional outage, submit this field alongside your topology reconfiguration to request a new regional outage resistant topology. Forced reconfigurations during an outage of the majority of electable nodes carry a risk of data loss if replicated writes (even majority committed writes) have not been replicated to the new primary node. MongoDB Atlas docs contain more information. To proceed with an operation which carries that risk, set `acceptDataRisksAndForceReplicaSetReconfig` to the current date. Learn more about Reconfiguring a Replica Set during a regional outage [here](https://dochub.mongodb.org/core/regional-outage-reconfigure-replica-set). */ - public readonly acceptDataRisksAndForceReplicaSetReconfig!: pulumi.Output; + public readonly acceptDataRisksAndForceReplicaSetReconfig!: pulumi.Output; public readonly advancedConfiguration!: pulumi.Output; public readonly autoScalingComputeEnabled!: pulumi.Output; /** @@ -55,7 +55,7 @@ export class Cluster extends pulumi.CustomResource { * - If this option is enabled, you must specify a value for `providerSettings.autoScaling.compute.minInstanceSize` */ public readonly autoScalingComputeScaleDownEnabled!: pulumi.Output; - public readonly autoScalingDiskGbEnabled!: pulumi.Output; + public readonly autoScalingDiskGbEnabled!: pulumi.Output; /** * Cloud service provider on which the server for a multi-tenant cluster is provisioned. * diff --git a/sdk/nodejs/getOnlineArchive.ts b/sdk/nodejs/getOnlineArchive.ts index ec448db5..01d82493 100644 --- a/sdk/nodejs/getOnlineArchive.ts +++ b/sdk/nodejs/getOnlineArchive.ts @@ -30,6 +30,7 @@ import * as utilities from "./utilities"; * * `collectionType` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `dataExpirationRule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `dataProcessRegion` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partitionFields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -45,6 +46,10 @@ import * as utilities from "./utilities"; * ### Data Expiration Rule * * `expireAfterDays` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloudProvider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. @@ -100,6 +105,7 @@ export interface GetOnlineArchiveResult { readonly collectionType: string; readonly criterias: outputs.GetOnlineArchiveCriteria[]; readonly dataExpirationRules: outputs.GetOnlineArchiveDataExpirationRule[]; + readonly dataProcessRegions: outputs.GetOnlineArchiveDataProcessRegion[]; readonly dbName: string; /** * The provider-assigned unique ID for this managed resource. @@ -135,6 +141,7 @@ export interface GetOnlineArchiveResult { * * `collectionType` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `dataExpirationRule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `dataProcessRegion` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partitionFields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -150,6 +157,10 @@ export interface GetOnlineArchiveResult { * ### Data Expiration Rule * * `expireAfterDays` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloudProvider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. diff --git a/sdk/nodejs/getOnlineArchives.ts b/sdk/nodejs/getOnlineArchives.ts index 3828183c..44ae8108 100644 --- a/sdk/nodejs/getOnlineArchives.ts +++ b/sdk/nodejs/getOnlineArchives.ts @@ -29,6 +29,7 @@ import * as utilities from "./utilities"; * * `collectionType` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `dataExpirationRule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `dataProcessRegion` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partitionFields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -44,6 +45,10 @@ import * as utilities from "./utilities"; * ### Data Expiration Rule * * `expireAfterDays` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloudProvider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). @@ -122,6 +127,7 @@ export interface GetOnlineArchivesResult { * * `collectionType` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * * `criteria` - Criteria to use for archiving data. See criteria. * * `dataExpirationRule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * * `dataProcessRegion` - Settings to configure the region where you wish to store your archived data. See data process region. * * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * * `partitionFields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -137,6 +143,10 @@ export interface GetOnlineArchivesResult { * ### Data Expiration Rule * * `expireAfterDays` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. * + * ### Data Process Region + * * `cloudProvider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + * * ### Schedule * * * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). diff --git a/sdk/nodejs/onlineArchive.ts b/sdk/nodejs/onlineArchive.ts index d1bd917e..6e3fbf38 100644 --- a/sdk/nodejs/onlineArchive.ts +++ b/sdk/nodejs/onlineArchive.ts @@ -131,6 +131,10 @@ export class OnlineArchive extends pulumi.CustomResource { * Rule for specifying when data should be deleted from the archive. See data expiration rule. */ public readonly dataExpirationRule!: pulumi.Output; + /** + * Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + */ + public readonly dataProcessRegion!: pulumi.Output; /** * Name of the database that contains the collection. */ @@ -176,6 +180,7 @@ export class OnlineArchive extends pulumi.CustomResource { resourceInputs["collectionType"] = state ? state.collectionType : undefined; resourceInputs["criteria"] = state ? state.criteria : undefined; resourceInputs["dataExpirationRule"] = state ? state.dataExpirationRule : undefined; + resourceInputs["dataProcessRegion"] = state ? state.dataProcessRegion : undefined; resourceInputs["dbName"] = state ? state.dbName : undefined; resourceInputs["partitionFields"] = state ? state.partitionFields : undefined; resourceInputs["paused"] = state ? state.paused : undefined; @@ -205,6 +210,7 @@ export class OnlineArchive extends pulumi.CustomResource { resourceInputs["collectionType"] = args ? args.collectionType : undefined; resourceInputs["criteria"] = args ? args.criteria : undefined; resourceInputs["dataExpirationRule"] = args ? args.dataExpirationRule : undefined; + resourceInputs["dataProcessRegion"] = args ? args.dataProcessRegion : undefined; resourceInputs["dbName"] = args ? args.dbName : undefined; resourceInputs["partitionFields"] = args ? args.partitionFields : undefined; resourceInputs["paused"] = args ? args.paused : undefined; @@ -247,6 +253,10 @@ export interface OnlineArchiveState { * Rule for specifying when data should be deleted from the archive. See data expiration rule. */ dataExpirationRule?: pulumi.Input; + /** + * Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + */ + dataProcessRegion?: pulumi.Input; /** * Name of the database that contains the collection. */ @@ -298,6 +308,10 @@ export interface OnlineArchiveArgs { * Rule for specifying when data should be deleted from the archive. See data expiration rule. */ dataExpirationRule?: pulumi.Input; + /** + * Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + */ + dataProcessRegion?: pulumi.Input; /** * Name of the database that contains the collection. */ diff --git a/sdk/nodejs/searchDeployment.ts b/sdk/nodejs/searchDeployment.ts index 8232fb44..fa25c340 100644 --- a/sdk/nodejs/searchDeployment.ts +++ b/sdk/nodejs/searchDeployment.ts @@ -10,7 +10,8 @@ import * as utilities from "./utilities"; * `mongodbatlas.SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster. * * > **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation). - * **NOTE:** Only a single search deployment resource can be defined for each cluster. + * + * > **NOTE:** Only a single search deployment resource can be defined for each cluster. * * ## Import * diff --git a/sdk/nodejs/types/input.ts b/sdk/nodejs/types/input.ts index 5fb18038..abbace47 100644 --- a/sdk/nodejs/types/input.ts +++ b/sdk/nodejs/types/input.ts @@ -1946,6 +1946,17 @@ export interface OnlineArchiveDataExpirationRule { expireAfterDays: pulumi.Input; } +export interface OnlineArchiveDataProcessRegion { + /** + * Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + */ + cloudProvider?: pulumi.Input; + /** + * Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + */ + region?: pulumi.Input; +} + export interface OnlineArchivePartitionField { /** * Human-readable label that identifies the parameter that MongoDB Cloud uses to partition data. To specify a nested parameter, use the dot notation. diff --git a/sdk/nodejs/types/output.ts b/sdk/nodejs/types/output.ts index f8b268db..7999173f 100644 --- a/sdk/nodejs/types/output.ts +++ b/sdk/nodejs/types/output.ts @@ -5119,6 +5119,11 @@ export interface GetOnlineArchiveDataExpirationRule { expireAfterDays: number; } +export interface GetOnlineArchiveDataProcessRegion { + cloudProvider: string; + region: string; +} + export interface GetOnlineArchivePartitionField { fieldName: string; fieldType: string; @@ -5149,6 +5154,7 @@ export interface GetOnlineArchivesResult { collectionType: string; criterias: outputs.GetOnlineArchivesResultCriteria[]; dataExpirationRules: outputs.GetOnlineArchivesResultDataExpirationRule[]; + dataProcessRegions: outputs.GetOnlineArchivesResultDataProcessRegion[]; dbName: string; partitionFields: outputs.GetOnlineArchivesResultPartitionField[]; paused: boolean; @@ -5172,6 +5178,11 @@ export interface GetOnlineArchivesResultDataExpirationRule { expireAfterDays: number; } +export interface GetOnlineArchivesResultDataProcessRegion { + cloudProvider: string; + region: string; +} + export interface GetOnlineArchivesResultPartitionField { fieldName: string; fieldType: string; @@ -5777,6 +5788,17 @@ export interface OnlineArchiveDataExpirationRule { expireAfterDays: number; } +export interface OnlineArchiveDataProcessRegion { + /** + * Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + */ + cloudProvider: string; + /** + * Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + */ + region: string; +} + export interface OnlineArchivePartitionField { /** * Human-readable label that identifies the parameter that MongoDB Cloud uses to partition data. To specify a nested parameter, use the dot notation. diff --git a/sdk/python/pulumi_mongodbatlas/_inputs.py b/sdk/python/pulumi_mongodbatlas/_inputs.py index 487921cf..d6fe32c8 100644 --- a/sdk/python/pulumi_mongodbatlas/_inputs.py +++ b/sdk/python/pulumi_mongodbatlas/_inputs.py @@ -98,6 +98,7 @@ 'LdapVerifyValidationArgs', 'OnlineArchiveCriteriaArgs', 'OnlineArchiveDataExpirationRuleArgs', + 'OnlineArchiveDataProcessRegionArgs', 'OnlineArchivePartitionFieldArgs', 'OnlineArchiveScheduleArgs', 'PrivateLinkEndpointServiceEndpointArgs', @@ -6428,6 +6429,45 @@ def expire_after_days(self, value: pulumi.Input[int]): pulumi.set(self, "expire_after_days", value) +@pulumi.input_type +class OnlineArchiveDataProcessRegionArgs: + def __init__(__self__, *, + cloud_provider: Optional[pulumi.Input[str]] = None, + region: Optional[pulumi.Input[str]] = None): + """ + :param pulumi.Input[str] cloud_provider: Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + :param pulumi.Input[str] region: Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + """ + if cloud_provider is not None: + pulumi.set(__self__, "cloud_provider", cloud_provider) + if region is not None: + pulumi.set(__self__, "region", region) + + @property + @pulumi.getter(name="cloudProvider") + def cloud_provider(self) -> Optional[pulumi.Input[str]]: + """ + Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + """ + return pulumi.get(self, "cloud_provider") + + @cloud_provider.setter + def cloud_provider(self, value: Optional[pulumi.Input[str]]): + pulumi.set(self, "cloud_provider", value) + + @property + @pulumi.getter + def region(self) -> Optional[pulumi.Input[str]]: + """ + Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + """ + return pulumi.get(self, "region") + + @region.setter + def region(self, value: Optional[pulumi.Input[str]]): + pulumi.set(self, "region", value) + + @pulumi.input_type class OnlineArchivePartitionFieldArgs: def __init__(__self__, *, diff --git a/sdk/python/pulumi_mongodbatlas/cluster.py b/sdk/python/pulumi_mongodbatlas/cluster.py index 03eaf374..b941f6cc 100644 --- a/sdk/python/pulumi_mongodbatlas/cluster.py +++ b/sdk/python/pulumi_mongodbatlas/cluster.py @@ -1830,7 +1830,7 @@ def get(resource_name: str, @property @pulumi.getter(name="acceptDataRisksAndForceReplicaSetReconfig") - def accept_data_risks_and_force_replica_set_reconfig(self) -> pulumi.Output[Optional[str]]: + def accept_data_risks_and_force_replica_set_reconfig(self) -> pulumi.Output[str]: """ If reconfiguration is necessary to regain a primary due to a regional outage, submit this field alongside your topology reconfiguration to request a new regional outage resistant topology. Forced reconfigurations during an outage of the majority of electable nodes carry a risk of data loss if replicated writes (even majority committed writes) have not been replicated to the new primary node. MongoDB Atlas docs contain more information. To proceed with an operation which carries that risk, set `accept_data_risks_and_force_replica_set_reconfig` to the current date. Learn more about Reconfiguring a Replica Set during a regional outage [here](https://dochub.mongodb.org/core/regional-outage-reconfigure-replica-set). """ @@ -1857,7 +1857,7 @@ def auto_scaling_compute_scale_down_enabled(self) -> pulumi.Output[bool]: @property @pulumi.getter(name="autoScalingDiskGbEnabled") - def auto_scaling_disk_gb_enabled(self) -> pulumi.Output[Optional[bool]]: + def auto_scaling_disk_gb_enabled(self) -> pulumi.Output[bool]: return pulumi.get(self, "auto_scaling_disk_gb_enabled") @property diff --git a/sdk/python/pulumi_mongodbatlas/get_online_archive.py b/sdk/python/pulumi_mongodbatlas/get_online_archive.py index 7d2f48e6..82635c51 100644 --- a/sdk/python/pulumi_mongodbatlas/get_online_archive.py +++ b/sdk/python/pulumi_mongodbatlas/get_online_archive.py @@ -22,7 +22,7 @@ class GetOnlineArchiveResult: """ A collection of values returned by getOnlineArchive. """ - def __init__(__self__, archive_id=None, cluster_name=None, coll_name=None, collection_type=None, criterias=None, data_expiration_rules=None, db_name=None, id=None, partition_fields=None, paused=None, project_id=None, schedules=None, state=None): + def __init__(__self__, archive_id=None, cluster_name=None, coll_name=None, collection_type=None, criterias=None, data_expiration_rules=None, data_process_regions=None, db_name=None, id=None, partition_fields=None, paused=None, project_id=None, schedules=None, state=None): if archive_id and not isinstance(archive_id, str): raise TypeError("Expected argument 'archive_id' to be a str") pulumi.set(__self__, "archive_id", archive_id) @@ -41,6 +41,9 @@ def __init__(__self__, archive_id=None, cluster_name=None, coll_name=None, colle if data_expiration_rules and not isinstance(data_expiration_rules, list): raise TypeError("Expected argument 'data_expiration_rules' to be a list") pulumi.set(__self__, "data_expiration_rules", data_expiration_rules) + if data_process_regions and not isinstance(data_process_regions, list): + raise TypeError("Expected argument 'data_process_regions' to be a list") + pulumi.set(__self__, "data_process_regions", data_process_regions) if db_name and not isinstance(db_name, str): raise TypeError("Expected argument 'db_name' to be a str") pulumi.set(__self__, "db_name", db_name) @@ -93,6 +96,11 @@ def criterias(self) -> Sequence['outputs.GetOnlineArchiveCriteriaResult']: def data_expiration_rules(self) -> Sequence['outputs.GetOnlineArchiveDataExpirationRuleResult']: return pulumi.get(self, "data_expiration_rules") + @property + @pulumi.getter(name="dataProcessRegions") + def data_process_regions(self) -> Sequence['outputs.GetOnlineArchiveDataProcessRegionResult']: + return pulumi.get(self, "data_process_regions") + @property @pulumi.getter(name="dbName") def db_name(self) -> str: @@ -144,6 +152,7 @@ def __await__(self): collection_type=self.collection_type, criterias=self.criterias, data_expiration_rules=self.data_expiration_rules, + data_process_regions=self.data_process_regions, db_name=self.db_name, id=self.id, partition_fields=self.partition_fields, @@ -179,6 +188,7 @@ def get_online_archive(archive_id: Optional[str] = None, * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * `criteria` - Criteria to use for archiving data. See criteria. * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -194,6 +204,10 @@ def get_online_archive(archive_id: Optional[str] = None, ### Data Expiration Rule * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. + ### Data Process Region + * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + ### Schedule * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. @@ -230,6 +244,7 @@ def get_online_archive(archive_id: Optional[str] = None, collection_type=pulumi.get(__ret__, 'collection_type'), criterias=pulumi.get(__ret__, 'criterias'), data_expiration_rules=pulumi.get(__ret__, 'data_expiration_rules'), + data_process_regions=pulumi.get(__ret__, 'data_process_regions'), db_name=pulumi.get(__ret__, 'db_name'), id=pulumi.get(__ret__, 'id'), partition_fields=pulumi.get(__ret__, 'partition_fields'), @@ -266,6 +281,7 @@ def get_online_archive_output(archive_id: Optional[pulumi.Input[str]] = None, * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * `criteria` - Criteria to use for archiving data. See criteria. * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -281,6 +297,10 @@ def get_online_archive_output(archive_id: Optional[pulumi.Input[str]] = None, ### Data Expiration Rule * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. + ### Data Process Region + * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + ### Schedule * `type` - Type of schedule. Valid values: `DEFAULT`, `DAILY`, `MONTHLY`, `WEEKLY`. diff --git a/sdk/python/pulumi_mongodbatlas/get_online_archives.py b/sdk/python/pulumi_mongodbatlas/get_online_archives.py index ba170db5..a804284d 100644 --- a/sdk/python/pulumi_mongodbatlas/get_online_archives.py +++ b/sdk/python/pulumi_mongodbatlas/get_online_archives.py @@ -105,6 +105,7 @@ def get_online_archives(cluster_name: Optional[str] = None, * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * `criteria` - Criteria to use for archiving data. See criteria. * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -120,6 +121,10 @@ def get_online_archives(cluster_name: Optional[str] = None, ### Data Expiration Rule * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. + ### Data Process Region + * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + ### Schedule * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). @@ -182,6 +187,7 @@ def get_online_archives_output(cluster_name: Optional[pulumi.Input[str]] = None, * `collection_type` - Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". * `criteria` - Criteria to use for archiving data. See criteria. * `data_expiration_rule` - Rule for specifying when data should be deleted from the archive. See data expiration rule. + * `data_process_region` - Settings to configure the region where you wish to store your archived data. See data process region. * `schedule` - Regular frequency and duration when archiving process occurs. See schedule. * `partition_fields` - Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. * `paused` - State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -197,6 +203,10 @@ def get_online_archives_output(cluster_name: Optional[pulumi.Input[str]] = None, ### Data Expiration Rule * `expire_after_days` - Number of days used in the date criteria for nominating documents for deletion. Value must be between 7 and 9215. + ### Data Process Region + * `cloud_provider` - Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + * `region` - Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + ### Schedule * `type` - Type of schedule (`DAILY`, `MONTHLY`, `WEEKLY`). diff --git a/sdk/python/pulumi_mongodbatlas/online_archive.py b/sdk/python/pulumi_mongodbatlas/online_archive.py index 14b73be4..2ee74690 100644 --- a/sdk/python/pulumi_mongodbatlas/online_archive.py +++ b/sdk/python/pulumi_mongodbatlas/online_archive.py @@ -23,6 +23,7 @@ def __init__(__self__, *, project_id: pulumi.Input[str], collection_type: Optional[pulumi.Input[str]] = None, data_expiration_rule: Optional[pulumi.Input['OnlineArchiveDataExpirationRuleArgs']] = None, + data_process_region: Optional[pulumi.Input['OnlineArchiveDataProcessRegionArgs']] = None, partition_fields: Optional[pulumi.Input[Sequence[pulumi.Input['OnlineArchivePartitionFieldArgs']]]] = None, paused: Optional[pulumi.Input[bool]] = None, schedule: Optional[pulumi.Input['OnlineArchiveScheduleArgs']] = None, @@ -36,6 +37,7 @@ def __init__(__self__, *, :param pulumi.Input[str] project_id: The unique ID for the project :param pulumi.Input[str] collection_type: Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". :param pulumi.Input['OnlineArchiveDataExpirationRuleArgs'] data_expiration_rule: Rule for specifying when data should be deleted from the archive. See data expiration rule. + :param pulumi.Input['OnlineArchiveDataProcessRegionArgs'] data_process_region: Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. :param pulumi.Input[Sequence[pulumi.Input['OnlineArchivePartitionFieldArgs']]] partition_fields: Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. :param pulumi.Input[bool] paused: State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. :param pulumi.Input['OnlineArchiveScheduleArgs'] schedule: Regular frequency and duration when archiving process occurs. See schedule. @@ -49,6 +51,8 @@ def __init__(__self__, *, pulumi.set(__self__, "collection_type", collection_type) if data_expiration_rule is not None: pulumi.set(__self__, "data_expiration_rule", data_expiration_rule) + if data_process_region is not None: + pulumi.set(__self__, "data_process_region", data_process_region) if partition_fields is not None: pulumi.set(__self__, "partition_fields", partition_fields) if paused is not None: @@ -142,6 +146,18 @@ def data_expiration_rule(self) -> Optional[pulumi.Input['OnlineArchiveDataExpira def data_expiration_rule(self, value: Optional[pulumi.Input['OnlineArchiveDataExpirationRuleArgs']]): pulumi.set(self, "data_expiration_rule", value) + @property + @pulumi.getter(name="dataProcessRegion") + def data_process_region(self) -> Optional[pulumi.Input['OnlineArchiveDataProcessRegionArgs']]: + """ + Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + """ + return pulumi.get(self, "data_process_region") + + @data_process_region.setter + def data_process_region(self, value: Optional[pulumi.Input['OnlineArchiveDataProcessRegionArgs']]): + pulumi.set(self, "data_process_region", value) + @property @pulumi.getter(name="partitionFields") def partition_fields(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['OnlineArchivePartitionFieldArgs']]]]: @@ -197,6 +213,7 @@ def __init__(__self__, *, collection_type: Optional[pulumi.Input[str]] = None, criteria: Optional[pulumi.Input['OnlineArchiveCriteriaArgs']] = None, data_expiration_rule: Optional[pulumi.Input['OnlineArchiveDataExpirationRuleArgs']] = None, + data_process_region: Optional[pulumi.Input['OnlineArchiveDataProcessRegionArgs']] = None, db_name: Optional[pulumi.Input[str]] = None, partition_fields: Optional[pulumi.Input[Sequence[pulumi.Input['OnlineArchivePartitionFieldArgs']]]] = None, paused: Optional[pulumi.Input[bool]] = None, @@ -212,6 +229,7 @@ def __init__(__self__, *, :param pulumi.Input[str] collection_type: Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". :param pulumi.Input['OnlineArchiveCriteriaArgs'] criteria: Criteria to use for archiving data. See criteria. :param pulumi.Input['OnlineArchiveDataExpirationRuleArgs'] data_expiration_rule: Rule for specifying when data should be deleted from the archive. See data expiration rule. + :param pulumi.Input['OnlineArchiveDataProcessRegionArgs'] data_process_region: Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. :param pulumi.Input[str] db_name: Name of the database that contains the collection. :param pulumi.Input[Sequence[pulumi.Input['OnlineArchivePartitionFieldArgs']]] partition_fields: Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. :param pulumi.Input[bool] paused: State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -231,6 +249,8 @@ def __init__(__self__, *, pulumi.set(__self__, "criteria", criteria) if data_expiration_rule is not None: pulumi.set(__self__, "data_expiration_rule", data_expiration_rule) + if data_process_region is not None: + pulumi.set(__self__, "data_process_region", data_process_region) if db_name is not None: pulumi.set(__self__, "db_name", db_name) if partition_fields is not None: @@ -318,6 +338,18 @@ def data_expiration_rule(self) -> Optional[pulumi.Input['OnlineArchiveDataExpira def data_expiration_rule(self, value: Optional[pulumi.Input['OnlineArchiveDataExpirationRuleArgs']]): pulumi.set(self, "data_expiration_rule", value) + @property + @pulumi.getter(name="dataProcessRegion") + def data_process_region(self) -> Optional[pulumi.Input['OnlineArchiveDataProcessRegionArgs']]: + """ + Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + """ + return pulumi.get(self, "data_process_region") + + @data_process_region.setter + def data_process_region(self, value: Optional[pulumi.Input['OnlineArchiveDataProcessRegionArgs']]): + pulumi.set(self, "data_process_region", value) + @property @pulumi.getter(name="dbName") def db_name(self) -> Optional[pulumi.Input[str]]: @@ -410,6 +442,7 @@ def __init__(__self__, collection_type: Optional[pulumi.Input[str]] = None, criteria: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveCriteriaArgs']]] = None, data_expiration_rule: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveDataExpirationRuleArgs']]] = None, + data_process_region: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveDataProcessRegionArgs']]] = None, db_name: Optional[pulumi.Input[str]] = None, partition_fields: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['OnlineArchivePartitionFieldArgs']]]]] = None, paused: Optional[pulumi.Input[bool]] = None, @@ -495,6 +528,7 @@ def __init__(__self__, :param pulumi.Input[str] collection_type: Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". :param pulumi.Input[pulumi.InputType['OnlineArchiveCriteriaArgs']] criteria: Criteria to use for archiving data. See criteria. :param pulumi.Input[pulumi.InputType['OnlineArchiveDataExpirationRuleArgs']] data_expiration_rule: Rule for specifying when data should be deleted from the archive. See data expiration rule. + :param pulumi.Input[pulumi.InputType['OnlineArchiveDataProcessRegionArgs']] data_process_region: Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. :param pulumi.Input[str] db_name: Name of the database that contains the collection. :param pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['OnlineArchivePartitionFieldArgs']]]] partition_fields: Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. :param pulumi.Input[bool] paused: State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -598,6 +632,7 @@ def _internal_init(__self__, collection_type: Optional[pulumi.Input[str]] = None, criteria: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveCriteriaArgs']]] = None, data_expiration_rule: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveDataExpirationRuleArgs']]] = None, + data_process_region: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveDataProcessRegionArgs']]] = None, db_name: Optional[pulumi.Input[str]] = None, partition_fields: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['OnlineArchivePartitionFieldArgs']]]]] = None, paused: Optional[pulumi.Input[bool]] = None, @@ -624,6 +659,7 @@ def _internal_init(__self__, raise TypeError("Missing required property 'criteria'") __props__.__dict__["criteria"] = criteria __props__.__dict__["data_expiration_rule"] = data_expiration_rule + __props__.__dict__["data_process_region"] = data_process_region if db_name is None and not opts.urn: raise TypeError("Missing required property 'db_name'") __props__.__dict__["db_name"] = db_name @@ -652,6 +688,7 @@ def get(resource_name: str, collection_type: Optional[pulumi.Input[str]] = None, criteria: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveCriteriaArgs']]] = None, data_expiration_rule: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveDataExpirationRuleArgs']]] = None, + data_process_region: Optional[pulumi.Input[pulumi.InputType['OnlineArchiveDataProcessRegionArgs']]] = None, db_name: Optional[pulumi.Input[str]] = None, partition_fields: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['OnlineArchivePartitionFieldArgs']]]]] = None, paused: Optional[pulumi.Input[bool]] = None, @@ -672,6 +709,7 @@ def get(resource_name: str, :param pulumi.Input[str] collection_type: Type of MongoDB collection that you want to return. This value can be "TIMESERIES" or "STANDARD". Default is "STANDARD". :param pulumi.Input[pulumi.InputType['OnlineArchiveCriteriaArgs']] criteria: Criteria to use for archiving data. See criteria. :param pulumi.Input[pulumi.InputType['OnlineArchiveDataExpirationRuleArgs']] data_expiration_rule: Rule for specifying when data should be deleted from the archive. See data expiration rule. + :param pulumi.Input[pulumi.InputType['OnlineArchiveDataProcessRegionArgs']] data_process_region: Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. :param pulumi.Input[str] db_name: Name of the database that contains the collection. :param pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['OnlineArchivePartitionFieldArgs']]]] partition_fields: Fields to use to partition data. You can specify up to two frequently queried fields to use for partitioning data. Queries that don’t contain the specified fields require a full collection scan of all archived documents, which takes longer and increases your costs. To learn more about how partition improves query performance, see [Data Structure in S3](https://docs.mongodb.com/datalake/admin/optimize-query-performance/#data-structure-in-s3). The value of a partition field can be up to a maximum of 700 characters. Documents with values exceeding 700 characters are not archived. See partition fields. :param pulumi.Input[bool] paused: State of the online archive. This is required for pausing an active online archive or resuming a paused online archive. If the collection has another active online archive, the resume request fails. @@ -689,6 +727,7 @@ def get(resource_name: str, __props__.__dict__["collection_type"] = collection_type __props__.__dict__["criteria"] = criteria __props__.__dict__["data_expiration_rule"] = data_expiration_rule + __props__.__dict__["data_process_region"] = data_process_region __props__.__dict__["db_name"] = db_name __props__.__dict__["partition_fields"] = partition_fields __props__.__dict__["paused"] = paused @@ -746,6 +785,14 @@ def data_expiration_rule(self) -> pulumi.Output[Optional['outputs.OnlineArchiveD """ return pulumi.get(self, "data_expiration_rule") + @property + @pulumi.getter(name="dataProcessRegion") + def data_process_region(self) -> pulumi.Output['outputs.OnlineArchiveDataProcessRegion']: + """ + Settings to configure the region where you wish to store your archived data. See data process region. This field is immutable hence cannot be updated. + """ + return pulumi.get(self, "data_process_region") + @property @pulumi.getter(name="dbName") def db_name(self) -> pulumi.Output[str]: diff --git a/sdk/python/pulumi_mongodbatlas/outputs.py b/sdk/python/pulumi_mongodbatlas/outputs.py index d9ca577d..3d60878f 100644 --- a/sdk/python/pulumi_mongodbatlas/outputs.py +++ b/sdk/python/pulumi_mongodbatlas/outputs.py @@ -99,6 +99,7 @@ 'LdapVerifyValidation', 'OnlineArchiveCriteria', 'OnlineArchiveDataExpirationRule', + 'OnlineArchiveDataProcessRegion', 'OnlineArchivePartitionField', 'OnlineArchiveSchedule', 'PrivateLinkEndpointServiceEndpoint', @@ -300,11 +301,13 @@ 'GetNetworkPeeringsResultResult', 'GetOnlineArchiveCriteriaResult', 'GetOnlineArchiveDataExpirationRuleResult', + 'GetOnlineArchiveDataProcessRegionResult', 'GetOnlineArchivePartitionFieldResult', 'GetOnlineArchiveScheduleResult', 'GetOnlineArchivesResultResult', 'GetOnlineArchivesResultCriteriaResult', 'GetOnlineArchivesResultDataExpirationRuleResult', + 'GetOnlineArchivesResultDataProcessRegionResult', 'GetOnlineArchivesResultPartitionFieldResult', 'GetOnlineArchivesResultScheduleResult', 'GetOrganizationLinkResult', @@ -6722,6 +6725,54 @@ def expire_after_days(self) -> int: return pulumi.get(self, "expire_after_days") +@pulumi.output_type +class OnlineArchiveDataProcessRegion(dict): + @staticmethod + def __key_warning(key: str): + suggest = None + if key == "cloudProvider": + suggest = "cloud_provider" + + if suggest: + pulumi.log.warn(f"Key '{key}' not found in OnlineArchiveDataProcessRegion. Access the value via the '{suggest}' property getter instead.") + + def __getitem__(self, key: str) -> Any: + OnlineArchiveDataProcessRegion.__key_warning(key) + return super().__getitem__(key) + + def get(self, key: str, default = None) -> Any: + OnlineArchiveDataProcessRegion.__key_warning(key) + return super().get(key, default) + + def __init__(__self__, *, + cloud_provider: Optional[str] = None, + region: Optional[str] = None): + """ + :param str cloud_provider: Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + :param str region: Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + """ + if cloud_provider is not None: + pulumi.set(__self__, "cloud_provider", cloud_provider) + if region is not None: + pulumi.set(__self__, "region", region) + + @property + @pulumi.getter(name="cloudProvider") + def cloud_provider(self) -> Optional[str]: + """ + Human-readable label that identifies the Cloud service provider where you wish to store your archived data. + """ + return pulumi.get(self, "cloud_provider") + + @property + @pulumi.getter + def region(self) -> Optional[str]: + """ + Human-readable label that identifies the geographic location of the region where you wish to store your archived data. For allowed values, see [MongoDB Atlas API documentation](https://www.mongodb.com/docs/atlas/reference/api-resources-spec/v2/#tag/Online-Archive/operation/createOnlineArchive) + """ + return pulumi.get(self, "region") + + @pulumi.output_type class OnlineArchivePartitionField(dict): @staticmethod @@ -18352,6 +18403,25 @@ def expire_after_days(self) -> int: return pulumi.get(self, "expire_after_days") +@pulumi.output_type +class GetOnlineArchiveDataProcessRegionResult(dict): + def __init__(__self__, *, + cloud_provider: str, + region: str): + pulumi.set(__self__, "cloud_provider", cloud_provider) + pulumi.set(__self__, "region", region) + + @property + @pulumi.getter(name="cloudProvider") + def cloud_provider(self) -> str: + return pulumi.get(self, "cloud_provider") + + @property + @pulumi.getter + def region(self) -> str: + return pulumi.get(self, "region") + + @pulumi.output_type class GetOnlineArchivePartitionFieldResult(dict): def __init__(__self__, *, @@ -18443,6 +18513,7 @@ def __init__(__self__, *, collection_type: str, criterias: Sequence['outputs.GetOnlineArchivesResultCriteriaResult'], data_expiration_rules: Sequence['outputs.GetOnlineArchivesResultDataExpirationRuleResult'], + data_process_regions: Sequence['outputs.GetOnlineArchivesResultDataProcessRegionResult'], db_name: str, partition_fields: Sequence['outputs.GetOnlineArchivesResultPartitionFieldResult'], paused: bool, @@ -18463,6 +18534,7 @@ def __init__(__self__, *, pulumi.set(__self__, "collection_type", collection_type) pulumi.set(__self__, "criterias", criterias) pulumi.set(__self__, "data_expiration_rules", data_expiration_rules) + pulumi.set(__self__, "data_process_regions", data_process_regions) pulumi.set(__self__, "db_name", db_name) pulumi.set(__self__, "partition_fields", partition_fields) pulumi.set(__self__, "paused", paused) @@ -18507,6 +18579,11 @@ def criterias(self) -> Sequence['outputs.GetOnlineArchivesResultCriteriaResult'] def data_expiration_rules(self) -> Sequence['outputs.GetOnlineArchivesResultDataExpirationRuleResult']: return pulumi.get(self, "data_expiration_rules") + @property + @pulumi.getter(name="dataProcessRegions") + def data_process_regions(self) -> Sequence['outputs.GetOnlineArchivesResultDataProcessRegionResult']: + return pulumi.get(self, "data_process_regions") + @property @pulumi.getter(name="dbName") def db_name(self) -> str: @@ -18593,6 +18670,25 @@ def expire_after_days(self) -> int: return pulumi.get(self, "expire_after_days") +@pulumi.output_type +class GetOnlineArchivesResultDataProcessRegionResult(dict): + def __init__(__self__, *, + cloud_provider: str, + region: str): + pulumi.set(__self__, "cloud_provider", cloud_provider) + pulumi.set(__self__, "region", region) + + @property + @pulumi.getter(name="cloudProvider") + def cloud_provider(self) -> str: + return pulumi.get(self, "cloud_provider") + + @property + @pulumi.getter + def region(self) -> str: + return pulumi.get(self, "region") + + @pulumi.output_type class GetOnlineArchivesResultPartitionFieldResult(dict): def __init__(__self__, *, diff --git a/sdk/python/pulumi_mongodbatlas/search_deployment.py b/sdk/python/pulumi_mongodbatlas/search_deployment.py index a6c06f3a..91452d03 100644 --- a/sdk/python/pulumi_mongodbatlas/search_deployment.py +++ b/sdk/python/pulumi_mongodbatlas/search_deployment.py @@ -176,7 +176,8 @@ def __init__(__self__, `SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster. > **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation). - **NOTE:** Only a single search deployment resource can be defined for each cluster. + + > **NOTE:** Only a single search deployment resource can be defined for each cluster. ## Import @@ -203,7 +204,8 @@ def __init__(__self__, `SearchDeployment` provides a Search Deployment resource. The resource lets you create, edit and delete dedicated search nodes in a cluster. > **NOTE:** For details on supported cloud providers and existing limitations you can visit the [Search Node Documentation](https://www.mongodb.com/docs/atlas/cluster-config/multi-cloud-distribution/#search-nodes-for-workload-isolation). - **NOTE:** Only a single search deployment resource can be defined for each cluster. + + > **NOTE:** Only a single search deployment resource can be defined for each cluster. ## Import