diff --git a/sdk/resourcemanager/datafactory/armdatafactory/CHANGELOG.md b/sdk/resourcemanager/datafactory/armdatafactory/CHANGELOG.md index 78a2241e1392..7f4cecfc7312 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/CHANGELOG.md +++ b/sdk/resourcemanager/datafactory/armdatafactory/CHANGELOG.md @@ -1,5 +1,68 @@ # Release History +## 5.0.0 (2024-01-17) +### Breaking Changes + +- Field `Pwd` of struct `MariaDBLinkedServiceTypeProperties` has been removed + +### Features Added + +- New enum type `SalesforceV2SinkWriteBehavior` with values `SalesforceV2SinkWriteBehaviorInsert`, `SalesforceV2SinkWriteBehaviorUpsert` +- New enum type `SnowflakeAuthenticationType` with values `SnowflakeAuthenticationTypeAADServicePrincipal`, `SnowflakeAuthenticationTypeBasic`, `SnowflakeAuthenticationTypeKeyPair` +- New function `*SalesforceServiceCloudV2LinkedService.GetLinkedService() *LinkedService` +- New function `*SalesforceServiceCloudV2ObjectDataset.GetDataset() *Dataset` +- New function `*SalesforceServiceCloudV2Sink.GetCopySink() *CopySink` +- New function `*SalesforceServiceCloudV2Source.GetCopySource() *CopySource` +- New function `*SalesforceV2LinkedService.GetLinkedService() *LinkedService` +- New function `*SalesforceV2ObjectDataset.GetDataset() *Dataset` +- New function `*SalesforceV2Sink.GetCopySink() *CopySink` +- New function `*SalesforceV2Source.GetCopySource() *CopySource` +- New function `*SalesforceV2Source.GetTabularSource() *TabularSource` +- New function `*SnowflakeV2Dataset.GetDataset() *Dataset` +- New function `*SnowflakeV2LinkedService.GetLinkedService() *LinkedService` +- New function `*SnowflakeV2Sink.GetCopySink() *CopySink` +- New function `*SnowflakeV2Source.GetCopySource() *CopySource` +- New function `*WarehouseLinkedService.GetLinkedService() *LinkedService` +- New function `*WarehouseSink.GetCopySink() *CopySink` +- New function `*WarehouseSource.GetCopySource() *CopySource` +- New function `*WarehouseSource.GetTabularSource() *TabularSource` +- New function `*WarehouseTableDataset.GetDataset() *Dataset` +- New struct `SalesforceServiceCloudV2LinkedService` +- New struct `SalesforceServiceCloudV2LinkedServiceTypeProperties` +- New struct `SalesforceServiceCloudV2ObjectDataset` +- New struct `SalesforceServiceCloudV2ObjectDatasetTypeProperties` +- New struct `SalesforceServiceCloudV2Sink` +- New struct `SalesforceServiceCloudV2Source` +- New struct `SalesforceV2LinkedService` +- New struct `SalesforceV2LinkedServiceTypeProperties` +- New struct `SalesforceV2ObjectDataset` +- New struct `SalesforceV2ObjectDatasetTypeProperties` +- New struct `SalesforceV2Sink` +- New struct `SalesforceV2Source` +- New struct `SnowflakeLinkedV2ServiceTypeProperties` +- New struct `SnowflakeV2Dataset` +- New struct `SnowflakeV2LinkedService` +- New struct `SnowflakeV2Sink` +- New struct `SnowflakeV2Source` +- New struct `WarehouseLinkedService` +- New struct `WarehouseLinkedServiceTypeProperties` +- New struct `WarehouseSink` +- New struct `WarehouseSource` +- New struct `WarehouseTableDataset` +- New struct `WarehouseTableDatasetTypeProperties` +- New field `Metadata` in struct `AzureBlobFSWriteSettings` +- New field `Metadata` in struct `AzureBlobStorageWriteSettings` +- New field `Metadata` in struct `AzureDataLakeStoreWriteSettings` +- New field `Metadata` in struct `AzureFileStorageWriteSettings` +- New field `Metadata` in struct `FileServerWriteSettings` +- New field `Metadata` in struct `LakeHouseWriteSettings` +- New field `Database`, `DriverVersion`, `Password`, `Port`, `Server`, `Username` in struct `MariaDBLinkedServiceTypeProperties` +- New field `Database`, `DriverVersion`, `Port`, `SSLMode`, `Server`, `UseSystemTrustStore`, `Username` in struct `MySQLLinkedServiceTypeProperties` +- New field `Metadata` in struct `SftpWriteSettings` +- New field `Metadata` in struct `StoreWriteSettings` +- New field `HTTPRequestTimeout`, `TurnOffAsync` in struct `WebActivityTypeProperties` + + ## 4.0.0 (2023-12-22) ### Breaking Changes diff --git a/sdk/resourcemanager/datafactory/armdatafactory/README.md b/sdk/resourcemanager/datafactory/armdatafactory/README.md index 17ed77923d27..386b3ce984e0 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/README.md +++ b/sdk/resourcemanager/datafactory/armdatafactory/README.md @@ -1,6 +1,6 @@ # Azure Data Factory Module for Go -[![PkgGoDev](https://pkg.go.dev/badge/github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4)](https://pkg.go.dev/github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4) +[![PkgGoDev](https://pkg.go.dev/badge/github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5)](https://pkg.go.dev/github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5) The `armdatafactory` module provides operations for working with Azure Data Factory. @@ -20,7 +20,7 @@ This project uses [Go modules](https://github.com/golang/go/wiki/Modules) for ve Install the Azure Data Factory module: ```sh -go get github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4 +go get github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5 ``` ## Authorization @@ -57,7 +57,7 @@ clientFactory, err := armdatafactory.NewClientFactory(, cred, & A client groups a set of related APIs, providing access to its functionality. Create one or more clients to access the APIs you require using client factory. ```go -client := clientFactory.NewTriggersClient() +client := clientFactory.NewActivityRunsClient() ``` ## Fakes diff --git a/sdk/resourcemanager/datafactory/armdatafactory/activityruns_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/activityruns_client_example_test.go deleted file mode 100644 index 5822f4655590..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/activityruns_client_example_test.go +++ /dev/null @@ -1,138 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "time" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ActivityRuns_QueryByPipelineRun.json -func ExampleActivityRunsClient_QueryByPipelineRun() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewActivityRunsClient().QueryByPipelineRun(ctx, "exampleResourceGroup", "exampleFactoryName", "2f7fdb90-5df1-4b8e-ac2f-064cfa58202b", armdatafactory.RunFilterParameters{ - LastUpdatedAfter: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:36:44.334Z"); return t }()), - LastUpdatedBefore: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:49:48.368Z"); return t }()), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ActivityRunsQueryResponse = armdatafactory.ActivityRunsQueryResponse{ - // Value: []*armdatafactory.ActivityRun{ - // { - // AdditionalProperties: map[string]any{ - // "retryAttempt": nil, - // "userProperties": map[string]any{ - // }, - // }, - // ActivityName: to.Ptr("ExampleForeachActivity"), - // ActivityRunEnd: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:38:11.544Z"); return t}()), - // ActivityRunID: to.Ptr("f30c5514-fb85-43ed-9fa4-768d42e58680"), - // ActivityRunStart: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:37:49.480Z"); return t}()), - // ActivityType: to.Ptr("ForEach"), - // DurationInMs: to.Ptr[int32](22064), - // Error: map[string]any{ - // "errorCode": "", - // "failureType": "", - // "message": "", - // "target": "ExampleForeachActivity", - // }, - // Input: map[string]any{ - // }, - // LinkedServiceName: to.Ptr(""), - // Output: map[string]any{ - // }, - // PipelineName: to.Ptr("examplePipeline"), - // PipelineRunID: to.Ptr("2f7fdb90-5df1-4b8e-ac2f-064cfa58202b"), - // Status: to.Ptr("Succeeded"), - // }, - // { - // AdditionalProperties: map[string]any{ - // "retryAttempt": nil, - // "userProperties": map[string]any{ - // }, - // }, - // ActivityName: to.Ptr("ExampleCopyActivity"), - // ActivityRunEnd: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:38:07.418Z"); return t}()), - // ActivityRunID: to.Ptr("a96678c8-7167-4f00-b629-afccfbad4e51"), - // ActivityRunStart: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:37:50.246Z"); return t}()), - // ActivityType: to.Ptr("Copy"), - // DurationInMs: to.Ptr[int32](17172), - // Error: map[string]any{ - // "errorCode": "", - // "failureType": "", - // "message": "", - // "target": "ExampleCopyActivity", - // }, - // Input: map[string]any{ - // "dataIntegrationUnits": float64(32), - // "sink":map[string]any{ - // "type": "BlobSink", - // }, - // "source":map[string]any{ - // "type": "BlobSource", - // }, - // }, - // LinkedServiceName: to.Ptr(""), - // Output: map[string]any{ - // "copyDuration": float64(6), - // "dataRead": float64(142000), - // "dataWritten": float64(142000), - // "effectiveIntegrationRuntime": "DefaultIntegrationRuntime (East US)", - // "errors":[]any{ - // }, - // "executionDetails":[]any{ - // map[string]any{ - // "detailedDurations":map[string]any{ - // "queuingDuration": float64(4), - // "transferDuration": float64(2), - // }, - // "duration": float64(6), - // "sink":map[string]any{ - // "type": "AzureBlob", - // }, - // "source":map[string]any{ - // "type": "AzureBlob", - // }, - // "start": "2018-06-16T00:37:50.68834Z", - // "status": "Succeeded", - // "usedCloudDataMovementUnits": float64(4), - // "usedParallelCopies": float64(1), - // }, - // }, - // "filesRead": float64(1), - // "filesWritten": float64(1), - // "throughput": float64(23.112), - // "usedCloudDataMovementUnits": float64(4), - // "usedParallelCopies": float64(1), - // }, - // PipelineName: to.Ptr("examplePipeline"), - // PipelineRunID: to.Ptr("2f7fdb90-5df1-4b8e-ac2f-064cfa58202b"), - // Status: to.Ptr("Succeeded"), - // }}, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/autorest.md b/sdk/resourcemanager/datafactory/armdatafactory/autorest.md index c90362d26431..fd5af3c2411a 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/autorest.md +++ b/sdk/resourcemanager/datafactory/armdatafactory/autorest.md @@ -5,9 +5,8 @@ ``` yaml azure-arm: true require: -- https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/readme.md -- https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/readme.go.md +- /mnt/vss/_work/1/s/azure-rest-api-specs/specification/datafactory/resource-manager/readme.md +- /mnt/vss/_work/1/s/azure-rest-api-specs/specification/datafactory/resource-manager/readme.go.md license-header: MICROSOFT_MIT_NO_VERSION -module-version: 4.0.0 -tag: package-2018-06 +module-version: 5.0.0 ``` \ No newline at end of file diff --git a/sdk/resourcemanager/datafactory/armdatafactory/changedatacapture_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/changedatacapture_client_example_test.go deleted file mode 100644 index 85ba48b4d23b..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/changedatacapture_client_example_test.go +++ /dev/null @@ -1,232 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_ListByFactory.json -func ExampleChangeDataCaptureClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewChangeDataCaptureClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.ChangeDataCaptureListResponse = armdatafactory.ChangeDataCaptureListResponse{ - // Value: []*armdatafactory.ChangeDataCaptureResource{ - // { - // Name: to.Ptr("exampleChangeDataCapture"), - // Type: to.Ptr("Microsoft.DataFactory/factories/adfcdcs"), - // Etag: to.Ptr("4200eefe-0000-0100-0000-641aa97a0000"), - // ID: to.Ptr("/subscriptions/d3bb3b2e-9a7e-4194-9960-5171bd192117/resourceGroups/amja-rg-03/providers/Microsoft.DataFactory/factories/amja-adf-04/adfcdcs/exampleChangeDataCapture"), - // Properties: &armdatafactory.ChangeDataCapture{ - // Description: to.Ptr("Sample demo change data capture to transfer data from delimited (csv) to Azure SQL Database with automapped and non-automapped mappings."), - // AllowVNetOverride: to.Ptr(false), - // Status: to.Ptr("Stopped"), - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_Create.json -func ExampleChangeDataCaptureClient_CreateOrUpdate_changeDataCaptureCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewChangeDataCaptureClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleChangeDataCapture", armdatafactory.ChangeDataCaptureResource{ - Properties: &armdatafactory.ChangeDataCapture{ - Description: to.Ptr("Sample demo change data capture to transfer data from delimited (csv) to Azure SQL Database with automapped and non-automapped mappings."), - AllowVNetOverride: to.Ptr(false), - }, - }, &armdatafactory.ChangeDataCaptureClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ChangeDataCaptureResource = armdatafactory.ChangeDataCaptureResource{ - // Name: to.Ptr("exampleChangeDataCapture"), - // Type: to.Ptr("Microsoft.DataFactory/factories/adfcdcs"), - // Etag: to.Ptr("4300f91d-0000-0100-0000-641ab34d0000"), - // ID: to.Ptr("/subscriptions/d3bb3b2e-9a7e-4194-9960-5171bd192117/resourceGroups/amja-rg-03/providers/Microsoft.DataFactory/factories/amja-adf-04/adfcdcs/exampleChangeDataCapture"), - // Properties: &armdatafactory.ChangeDataCapture{ - // Description: to.Ptr("Sample demo change data capture to transfer data from delimited (csv) to Azure SQL Database. Updating table mappings."), - // AllowVNetOverride: to.Ptr(false), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_Update.json -func ExampleChangeDataCaptureClient_CreateOrUpdate_changeDataCaptureUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewChangeDataCaptureClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleChangeDataCapture", armdatafactory.ChangeDataCaptureResource{ - Properties: &armdatafactory.ChangeDataCapture{ - Description: to.Ptr("Sample demo change data capture to transfer data from delimited (csv) to Azure SQL Database. Updating table mappings."), - AllowVNetOverride: to.Ptr(false), - Status: to.Ptr("Stopped"), - }, - }, &armdatafactory.ChangeDataCaptureClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ChangeDataCaptureResource = armdatafactory.ChangeDataCaptureResource{ - // Name: to.Ptr("exampleChangeDataCapture"), - // Type: to.Ptr("Microsoft.DataFactory/factories/adfcdcs"), - // Etag: to.Ptr("43008e18-0000-0100-0000-641aafca0000"), - // ID: to.Ptr("/subscriptions/d3bb3b2e-9a7e-4194-9960-5171bd192117/resourceGroups/amja-rg-03/providers/Microsoft.DataFactory/factories/amja-adf-04/adfcdcs/exampleChangeDataCapture"), - // Properties: &armdatafactory.ChangeDataCapture{ - // Description: to.Ptr("Sample demo change data capture to transfer data from delimited (csv) to Azure SQL Database. Updating table mappings."), - // AllowVNetOverride: to.Ptr(false), - // Status: to.Ptr("Stopped"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_Get.json -func ExampleChangeDataCaptureClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewChangeDataCaptureClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleChangeDataCapture", &armdatafactory.ChangeDataCaptureClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ChangeDataCaptureResource = armdatafactory.ChangeDataCaptureResource{ - // Name: to.Ptr("exampleChangeDataCapture"), - // Type: to.Ptr("Microsoft.DataFactory/factories/adfcdcs"), - // Etag: to.Ptr("4300f91d-0000-0100-0000-641ab34d0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/adfcdcs/exampleChangeDataCapture"), - // Properties: &armdatafactory.ChangeDataCapture{ - // Description: to.Ptr("Sample demo change data capture to transfer data from delimited (csv) to Azure SQL Database. Updating table mappings."), - // AllowVNetOverride: to.Ptr(false), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_Delete.json -func ExampleChangeDataCaptureClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewChangeDataCaptureClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleChangeDataCapture", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_Start.json -func ExampleChangeDataCaptureClient_Start() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewChangeDataCaptureClient().Start(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleChangeDataCapture", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_Stop.json -func ExampleChangeDataCaptureClient_Stop() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewChangeDataCaptureClient().Stop(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleChangeDataCapture", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ChangeDataCapture_Status.json -func ExampleChangeDataCaptureClient_Status() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewChangeDataCaptureClient().Status(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleChangeDataCapture", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.Value = "Stopped" -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/constants.go b/sdk/resourcemanager/datafactory/armdatafactory/constants.go index cb40cdb62708..7a0287ff1885 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/constants.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/constants.go @@ -10,7 +10,7 @@ package armdatafactory const ( moduleName = "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory" - moduleVersion = "v4.0.0" + moduleVersion = "v5.0.0" ) // ActivityOnInactiveMarkAs - Status result of the activity when the state is set to Inactive. This is an optional property @@ -1324,6 +1324,22 @@ func PossibleSalesforceSinkWriteBehaviorValues() []SalesforceSinkWriteBehavior { } } +// SalesforceV2SinkWriteBehavior - The write behavior for the operation. Default is Insert. +type SalesforceV2SinkWriteBehavior string + +const ( + SalesforceV2SinkWriteBehaviorInsert SalesforceV2SinkWriteBehavior = "Insert" + SalesforceV2SinkWriteBehaviorUpsert SalesforceV2SinkWriteBehavior = "Upsert" +) + +// PossibleSalesforceV2SinkWriteBehaviorValues returns the possible values for the SalesforceV2SinkWriteBehavior const type. +func PossibleSalesforceV2SinkWriteBehaviorValues() []SalesforceV2SinkWriteBehavior { + return []SalesforceV2SinkWriteBehavior{ + SalesforceV2SinkWriteBehaviorInsert, + SalesforceV2SinkWriteBehaviorUpsert, + } +} + // SapCloudForCustomerSinkWriteBehavior - The write behavior for the operation. Default is 'Insert'. type SapCloudForCustomerSinkWriteBehavior string @@ -1502,6 +1518,24 @@ func PossibleSftpAuthenticationTypeValues() []SftpAuthenticationType { } } +// SnowflakeAuthenticationType - The type used for authentication. Type: string. +type SnowflakeAuthenticationType string + +const ( + SnowflakeAuthenticationTypeAADServicePrincipal SnowflakeAuthenticationType = "AADServicePrincipal" + SnowflakeAuthenticationTypeBasic SnowflakeAuthenticationType = "Basic" + SnowflakeAuthenticationTypeKeyPair SnowflakeAuthenticationType = "KeyPair" +) + +// PossibleSnowflakeAuthenticationTypeValues returns the possible values for the SnowflakeAuthenticationType const type. +func PossibleSnowflakeAuthenticationTypeValues() []SnowflakeAuthenticationType { + return []SnowflakeAuthenticationType{ + SnowflakeAuthenticationTypeAADServicePrincipal, + SnowflakeAuthenticationTypeBasic, + SnowflakeAuthenticationTypeKeyPair, + } +} + // SparkAuthenticationType - The authentication method used to access the Spark server. type SparkAuthenticationType string diff --git a/sdk/resourcemanager/datafactory/armdatafactory/credentialoperations_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/credentialoperations_client_example_test.go deleted file mode 100644 index 76a91f66e78c..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/credentialoperations_client_example_test.go +++ /dev/null @@ -1,149 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Credentials_ListByFactory.json -func ExampleCredentialOperationsClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewCredentialOperationsClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.CredentialListResponse = armdatafactory.CredentialListResponse{ - // Value: []*armdatafactory.ManagedIdentityCredentialResource{ - // { - // Name: to.Ptr("exampleLinkedService"), - // Type: to.Ptr("Microsoft.DataFactory/factories/credentials"), - // Etag: to.Ptr("0a0064d4-0000-0000-0000-5b245bd00000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/credentials/exampleCredential"), - // Properties: &armdatafactory.ManagedIdentityCredential{ - // Type: to.Ptr("ManagedIdentity"), - // Description: to.Ptr("Example description"), - // TypeProperties: &armdatafactory.ManagedIdentityTypeProperties{ - // ResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourcegroups/exampleResourceGroup/providers/Microsoft.ManagedIdentity/userAssignedIdentities/exampleUami"), - // }, - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Credentials_Create.json -func ExampleCredentialOperationsClient_CreateOrUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewCredentialOperationsClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleCredential", armdatafactory.ManagedIdentityCredentialResource{ - Properties: &armdatafactory.ManagedIdentityCredential{ - Type: to.Ptr("ManagedIdentity"), - TypeProperties: &armdatafactory.ManagedIdentityTypeProperties{ - ResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourcegroups/exampleResourceGroup/providers/Microsoft.ManagedIdentity/userAssignedIdentities/exampleUami"), - }, - }, - }, &armdatafactory.CredentialOperationsClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ManagedIdentityCredentialResource = armdatafactory.ManagedIdentityCredentialResource{ - // Name: to.Ptr("exampleCredential"), - // Type: to.Ptr("Microsoft.DataFactory/factories/credentials"), - // Etag: to.Ptr("0a0062d4-0000-0000-0000-5b245bcf0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/credentials/exampleCredential"), - // Properties: &armdatafactory.ManagedIdentityCredential{ - // Type: to.Ptr("ManagedIdentity"), - // TypeProperties: &armdatafactory.ManagedIdentityTypeProperties{ - // ResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourcegroups/exampleResourceGroup/providers/Microsoft.ManagedIdentity/userAssignedIdentities/exampleUami"), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Credentials_Get.json -func ExampleCredentialOperationsClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewCredentialOperationsClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleCredential", &armdatafactory.CredentialOperationsClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ManagedIdentityCredentialResource = armdatafactory.ManagedIdentityCredentialResource{ - // Name: to.Ptr("exampleLinkedService"), - // Type: to.Ptr("Microsoft.DataFactory/factories/credentials"), - // Etag: to.Ptr("1500474f-0000-0200-0000-5cbe090d0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/credentials/exampleCredential"), - // Properties: &armdatafactory.ManagedIdentityCredential{ - // Type: to.Ptr("ManagedIdentity"), - // Description: to.Ptr("Example description"), - // TypeProperties: &armdatafactory.ManagedIdentityTypeProperties{ - // ResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourcegroups/exampleResourceGroup/providers/Microsoft.ManagedIdentity/userAssignedIdentities/exampleUami"), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Credentials_Delete.json -func ExampleCredentialOperationsClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewCredentialOperationsClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleCredential", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/datafactory_live_test.go b/sdk/resourcemanager/datafactory/armdatafactory/datafactory_live_test.go index 807185b52c40..1dafc78d3e48 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/datafactory_live_test.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/datafactory_live_test.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/arm" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" "github.com/Azure/azure-sdk-for-go/sdk/internal/recording" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/internal/v2/testutil" "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/resources/armresources" "github.com/stretchr/testify/suite" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/dataflowdebugsession_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/dataflowdebugsession_client_example_test.go deleted file mode 100644 index 871470762f34..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/dataflowdebugsession_client_example_test.go +++ /dev/null @@ -1,265 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlowDebugSession_Create.json -func ExampleDataFlowDebugSessionClient_BeginCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewDataFlowDebugSessionClient().BeginCreate(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.CreateDataFlowDebugSessionRequest{ - IntegrationRuntime: &armdatafactory.IntegrationRuntimeDebugResource{ - Name: to.Ptr("ir1"), - Properties: &armdatafactory.ManagedIntegrationRuntime{ - Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeManaged), - TypeProperties: &armdatafactory.ManagedIntegrationRuntimeTypeProperties{ - ComputeProperties: &armdatafactory.IntegrationRuntimeComputeProperties{ - DataFlowProperties: &armdatafactory.IntegrationRuntimeDataFlowProperties{ - ComputeType: to.Ptr(armdatafactory.DataFlowComputeTypeGeneral), - CoreCount: to.Ptr[int32](48), - TimeToLive: to.Ptr[int32](10), - }, - Location: to.Ptr("AutoResolve"), - }, - }, - }, - }, - TimeToLive: to.Ptr[int32](60), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - res, err := poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.CreateDataFlowDebugSessionResponse = armdatafactory.CreateDataFlowDebugSessionResponse{ - // SessionID: to.Ptr("229c688c-944c-44ac-b31a-82d50f347154"), - // Status: to.Ptr("Succeeded"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlowDebugSession_QueryByFactory.json -func ExampleDataFlowDebugSessionClient_NewQueryByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewDataFlowDebugSessionClient().NewQueryByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.QueryDataFlowDebugSessionsResponse = armdatafactory.QueryDataFlowDebugSessionsResponse{ - // Value: []*armdatafactory.DataFlowDebugSessionInfo{ - // { - // AdditionalProperties: map[string]any{ - // "dataflowName": "DebugSession-0a7e0d6e-f2b7-48cc-8cd8-618326f5662f", - // "userObjectId": "0a7e0d6e-f2b7-48cc-8cd8-618326f5662f", - // }, - // ComputeType: to.Ptr("General"), - // CoreCount: to.Ptr[int32](48), - // LastActivityTime: to.Ptr("2019-09-05T18:28:00.9459674+00:00"), - // SessionID: to.Ptr("229c688c-944c-44ac-b31a-82d50f347154"), - // StartTime: to.Ptr("2019-09-05T18:23:20.3257799+00:00"), - // TimeToLiveInMinutes: to.Ptr[int32](60), - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlowDebugSession_AddDataFlow.json -func ExampleDataFlowDebugSessionClient_AddDataFlow() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewDataFlowDebugSessionClient().AddDataFlow(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.DataFlowDebugPackage{ - DataFlow: &armdatafactory.DataFlowDebugResource{ - Name: to.Ptr("dataflow1"), - Properties: &armdatafactory.MappingDataFlow{ - Type: to.Ptr("MappingDataFlow"), - TypeProperties: &armdatafactory.MappingDataFlowTypeProperties{ - Script: to.Ptr("\n\nsource(output(\n Column_1 as string\n ),\n allowSchemaDrift: true,\n validateSchema: false) ~> source1"), - Sinks: []*armdatafactory.DataFlowSink{}, - Sources: []*armdatafactory.DataFlowSource{ - { - Name: to.Ptr("source1"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("DelimitedText2"), - }, - }}, - Transformations: []*armdatafactory.Transformation{}, - }, - }, - }, - Datasets: []*armdatafactory.DatasetDebugResource{ - { - Name: to.Ptr("dataset1"), - Properties: &armdatafactory.DelimitedTextDataset{ - Type: to.Ptr("DelimitedText"), - Schema: []any{ - map[string]any{ - "type": "String", - }, - }, - Annotations: []any{}, - LinkedServiceName: &armdatafactory.LinkedServiceReference{ - Type: to.Ptr(armdatafactory.LinkedServiceReferenceTypeLinkedServiceReference), - ReferenceName: to.Ptr("linkedService5"), - }, - TypeProperties: &armdatafactory.DelimitedTextDatasetTypeProperties{ - ColumnDelimiter: ",", - EscapeChar: "\\", - FirstRowAsHeader: true, - Location: &armdatafactory.AzureBlobStorageLocation{ - Type: to.Ptr("AzureBlobStorageLocation"), - FileName: "Ansiencoding.csv", - Container: "dataflow-sample-data", - }, - QuoteChar: "\"", - }, - }, - }}, - DebugSettings: &armdatafactory.DataFlowDebugPackageDebugSettings{ - DatasetParameters: map[string]any{ - "Movies": map[string]any{ - "path": "abc", - }, - "Output": map[string]any{ - "time": "def", - }, - }, - Parameters: map[string]any{ - "sourcePath": "Toy", - }, - SourceSettings: []*armdatafactory.DataFlowSourceSetting{ - { - RowLimit: to.Ptr[int32](1000), - SourceName: to.Ptr("source1"), - }, - { - RowLimit: to.Ptr[int32](222), - SourceName: to.Ptr("source2"), - }}, - }, - LinkedServices: []*armdatafactory.LinkedServiceDebugResource{ - { - Name: to.Ptr("linkedService1"), - Properties: &armdatafactory.AzureBlobStorageLinkedService{ - Type: to.Ptr("AzureBlobStorage"), - Annotations: []any{}, - TypeProperties: &armdatafactory.AzureBlobStorageLinkedServiceTypeProperties{ - ConnectionString: "DefaultEndpointsProtocol=https;AccountName=;EndpointSuffix=core.windows.net;", - EncryptedCredential: to.Ptr(""), - }, - }, - }}, - SessionID: to.Ptr("f06ed247-9d07-49b2-b05e-2cb4a2fc871e"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.AddDataFlowToDebugSessionResponse = armdatafactory.AddDataFlowToDebugSessionResponse{ - // JobVersion: to.Ptr("e5328ee7-c524-4207-8ba4-b709010db33d"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlowDebugSession_Delete.json -func ExampleDataFlowDebugSessionClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewDataFlowDebugSessionClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.DeleteDataFlowDebugSessionRequest{ - SessionID: to.Ptr("91fb57e0-8292-47be-89ff-c8f2d2bb2a7e"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlowDebugSession_ExecuteCommand.json -func ExampleDataFlowDebugSessionClient_BeginExecuteCommand() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewDataFlowDebugSessionClient().BeginExecuteCommand(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.DataFlowDebugCommandRequest{ - Command: to.Ptr(armdatafactory.DataFlowDebugCommandTypeExecutePreviewQuery), - CommandPayload: &armdatafactory.DataFlowDebugCommandPayload{ - RowLimits: to.Ptr[int32](100), - StreamName: to.Ptr("source1"), - }, - SessionID: to.Ptr("f06ed247-9d07-49b2-b05e-2cb4a2fc871e"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - res, err := poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.DataFlowDebugCommandResponse = armdatafactory.DataFlowDebugCommandResponse{ - // Data: to.Ptr("some output"), - // Status: to.Ptr("Succeeded"), - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/dataflows_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/dataflows_client_example_test.go deleted file mode 100644 index e5f4afe34388..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/dataflows_client_example_test.go +++ /dev/null @@ -1,460 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlows_Create.json -func ExampleDataFlowsClient_CreateOrUpdate_dataFlowsCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewDataFlowsClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataFlow", armdatafactory.DataFlowResource{ - Properties: &armdatafactory.MappingDataFlow{ - Type: to.Ptr("MappingDataFlow"), - Description: to.Ptr("Sample demo data flow to convert currencies showing usage of union, derive and conditional split transformation."), - TypeProperties: &armdatafactory.MappingDataFlowTypeProperties{ - ScriptLines: []*string{ - to.Ptr("source(output("), - to.Ptr("PreviousConversionRate as double,"), - to.Ptr("Country as string,"), - to.Ptr("DateTime1 as string,"), - to.Ptr("CurrentConversionRate as double"), - to.Ptr("),"), - to.Ptr("allowSchemaDrift: false,"), - to.Ptr("validateSchema: false) ~> USDCurrency"), - to.Ptr("source(output("), - to.Ptr("PreviousConversionRate as double,"), - to.Ptr("Country as string,"), - to.Ptr("DateTime1 as string,"), - to.Ptr("CurrentConversionRate as double"), - to.Ptr("),"), - to.Ptr("allowSchemaDrift: true,"), - to.Ptr("validateSchema: false) ~> CADSource"), - to.Ptr("USDCurrency, CADSource union(byName: true)~> Union"), - to.Ptr("Union derive(NewCurrencyRate = round(CurrentConversionRate*1.25)) ~> NewCurrencyColumn"), - to.Ptr("NewCurrencyColumn split(Country == 'USD',"), - to.Ptr("Country == 'CAD',disjoint: false) ~> ConditionalSplit1@(USD, CAD)"), - to.Ptr("ConditionalSplit1@USD sink(saveMode:'overwrite' ) ~> USDSink"), - to.Ptr("ConditionalSplit1@CAD sink(saveMode:'overwrite' ) ~> CADSink")}, - Sinks: []*armdatafactory.DataFlowSink{ - { - Name: to.Ptr("USDSink"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("USDOutput"), - }, - }, - { - Name: to.Ptr("CADSink"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("CADOutput"), - }, - }}, - Sources: []*armdatafactory.DataFlowSource{ - { - Name: to.Ptr("USDCurrency"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("CurrencyDatasetUSD"), - }, - }, - { - Name: to.Ptr("CADSource"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("CurrencyDatasetCAD"), - }, - }}, - }, - }, - }, &armdatafactory.DataFlowsClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.DataFlowResource = armdatafactory.DataFlowResource{ - // Name: to.Ptr("exampleDataFlow"), - // Type: to.Ptr("Microsoft.DataFactory/factories/dataflows"), - // Etag: to.Ptr("0a0066d4-0000-0000-0000-5b245bd20000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/datasets/exampleDataset"), - // Properties: &armdatafactory.MappingDataFlow{ - // Type: to.Ptr("MappingDataFlow"), - // Description: to.Ptr("Sample demo data flow to convert currencies showing usage of union, derive and conditional split transformation."), - // TypeProperties: &armdatafactory.MappingDataFlowTypeProperties{ - // ScriptLines: []*string{ - // to.Ptr("source(output("), - // to.Ptr("PreviousConversionRate as double,"), - // to.Ptr("Country as string,"), - // to.Ptr("DateTime1 as string,"), - // to.Ptr("CurrentConversionRate as double"), - // to.Ptr("),"), - // to.Ptr("allowSchemaDrift: false,"), - // to.Ptr("validateSchema: false) ~> USDCurrency"), - // to.Ptr("source(output("), - // to.Ptr("PreviousConversionRate as double,"), - // to.Ptr("Country as string,"), - // to.Ptr("DateTime1 as string,"), - // to.Ptr("CurrentConversionRate as double"), - // to.Ptr("),"), - // to.Ptr("allowSchemaDrift: true,"), - // to.Ptr("validateSchema: false) ~> CADSource"), - // to.Ptr("USDCurrency, CADSource union(byName: true)~> Union"), - // to.Ptr("Union derive(NewCurrencyRate = round(CurrentConversionRate*1.25)) ~> NewCurrencyColumn"), - // to.Ptr("NewCurrencyColumn split(Country == 'USD',"), - // to.Ptr("Country == 'CAD',disjoint: false) ~> ConditionalSplit1@(USD, CAD)"), - // to.Ptr("ConditionalSplit1@USD sink(saveMode:'overwrite' ) ~> USDSink"), - // to.Ptr("ConditionalSplit1@CAD sink(saveMode:'overwrite' ) ~> CADSink")}, - // Sinks: []*armdatafactory.DataFlowSink{ - // { - // Name: to.Ptr("USDSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("USDOutput"), - // }, - // }, - // { - // Name: to.Ptr("CADSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CADOutput"), - // }, - // }}, - // Sources: []*armdatafactory.DataFlowSource{ - // { - // Name: to.Ptr("USDCurrency"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetUSD"), - // }, - // }, - // { - // Name: to.Ptr("CADSource"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetCAD"), - // }, - // }}, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlows_Update.json -func ExampleDataFlowsClient_CreateOrUpdate_dataFlowsUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewDataFlowsClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataFlow", armdatafactory.DataFlowResource{ - Properties: &armdatafactory.MappingDataFlow{ - Type: to.Ptr("MappingDataFlow"), - Description: to.Ptr("Sample demo data flow to convert currencies showing usage of union, derive and conditional split transformation."), - TypeProperties: &armdatafactory.MappingDataFlowTypeProperties{ - ScriptLines: []*string{ - to.Ptr("source(output("), - to.Ptr("PreviousConversionRate as double,"), - to.Ptr("Country as string,"), - to.Ptr("DateTime1 as string,"), - to.Ptr("CurrentConversionRate as double"), - to.Ptr("),"), - to.Ptr("allowSchemaDrift: false,"), - to.Ptr("validateSchema: false) ~> USDCurrency"), - to.Ptr("source(output("), - to.Ptr("PreviousConversionRate as double,"), - to.Ptr("Country as string,"), - to.Ptr("DateTime1 as string,"), - to.Ptr("CurrentConversionRate as double"), - to.Ptr("),"), - to.Ptr("allowSchemaDrift: true,"), - to.Ptr("validateSchema: false) ~> CADSource"), - to.Ptr("USDCurrency, CADSource union(byName: true)~> Union"), - to.Ptr("Union derive(NewCurrencyRate = round(CurrentConversionRate*1.25)) ~> NewCurrencyColumn"), - to.Ptr("NewCurrencyColumn split(Country == 'USD',"), - to.Ptr("Country == 'CAD',disjoint: false) ~> ConditionalSplit1@(USD, CAD)"), - to.Ptr("ConditionalSplit1@USD sink(saveMode:'overwrite' ) ~> USDSink"), - to.Ptr("ConditionalSplit1@CAD sink(saveMode:'overwrite' ) ~> CADSink")}, - Sinks: []*armdatafactory.DataFlowSink{ - { - Name: to.Ptr("USDSink"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("USDOutput"), - }, - }, - { - Name: to.Ptr("CADSink"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("CADOutput"), - }, - }}, - Sources: []*armdatafactory.DataFlowSource{ - { - Name: to.Ptr("USDCurrency"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("CurrencyDatasetUSD"), - }, - }, - { - Name: to.Ptr("CADSource"), - Dataset: &armdatafactory.DatasetReference{ - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - ReferenceName: to.Ptr("CurrencyDatasetCAD"), - }, - }}, - }, - }, - }, &armdatafactory.DataFlowsClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.DataFlowResource = armdatafactory.DataFlowResource{ - // Name: to.Ptr("exampleDataFlow"), - // Type: to.Ptr("Microsoft.DataFactory/factories/dataflows"), - // Etag: to.Ptr("0a0068d4-0000-0000-0000-5b245bd30002"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/datasets/exampleDataset"), - // Properties: &armdatafactory.MappingDataFlow{ - // Type: to.Ptr("MappingDataFlow"), - // Description: to.Ptr("Sample demo data flow to convert currencies showing usage of union, derive and conditional split transformation."), - // TypeProperties: &armdatafactory.MappingDataFlowTypeProperties{ - // ScriptLines: []*string{ - // to.Ptr("source(output("), - // to.Ptr("PreviousConversionRate as double,"), - // to.Ptr("Country as string,"), - // to.Ptr("DateTime1 as string,"), - // to.Ptr("CurrentConversionRate as double"), - // to.Ptr("),"), - // to.Ptr("allowSchemaDrift: false,"), - // to.Ptr("validateSchema: false) ~> USDCurrency"), - // to.Ptr("source(output("), - // to.Ptr("PreviousConversionRate as double,"), - // to.Ptr("Country as string,"), - // to.Ptr("DateTime1 as string,"), - // to.Ptr("CurrentConversionRate as double"), - // to.Ptr("),"), - // to.Ptr("allowSchemaDrift: true,"), - // to.Ptr("validateSchema: false) ~> CADSource"), - // to.Ptr("USDCurrency, CADSource union(byName: true)~> Union"), - // to.Ptr("Union derive(NewCurrencyRate = round(CurrentConversionRate*1.25)) ~> NewCurrencyColumn"), - // to.Ptr("NewCurrencyColumn split(Country == 'USD',"), - // to.Ptr("Country == 'CAD',disjoint: false) ~> ConditionalSplit1@(USD, CAD)"), - // to.Ptr("ConditionalSplit1@USD sink(saveMode:'overwrite' ) ~> USDSink"), - // to.Ptr("ConditionalSplit1@CAD sink(saveMode:'overwrite' ) ~> CADSink")}, - // Sinks: []*armdatafactory.DataFlowSink{ - // { - // Name: to.Ptr("USDSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("USDOutput"), - // }, - // }, - // { - // Name: to.Ptr("CADSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CADOutput"), - // }, - // }}, - // Sources: []*armdatafactory.DataFlowSource{ - // { - // Name: to.Ptr("USDCurrency"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetUSD"), - // }, - // }, - // { - // Name: to.Ptr("CADSource"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetCAD"), - // }, - // }}, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlows_Get.json -func ExampleDataFlowsClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewDataFlowsClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataFlow", &armdatafactory.DataFlowsClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.DataFlowResource = armdatafactory.DataFlowResource{ - // Name: to.Ptr("exampleDataFlow"), - // Type: to.Ptr("Microsoft.DataFactory/factories/dataflows"), - // Etag: to.Ptr("15004c4f-0000-0200-0000-5cbe090e0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/dataflows/exampleDataFlow"), - // Properties: &armdatafactory.MappingDataFlow{ - // Type: to.Ptr("MappingDataFlow"), - // Description: to.Ptr("Sample demo data flow to convert currencies showing usage of union, derive and conditional split transformation."), - // TypeProperties: &armdatafactory.MappingDataFlowTypeProperties{ - // Script: to.Ptr("source(output(PreviousConversionRate as double,Country as string,DateTime1 as string,CurrentConversionRate as double),allowSchemaDrift: false,validateSchema: false) ~> USDCurrency\nsource(output(PreviousConversionRate as double,Country as string,DateTime1 as string,CurrentConversionRate as double),allowSchemaDrift: true,validateSchema: false) ~> CADSource\nUSDCurrency, CADSource union(byName: true)~> Union\nUnion derive(NewCurrencyRate = round(CurrentConversionRate*1.25)) ~> NewCurrencyColumn\nNewCurrencyColumn split(Country == 'USD',Country == 'CAD',disjoint: false) ~> ConditionalSplit1@(USD, CAD)\nConditionalSplit1@USD sink(saveMode:'overwrite' ) ~> USDSink\nConditionalSplit1@CAD sink(saveMode:'overwrite' ) ~> CADSink"), - // Sinks: []*armdatafactory.DataFlowSink{ - // { - // Name: to.Ptr("USDSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("USDOutput"), - // }, - // }, - // { - // Name: to.Ptr("CADSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CADOutput"), - // }, - // }}, - // Sources: []*armdatafactory.DataFlowSource{ - // { - // Name: to.Ptr("USDCurrency"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetUSD"), - // }, - // }, - // { - // Name: to.Ptr("CADSource"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetCAD"), - // }, - // }}, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlows_Delete.json -func ExampleDataFlowsClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewDataFlowsClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataFlow", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DataFlows_ListByFactory.json -func ExampleDataFlowsClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewDataFlowsClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.DataFlowListResponse = armdatafactory.DataFlowListResponse{ - // Value: []*armdatafactory.DataFlowResource{ - // { - // Name: to.Ptr("exampleDataFlow"), - // Type: to.Ptr("Microsoft.DataFactory/factories/dataflows"), - // Etag: to.Ptr("0a0068d4-0000-0000-0000-5b245bd30000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/dataflows/exampleDataFlow"), - // Properties: &armdatafactory.MappingDataFlow{ - // Type: to.Ptr("MappingDataFlow"), - // Description: to.Ptr("Sample demo data flow to convert currencies showing usage of union, derive and conditional split transformation."), - // TypeProperties: &armdatafactory.MappingDataFlowTypeProperties{ - // Script: to.Ptr("source(output(PreviousConversionRate as double,Country as string,DateTime1 as string,CurrentConversionRate as double),allowSchemaDrift: false,validateSchema: false) ~> USDCurrency\nsource(output(PreviousConversionRate as double,Country as string,DateTime1 as string,CurrentConversionRate as double),allowSchemaDrift: true,validateSchema: false) ~> CADSource\nUSDCurrency, CADSource union(byName: true)~> Union\nUnion derive(NewCurrencyRate = round(CurrentConversionRate*1.25)) ~> NewCurrencyColumn\nNewCurrencyColumn split(Country == 'USD',Country == 'CAD',disjoint: false) ~> ConditionalSplit1@(USD, CAD)\nConditionalSplit1@USD sink(saveMode:'overwrite' ) ~> USDSink\nConditionalSplit1@CAD sink(saveMode:'overwrite' ) ~> CADSink"), - // Sinks: []*armdatafactory.DataFlowSink{ - // { - // Name: to.Ptr("USDSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("USDOutput"), - // }, - // }, - // { - // Name: to.Ptr("CADSink"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CADOutput"), - // }, - // }}, - // Sources: []*armdatafactory.DataFlowSource{ - // { - // Name: to.Ptr("USDCurrency"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetUSD"), - // }, - // }, - // { - // Name: to.Ptr("CADSource"), - // Dataset: &armdatafactory.DatasetReference{ - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // ReferenceName: to.Ptr("CurrencyDatasetCAD"), - // }, - // }}, - // }, - // }, - // }}, - // } - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/datasets_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/datasets_client_example_test.go deleted file mode 100644 index 1b8452fa6d36..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/datasets_client_example_test.go +++ /dev/null @@ -1,334 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Datasets_ListByFactory.json -func ExampleDatasetsClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewDatasetsClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.DatasetListResponse = armdatafactory.DatasetListResponse{ - // Value: []*armdatafactory.DatasetResource{ - // { - // Name: to.Ptr("exampleDataset"), - // Type: to.Ptr("Microsoft.DataFactory/factories/datasets"), - // Etag: to.Ptr("0a0068d4-0000-0000-0000-5b245bd30000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/datasets/exampleDataset"), - // Properties: &armdatafactory.AzureBlobDataset{ - // Type: to.Ptr("AzureBlob"), - // Description: to.Ptr("Example description"), - // LinkedServiceName: &armdatafactory.LinkedServiceReference{ - // Type: to.Ptr(armdatafactory.LinkedServiceReferenceTypeLinkedServiceReference), - // ReferenceName: to.Ptr("exampleLinkedService"), - // }, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "MyFileName": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // "MyFolderPath": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // }, - // TypeProperties: &armdatafactory.AzureBlobDatasetTypeProperties{ - // Format: &armdatafactory.TextFormat{ - // Type: to.Ptr("TextFormat"), - // }, - // FileName: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFileName", - // }, - // FolderPath: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFolderPath", - // }, - // }, - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Datasets_Create.json -func ExampleDatasetsClient_CreateOrUpdate_datasetsCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewDatasetsClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataset", armdatafactory.DatasetResource{ - Properties: &armdatafactory.AzureBlobDataset{ - Type: to.Ptr("AzureBlob"), - LinkedServiceName: &armdatafactory.LinkedServiceReference{ - Type: to.Ptr(armdatafactory.LinkedServiceReferenceTypeLinkedServiceReference), - ReferenceName: to.Ptr("exampleLinkedService"), - }, - Parameters: map[string]*armdatafactory.ParameterSpecification{ - "MyFileName": { - Type: to.Ptr(armdatafactory.ParameterTypeString), - }, - "MyFolderPath": { - Type: to.Ptr(armdatafactory.ParameterTypeString), - }, - }, - TypeProperties: &armdatafactory.AzureBlobDatasetTypeProperties{ - Format: &armdatafactory.TextFormat{ - Type: to.Ptr("TextFormat"), - }, - FileName: map[string]any{ - "type": "Expression", - "value": "@dataset().MyFileName", - }, - FolderPath: map[string]any{ - "type": "Expression", - "value": "@dataset().MyFolderPath", - }, - }, - }, - }, &armdatafactory.DatasetsClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.DatasetResource = armdatafactory.DatasetResource{ - // Name: to.Ptr("exampleDataset"), - // Type: to.Ptr("Microsoft.DataFactory/factories/datasets"), - // Etag: to.Ptr("0a0066d4-0000-0000-0000-5b245bd20000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/datasets/exampleDataset"), - // Properties: &armdatafactory.AzureBlobDataset{ - // Type: to.Ptr("AzureBlob"), - // Schema: []any{ - // map[string]any{ - // "name": "col1", - // "type": "INT_32", - // }, - // map[string]any{ - // "name": "col2", - // "type": "Decimal", - // "precision": "38", - // "scale": "2", - // }, - // }, - // LinkedServiceName: &armdatafactory.LinkedServiceReference{ - // Type: to.Ptr(armdatafactory.LinkedServiceReferenceTypeLinkedServiceReference), - // ReferenceName: to.Ptr("exampleLinkedService"), - // }, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "MyFileName": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // "MyFolderPath": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // }, - // TypeProperties: &armdatafactory.AzureBlobDatasetTypeProperties{ - // Format: &armdatafactory.TextFormat{ - // Type: to.Ptr("TextFormat"), - // }, - // FileName: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFileName", - // }, - // FolderPath: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFolderPath", - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Datasets_Update.json -func ExampleDatasetsClient_CreateOrUpdate_datasetsUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewDatasetsClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataset", armdatafactory.DatasetResource{ - Properties: &armdatafactory.AzureBlobDataset{ - Type: to.Ptr("AzureBlob"), - Description: to.Ptr("Example description"), - LinkedServiceName: &armdatafactory.LinkedServiceReference{ - Type: to.Ptr(armdatafactory.LinkedServiceReferenceTypeLinkedServiceReference), - ReferenceName: to.Ptr("exampleLinkedService"), - }, - Parameters: map[string]*armdatafactory.ParameterSpecification{ - "MyFileName": { - Type: to.Ptr(armdatafactory.ParameterTypeString), - }, - "MyFolderPath": { - Type: to.Ptr(armdatafactory.ParameterTypeString), - }, - }, - TypeProperties: &armdatafactory.AzureBlobDatasetTypeProperties{ - Format: &armdatafactory.TextFormat{ - Type: to.Ptr("TextFormat"), - }, - FileName: map[string]any{ - "type": "Expression", - "value": "@dataset().MyFileName", - }, - FolderPath: map[string]any{ - "type": "Expression", - "value": "@dataset().MyFolderPath", - }, - }, - }, - }, &armdatafactory.DatasetsClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.DatasetResource = armdatafactory.DatasetResource{ - // Name: to.Ptr("exampleDataset"), - // Type: to.Ptr("Microsoft.DataFactory/factories/datasets"), - // Etag: to.Ptr("0a0068d4-0000-0000-0000-5b245bd30000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/datasets/exampleDataset"), - // Properties: &armdatafactory.AzureBlobDataset{ - // Type: to.Ptr("AzureBlob"), - // Description: to.Ptr("Example description"), - // LinkedServiceName: &armdatafactory.LinkedServiceReference{ - // Type: to.Ptr(armdatafactory.LinkedServiceReferenceTypeLinkedServiceReference), - // ReferenceName: to.Ptr("exampleLinkedService"), - // }, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "MyFileName": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // "MyFolderPath": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // }, - // TypeProperties: &armdatafactory.AzureBlobDatasetTypeProperties{ - // Format: &armdatafactory.TextFormat{ - // Type: to.Ptr("TextFormat"), - // }, - // FileName: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFileName", - // }, - // FolderPath: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFolderPath", - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Datasets_Get.json -func ExampleDatasetsClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewDatasetsClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataset", &armdatafactory.DatasetsClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.DatasetResource = armdatafactory.DatasetResource{ - // Name: to.Ptr("exampleDataset"), - // Type: to.Ptr("Microsoft.DataFactory/factories/datasets"), - // Etag: to.Ptr("15004c4f-0000-0200-0000-5cbe090e0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/datasets/exampleDataset"), - // Properties: &armdatafactory.AzureBlobDataset{ - // Type: to.Ptr("AzureBlob"), - // Description: to.Ptr("Example description"), - // LinkedServiceName: &armdatafactory.LinkedServiceReference{ - // Type: to.Ptr(armdatafactory.LinkedServiceReferenceTypeLinkedServiceReference), - // ReferenceName: to.Ptr("exampleLinkedService"), - // }, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "MyFileName": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // "MyFolderPath": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // }, - // TypeProperties: &armdatafactory.AzureBlobDatasetTypeProperties{ - // Format: &armdatafactory.TextFormat{ - // Type: to.Ptr("TextFormat"), - // }, - // FileName: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFileName", - // }, - // FolderPath: map[string]any{ - // "type": "Expression", - // "value": "@dataset().MyFolderPath", - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Datasets_Delete.json -func ExampleDatasetsClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewDatasetsClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleDataset", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/exposurecontrol_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/exposurecontrol_client_example_test.go deleted file mode 100644 index 620eb3f00116..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/exposurecontrol_client_example_test.go +++ /dev/null @@ -1,114 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ExposureControl_GetFeatureValue.json -func ExampleExposureControlClient_GetFeatureValue() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewExposureControlClient().GetFeatureValue(ctx, "WestEurope", armdatafactory.ExposureControlRequest{ - FeatureName: to.Ptr("ADFIntegrationRuntimeSharingRbac"), - FeatureType: to.Ptr("Feature"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ExposureControlResponse = armdatafactory.ExposureControlResponse{ - // FeatureName: to.Ptr("ADFIntegrationRuntimeSharingRbac"), - // Value: to.Ptr("False"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ExposureControl_GetFeatureValueByFactory.json -func ExampleExposureControlClient_GetFeatureValueByFactory() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewExposureControlClient().GetFeatureValueByFactory(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.ExposureControlRequest{ - FeatureName: to.Ptr("ADFIntegrationRuntimeSharingRbac"), - FeatureType: to.Ptr("Feature"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ExposureControlResponse = armdatafactory.ExposureControlResponse{ - // FeatureName: to.Ptr("ADFIntegrationRuntimeSharingRbac"), - // Value: to.Ptr("False"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ExposureControl_QueryFeatureValuesByFactory.json -func ExampleExposureControlClient_QueryFeatureValuesByFactory() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewExposureControlClient().QueryFeatureValuesByFactory(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.ExposureControlBatchRequest{ - ExposureControlRequests: []*armdatafactory.ExposureControlRequest{ - { - FeatureName: to.Ptr("ADFIntegrationRuntimeSharingRbac"), - FeatureType: to.Ptr("Feature"), - }, - { - FeatureName: to.Ptr("ADFSampleFeature"), - FeatureType: to.Ptr("Feature"), - }}, - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ExposureControlBatchResponse = armdatafactory.ExposureControlBatchResponse{ - // ExposureControlResponses: []*armdatafactory.ExposureControlResponse{ - // { - // FeatureName: to.Ptr("ADFIntegrationRuntimeSharingRbac"), - // Value: to.Ptr("False"), - // }, - // { - // FeatureName: to.Ptr("ADFSampleFeature"), - // Value: to.Ptr("True"), - // }}, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/factories_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/factories_client_example_test.go deleted file mode 100644 index e6d8f2305285..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/factories_client_example_test.go +++ /dev/null @@ -1,592 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_List.json -func ExampleFactoriesClient_NewListPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewFactoriesClient().NewListPager(nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.FactoryListResponse = armdatafactory.FactoryListResponse{ - // Value: []*armdatafactory.Factory{ - // { - // Name: to.Ptr("rpV2OrigDF-72c7d3d4-5e17-4ec6-91de-9ab433f15e79"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"0000aa0d-0000-0000-0000-5b0d58170000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/rg-yanzhang-dfv2/providers/Microsoft.DataFactory/factories/rpv2origdf-72c7d3d4-5e17-4ec6-91de-9ab433f15e79"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Identity: &armdatafactory.FactoryIdentity{ - // Type: to.Ptr(armdatafactory.FactoryIdentityTypeSystemAssigned), - // PrincipalID: to.Ptr("399c3de2-6072-4326-bfa9-4d0c116f1a7b"), - // TenantID: to.Ptr("12345678-1234-1234-1234-123456789abc"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-05-29T13:39:35.615Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2017-09-01-preview"), - // }, - // }, - // { - // Name: to.Ptr("df-dogfood-yanzhang-we"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"0000f301-0000-0000-0000-5b21b16c0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/rg-yanzhang-dfv2/providers/Microsoft.DataFactory/factories/df-dogfood-yanzhang-we"), - // Location: to.Ptr("West Europe"), - // Tags: map[string]*string{ - // }, - // Identity: &armdatafactory.FactoryIdentity{ - // Type: to.Ptr(armdatafactory.FactoryIdentityTypeSystemAssigned), - // PrincipalID: to.Ptr("e8dd6df9-bad5-4dea-8fb8-0d13d1845d9e"), - // TenantID: to.Ptr("12345678-1234-1234-1234-123456789abc"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T00:06:04.666Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2017-09-01-preview"), - // }, - // }, - // { - // Name: to.Ptr("exampleFactoryName-linked"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00008a02-0000-0000-0000-5b237f270000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName-linked"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Identity: &armdatafactory.FactoryIdentity{ - // Type: to.Ptr(armdatafactory.FactoryIdentityTypeSystemAssigned), - // PrincipalID: to.Ptr("10743799-44d2-42fe-8c4d-5bc5c51c0684"), - // TenantID: to.Ptr("12345678-1234-1234-1234-123456789abc"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-15T08:56:07.182Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2017-09-01-preview"), - // }, - // }, - // { - // Name: to.Ptr("FactoryToUpgrade"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00003d04-0000-0000-0000-5b28962f0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/factorytoupgrade"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:35:35.713Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2018-06-01"), - // }, - // }, - // { - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00004004-0000-0000-0000-5b28979e0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // "exampleTag": to.Ptr("exampleValue"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:41:50.004Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // PurviewConfiguration: &armdatafactory.PurviewConfiguration{ - // PurviewResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.Purview/accounts/examplePurview"), - // }, - // RepoConfiguration: &armdatafactory.FactoryVSTSConfiguration{ - // Type: to.Ptr("FactoryVSTSConfiguration"), - // AccountName: to.Ptr("ADF"), - // CollaborationBranch: to.Ptr("master"), - // LastCommitID: to.Ptr(""), - // RepositoryName: to.Ptr("repo"), - // RootFolder: to.Ptr("/"), - // ProjectName: to.Ptr("project"), - // TenantID: to.Ptr(""), - // }, - // Version: to.Ptr("2018-06-01"), - // }, - // }, - // { - // Name: to.Ptr("rpV2OrigDF-72c7d3d4-5e17-4ec6-91de-9ab433f15e79"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"0000aa0d-0000-0000-0000-5b0d58170000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/rg-yanzhang-dfv2/providers/Microsoft.DataFactory/factories/rpv2origdf-72c7d3d4-5e17-4ec6-91de-9ab433f15e79"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Identity: &armdatafactory.FactoryIdentity{ - // Type: to.Ptr(armdatafactory.FactoryIdentityTypeSystemAssigned), - // PrincipalID: to.Ptr("399c3de2-6072-4326-bfa9-4d0c116f1a7b"), - // TenantID: to.Ptr("12345678-1234-1234-1234-123456789abc"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-05-29T13:39:35.615Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2017-09-01-preview"), - // }, - // }, - // { - // Name: to.Ptr("df-dogfood-yanzhang-we"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"0000f301-0000-0000-0000-5b21b16c0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/rg-yanzhang-dfv2/providers/Microsoft.DataFactory/factories/df-dogfood-yanzhang-we"), - // Location: to.Ptr("West Europe"), - // Tags: map[string]*string{ - // }, - // Identity: &armdatafactory.FactoryIdentity{ - // Type: to.Ptr(armdatafactory.FactoryIdentityTypeSystemAssigned), - // PrincipalID: to.Ptr("e8dd6df9-bad5-4dea-8fb8-0d13d1845d9e"), - // TenantID: to.Ptr("12345678-1234-1234-1234-123456789abc"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T00:06:04.666Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2017-09-01-preview"), - // }, - // }, - // { - // Name: to.Ptr("exampleFactoryName-linked"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00008a02-0000-0000-0000-5b237f270000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName-linked"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Identity: &armdatafactory.FactoryIdentity{ - // Type: to.Ptr(armdatafactory.FactoryIdentityTypeSystemAssigned), - // PrincipalID: to.Ptr("10743799-44d2-42fe-8c4d-5bc5c51c0684"), - // TenantID: to.Ptr("12345678-1234-1234-1234-123456789abc"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-15T08:56:07.182Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2017-09-01-preview"), - // }, - // }, - // { - // Name: to.Ptr("FactoryToUpgrade"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00003d04-0000-0000-0000-5b28962f0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/factorytoupgrade"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:35:35.713Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2018-06-01"), - // }, - // }, - // { - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00004004-0000-0000-0000-5b28979e0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // "exampleTag": to.Ptr("exampleValue"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:41:50.004Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // PurviewConfiguration: &armdatafactory.PurviewConfiguration{ - // PurviewResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.Purview/accounts/examplePurview"), - // }, - // RepoConfiguration: &armdatafactory.FactoryVSTSConfiguration{ - // Type: to.Ptr("FactoryVSTSConfiguration"), - // AccountName: to.Ptr("ADF"), - // CollaborationBranch: to.Ptr("master"), - // LastCommitID: to.Ptr(""), - // RepositoryName: to.Ptr("repo"), - // RootFolder: to.Ptr("/"), - // ProjectName: to.Ptr("project"), - // TenantID: to.Ptr(""), - // }, - // Version: to.Ptr("2018-06-01"), - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_ConfigureFactoryRepo.json -func ExampleFactoriesClient_ConfigureFactoryRepo() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewFactoriesClient().ConfigureFactoryRepo(ctx, "East US", armdatafactory.FactoryRepoUpdate{ - FactoryResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - RepoConfiguration: &armdatafactory.FactoryVSTSConfiguration{ - Type: to.Ptr("FactoryVSTSConfiguration"), - AccountName: to.Ptr("ADF"), - CollaborationBranch: to.Ptr("master"), - LastCommitID: to.Ptr(""), - RepositoryName: to.Ptr("repo"), - RootFolder: to.Ptr("/"), - ProjectName: to.Ptr("project"), - TenantID: to.Ptr(""), - }, - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.Factory = armdatafactory.Factory{ - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00004004-0000-0000-0000-5b28979e0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // "exampleTag": to.Ptr("exampleValue"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:41:50.004Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // RepoConfiguration: &armdatafactory.FactoryVSTSConfiguration{ - // Type: to.Ptr("FactoryVSTSConfiguration"), - // AccountName: to.Ptr("ADF"), - // CollaborationBranch: to.Ptr("master"), - // LastCommitID: to.Ptr(""), - // RepositoryName: to.Ptr("repo"), - // RootFolder: to.Ptr("/"), - // ProjectName: to.Ptr("project"), - // TenantID: to.Ptr(""), - // }, - // Version: to.Ptr("2018-06-01"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_ListByResourceGroup.json -func ExampleFactoriesClient_NewListByResourceGroupPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewFactoriesClient().NewListByResourceGroupPager("exampleResourceGroup", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.FactoryListResponse = armdatafactory.FactoryListResponse{ - // Value: []*armdatafactory.Factory{ - // { - // Name: to.Ptr("exampleFactoryName-linked"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00008a02-0000-0000-0000-5b237f270000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName-linked"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Identity: &armdatafactory.FactoryIdentity{ - // Type: to.Ptr(armdatafactory.FactoryIdentityTypeSystemAssigned), - // PrincipalID: to.Ptr("10743799-44d2-42fe-8c4d-5bc5c51c0684"), - // TenantID: to.Ptr("12345678-1234-1234-1234-123456789abc"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-15T08:56:07.182Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2017-09-01-preview"), - // }, - // }, - // { - // Name: to.Ptr("FactoryToUpgrade"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00003d04-0000-0000-0000-5b28962f0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/factorytoupgrade"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:35:35.713Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2018-06-01"), - // }, - // }, - // { - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00004004-0000-0000-0000-5b28979e0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // "exampleTag": to.Ptr("exampleValue"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:41:50.004Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // RepoConfiguration: &armdatafactory.FactoryVSTSConfiguration{ - // Type: to.Ptr("FactoryVSTSConfiguration"), - // AccountName: to.Ptr("ADF"), - // CollaborationBranch: to.Ptr("master"), - // LastCommitID: to.Ptr(""), - // RepositoryName: to.Ptr("repo"), - // RootFolder: to.Ptr("/"), - // ProjectName: to.Ptr("project"), - // TenantID: to.Ptr(""), - // }, - // Version: to.Ptr("2018-06-01"), - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_CreateOrUpdate.json -func ExampleFactoriesClient_CreateOrUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewFactoriesClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.Factory{ - Location: to.Ptr("East US"), - }, &armdatafactory.FactoriesClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.Factory = armdatafactory.Factory{ - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00003e04-0000-0000-0000-5b28979e0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:41:50.004Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2018-06-01"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_Update.json -func ExampleFactoriesClient_Update() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewFactoriesClient().Update(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.FactoryUpdateParameters{ - Tags: map[string]*string{ - "exampleTag": to.Ptr("exampleValue"), - }, - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.Factory = armdatafactory.Factory{ - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00003f04-0000-0000-0000-5b28979e0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // "exampleTag": to.Ptr("exampleValue"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:41:50.004Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // Version: to.Ptr("2018-06-01"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_Get.json -func ExampleFactoriesClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewFactoriesClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", &armdatafactory.FactoriesClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.Factory = armdatafactory.Factory{ - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories"), - // ETag: to.Ptr("\"00004004-0000-0000-0000-5b28979e0000\""), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Location: to.Ptr("East US"), - // Tags: map[string]*string{ - // "exampleTag": to.Ptr("exampleValue"), - // }, - // Properties: &armdatafactory.FactoryProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-19T05:41:50.004Z"); return t}()), - // ProvisioningState: to.Ptr("Succeeded"), - // PurviewConfiguration: &armdatafactory.PurviewConfiguration{ - // PurviewResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.Purview/accounts/examplePurview"), - // }, - // RepoConfiguration: &armdatafactory.FactoryVSTSConfiguration{ - // Type: to.Ptr("FactoryVSTSConfiguration"), - // AccountName: to.Ptr("ADF"), - // CollaborationBranch: to.Ptr("master"), - // LastCommitID: to.Ptr(""), - // RepositoryName: to.Ptr("repo"), - // RootFolder: to.Ptr("/"), - // ProjectName: to.Ptr("project"), - // TenantID: to.Ptr(""), - // }, - // Version: to.Ptr("2018-06-01"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_Delete.json -func ExampleFactoriesClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewFactoriesClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_GetGitHubAccessToken.json -func ExampleFactoriesClient_GetGitHubAccessToken() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewFactoriesClient().GetGitHubAccessToken(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.GitHubAccessTokenRequest{ - GitHubAccessCode: to.Ptr("some"), - GitHubAccessTokenBaseURL: to.Ptr("some"), - GitHubClientID: to.Ptr("some"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.GitHubAccessTokenResponse = armdatafactory.GitHubAccessTokenResponse{ - // GitHubAccessToken: to.Ptr("myAccessTokenExample"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_GetDataPlaneAccess.json -func ExampleFactoriesClient_GetDataPlaneAccess() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewFactoriesClient().GetDataPlaneAccess(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.UserAccessPolicy{ - AccessResourcePath: to.Ptr(""), - ExpireTime: to.Ptr("2018-11-10T09:46:20.2659347Z"), - Permissions: to.Ptr("r"), - ProfileName: to.Ptr("DefaultProfile"), - StartTime: to.Ptr("2018-11-10T02:46:20.2659347Z"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.AccessPolicyResponse = armdatafactory.AccessPolicyResponse{ - // AccessToken: to.Ptr("**********"), - // DataPlaneURL: to.Ptr("https://rpeastus.svc.datafactory.azure.com:4433"), - // Policy: &armdatafactory.UserAccessPolicy{ - // AccessResourcePath: to.Ptr(""), - // ExpireTime: to.Ptr("2018-11-10T09:46:20.2659347Z"), - // Permissions: to.Ptr("r"), - // ProfileName: to.Ptr("DefaultProfile"), - // StartTime: to.Ptr("2018-11-10T02:46:20.2659347Z"), - // }, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/activityruns_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/activityruns_server.go index d4f0cd0206d3..f35e056eb378 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/activityruns_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/activityruns_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/changedatacapture_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/changedatacapture_server.go index 99fa45f9061d..a9c509e943fe 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/changedatacapture_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/changedatacapture_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/credentialoperations_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/credentialoperations_server.go index 29faa8799377..43ac373f436e 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/credentialoperations_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/credentialoperations_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflowdebugsession_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflowdebugsession_server.go index 05048812c4e6..5677a8bfbce5 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflowdebugsession_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflowdebugsession_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflows_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflows_server.go index acd7840e247e..0af095a3ff6d 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflows_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/dataflows_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/datasets_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/datasets_server.go index 789cb6ee63f3..68f449c776e2 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/datasets_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/datasets_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/exposurecontrol_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/exposurecontrol_server.go index b8092dedda79..6c492d9052a9 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/exposurecontrol_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/exposurecontrol_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/factories_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/factories_server.go index dcc5eec1ade5..7dac2b396254 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/factories_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/factories_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/globalparameters_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/globalparameters_server.go index 0afdd3ff4e20..bd34cb7f4d4d 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/globalparameters_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/globalparameters_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimenodes_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimenodes_server.go index c411c7f21203..d3bb4582c902 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimenodes_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimenodes_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimeobjectmetadata_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimeobjectmetadata_server.go index 2bec7d92e69e..ecaaa74c2c78 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimeobjectmetadata_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimeobjectmetadata_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "reflect" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimes_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimes_server.go index 50e026d56d7f..f9816c2e4bec 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimes_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/integrationruntimes_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/linkedservices_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/linkedservices_server.go index 39124efc71a2..c08ded02a265 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/linkedservices_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/linkedservices_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/managedprivateendpoints_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/managedprivateendpoints_server.go index efe4ef8cbd3d..af96c35e17ad 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/managedprivateendpoints_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/managedprivateendpoints_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/managedvirtualnetworks_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/managedvirtualnetworks_server.go index fb77c0b801b0..9cb7e0d2e1a2 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/managedvirtualnetworks_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/managedvirtualnetworks_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/operations_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/operations_server.go index 57a8307b1b33..648245ff960e 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/operations_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/operations_server.go @@ -15,7 +15,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" ) diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelineruns_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelineruns_server.go index a4b4d7fbc582..264bd7b43077 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelineruns_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelineruns_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelines_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelines_server.go index b89763eb792b..59b41d58e356 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelines_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/pipelines_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "reflect" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnection_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnection_server.go index a4206ede6255..90f2988cb914 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnection_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnection_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnections_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnections_server.go index 827b2eb76188..bf9ec5faa92f 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnections_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/privateendpointconnections_server.go @@ -15,7 +15,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/privatelinkresources_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/privatelinkresources_server.go index 7155fd7c54d7..7c86b5294df0 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/privatelinkresources_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/privatelinkresources_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/triggerruns_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/triggerruns_server.go index 5a330b02dac5..d9552700fddf 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/triggerruns_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/triggerruns_server.go @@ -15,7 +15,7 @@ import ( azfake "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake" "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/fake/triggers_server.go b/sdk/resourcemanager/datafactory/armdatafactory/fake/triggers_server.go index cd6232dbd198..db0449446b06 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/fake/triggers_server.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/fake/triggers_server.go @@ -16,7 +16,7 @@ import ( "github.com/Azure/azure-sdk-for-go/sdk/azcore/fake/server" "github.com/Azure/azure-sdk-for-go/sdk/azcore/runtime" "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" + "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5" "net/http" "net/url" "regexp" diff --git a/sdk/resourcemanager/datafactory/armdatafactory/globalparameters_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/globalparameters_client_example_test.go deleted file mode 100644 index 125c75c4a986..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/globalparameters_client_example_test.go +++ /dev/null @@ -1,209 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/GlobalParameters_ListByFactory.json -func ExampleGlobalParametersClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewGlobalParametersClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.GlobalParameterListResponse = armdatafactory.GlobalParameterListResponse{ - // Value: []*armdatafactory.GlobalParameterResource{ - // { - // Name: to.Ptr("default"), - // Type: to.Ptr("Microsoft.DataFactory/factories/globalParameters"), - // Etag: to.Ptr("da00a1c3-0000-0400-0000-6241f3290000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/globalParameters/default"), - // Properties: map[string]*armdatafactory.GlobalParameterSpecification{ - // "copyPipelineTest": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeObject), - // Value: map[string]any{ - // "mySinkDatasetFolderPath": "exampleOutput", - // "mySourceDatasetFolderPath": "exampleInput/", - // "testingEmptyContextParams": "", - // }, - // }, - // "mySourceDatasetFolderPath": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeString), - // Value: "input", - // }, - // "url": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeString), - // Value: "https://testuri.test", - // }, - // "validADFOffice365Uris": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeArray), - // Value: []any{ - // "https://testuri.test", - // "https://testuri.test", - // }, - // }, - // "waitTime": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeInt), - // Value: float64(5), - // }, - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/GlobalParameters_Get.json -func ExampleGlobalParametersClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewGlobalParametersClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "default", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.GlobalParameterResource = armdatafactory.GlobalParameterResource{ - // Name: to.Ptr("default"), - // Type: to.Ptr("Microsoft.DataFactory/factories/globalParameters"), - // Etag: to.Ptr("72001a6a-0000-0400-0000-623d058f0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/globalParameters/default"), - // Properties: map[string]*armdatafactory.GlobalParameterSpecification{ - // "waitTime": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeInt), - // Value: float64(10), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/GlobalParameters_Create.json -func ExampleGlobalParametersClient_CreateOrUpdate_globalParametersCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewGlobalParametersClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "default", armdatafactory.GlobalParameterResource{ - Properties: map[string]*armdatafactory.GlobalParameterSpecification{ - "waitTime": { - Type: to.Ptr(armdatafactory.GlobalParameterTypeInt), - Value: float64(5), - }, - }, - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.GlobalParameterResource = armdatafactory.GlobalParameterResource{ - // Name: to.Ptr("default"), - // Type: to.Ptr("Microsoft.DataFactory/factories/globalParameters"), - // Etag: to.Ptr("0a008ad4-0000-0000-0000-5b245c6e0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/globalParameters/default"), - // Properties: map[string]*armdatafactory.GlobalParameterSpecification{ - // "waitTime": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeInt), - // Value: float64(5), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/GlobalParameters_Update.json -func ExampleGlobalParametersClient_CreateOrUpdate_globalParametersUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewGlobalParametersClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "default", armdatafactory.GlobalParameterResource{ - Properties: map[string]*armdatafactory.GlobalParameterSpecification{ - "waitTime": { - Type: to.Ptr(armdatafactory.GlobalParameterTypeInt), - Value: float64(5), - }, - }, - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.GlobalParameterResource = armdatafactory.GlobalParameterResource{ - // Name: to.Ptr("default"), - // Type: to.Ptr("Microsoft.DataFactory/factories/globalParameters"), - // Etag: to.Ptr("0a008ad4-0000-0000-0000-5b245c6e0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/globalParameters/default"), - // Properties: map[string]*armdatafactory.GlobalParameterSpecification{ - // "waitTime": &armdatafactory.GlobalParameterSpecification{ - // Type: to.Ptr(armdatafactory.GlobalParameterTypeInt), - // Value: float64(5), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/GlobalParameters_Delete.json -func ExampleGlobalParametersClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewGlobalParametersClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "default", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/go.mod b/sdk/resourcemanager/datafactory/armdatafactory/go.mod index d2cc0abd969f..b8d3e7bed049 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/go.mod +++ b/sdk/resourcemanager/datafactory/armdatafactory/go.mod @@ -1,10 +1,9 @@ -module github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4 +module github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v5 go 1.18 require ( github.com/Azure/azure-sdk-for-go/sdk/azcore v1.9.0 - github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.4.0 github.com/Azure/azure-sdk-for-go/sdk/internal v1.5.0 github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/internal/v2 v2.0.0 github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/resources/armresources v1.1.1 @@ -12,6 +11,7 @@ require ( ) require ( + github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.4.0 // indirect github.com/AzureAD/microsoft-authentication-library-for-go v1.1.1 // indirect github.com/davecgh/go-spew v1.1.1 // indirect github.com/dnaeon/go-vcr v1.2.0 // indirect diff --git a/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimenodes_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimenodes_client_example_test.go deleted file mode 100644 index ee44c6480c70..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimenodes_client_example_test.go +++ /dev/null @@ -1,144 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimeNodes_Get.json -func ExampleIntegrationRuntimeNodesClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimeNodesClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", "Node_1", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.SelfHostedIntegrationRuntimeNode = armdatafactory.SelfHostedIntegrationRuntimeNode{ - // Capabilities: map[string]*string{ - // "connectedToResourceManager": to.Ptr("True"), - // "credentialInSync": to.Ptr("True"), - // "httpsPortEnabled": to.Ptr("True"), - // "nodeEnabled": to.Ptr("True"), - // "serviceBusConnected": to.Ptr("True"), - // }, - // HostServiceURI: to.Ptr("https://yanzhang-dt.fareast.corp.microsoft.com:8050/HostServiceRemote.svc/"), - // IsActiveDispatcher: to.Ptr(true), - // LastConnectTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T06:30:46.626Z"); return t}()), - // LastStartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T03:45:30.849Z"); return t}()), - // LastUpdateResult: to.Ptr(armdatafactory.IntegrationRuntimeUpdateResultNone), - // MachineName: to.Ptr("YANZHANG-DT"), - // MaxConcurrentJobs: to.Ptr[int32](20), - // NodeName: to.Ptr("Node_1"), - // RegisterTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T03:44:55.801Z"); return t}()), - // Status: to.Ptr(armdatafactory.SelfHostedIntegrationRuntimeNodeStatusOnline), - // Version: to.Ptr("3.8.6743.6"), - // VersionStatus: to.Ptr("UpToDate"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimeNodes_Delete.json -func ExampleIntegrationRuntimeNodesClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewIntegrationRuntimeNodesClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", "Node_1", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimeNodes_Update.json -func ExampleIntegrationRuntimeNodesClient_Update() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimeNodesClient().Update(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", "Node_1", armdatafactory.UpdateIntegrationRuntimeNodeRequest{ - ConcurrentJobsLimit: to.Ptr[int32](2), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.SelfHostedIntegrationRuntimeNode = armdatafactory.SelfHostedIntegrationRuntimeNode{ - // Capabilities: map[string]*string{ - // "connectedToResourceManager": to.Ptr("True"), - // "credentialInSync": to.Ptr("True"), - // "httpsPortEnabled": to.Ptr("True"), - // "nodeEnabled": to.Ptr("True"), - // "serviceBusConnected": to.Ptr("True"), - // }, - // ConcurrentJobsLimit: to.Ptr[int32](2), - // HostServiceURI: to.Ptr("https://yanzhang-dt.fareast.corp.microsoft.com:8050/HostServiceRemote.svc/"), - // IsActiveDispatcher: to.Ptr(true), - // LastConnectTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T15:00:35.754Z"); return t}()), - // LastStartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T14:52:59.893Z"); return t}()), - // LastUpdateResult: to.Ptr(armdatafactory.IntegrationRuntimeUpdateResultNone), - // MachineName: to.Ptr("YANZHANG-DT"), - // MaxConcurrentJobs: to.Ptr[int32](56), - // NodeName: to.Ptr("Node_1"), - // RegisterTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T14:51:44.923Z"); return t}()), - // Status: to.Ptr(armdatafactory.SelfHostedIntegrationRuntimeNodeStatusOnline), - // Version: to.Ptr("3.8.6730.2"), - // VersionStatus: to.Ptr("UpToDate"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimeNodes_GetIpAddress.json -func ExampleIntegrationRuntimeNodesClient_GetIPAddress() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimeNodesClient().GetIPAddress(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", "Node_1", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeNodeIPAddress = armdatafactory.IntegrationRuntimeNodeIPAddress{ - // IPAddress: to.Ptr("**********"), - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimeobjectmetadata_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimeobjectmetadata_client_example_test.go deleted file mode 100644 index f28ae4aaba18..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimeobjectmetadata_client_example_test.go +++ /dev/null @@ -1,577 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimeObjectMetadata_Refresh.json -func ExampleIntegrationRuntimeObjectMetadataClient_BeginRefresh() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewIntegrationRuntimeObjectMetadataClient().BeginRefresh(ctx, "exampleResourceGroup", "exampleFactoryName", "testactivityv2", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - res, err := poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.SsisObjectMetadataStatusResponse = armdatafactory.SsisObjectMetadataStatusResponse{ - // Name: to.Ptr("ca63c855b72d44959653ffcc6eb0b96c"), - // Status: to.Ptr("Succeeded"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimeObjectMetadata_Get.json -func ExampleIntegrationRuntimeObjectMetadataClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimeObjectMetadataClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "testactivityv2", &armdatafactory.IntegrationRuntimeObjectMetadataClientGetOptions{GetMetadataRequest: &armdatafactory.GetSsisObjectMetadataRequest{ - MetadataPath: to.Ptr("ssisFolders"), - }, - }) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.SsisObjectMetadataListResponse = armdatafactory.SsisObjectMetadataListResponse{ - // Value: []armdatafactory.SsisObjectMetadataClassification{ - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("TestFolder"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // Description: to.Ptr(""), - // ID: to.Ptr[int64](1), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("EnvironmentFolder"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // Description: to.Ptr(""), - // ID: to.Ptr[int64](2), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("ActivityTest"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // Description: to.Ptr(""), - // ID: to.Ptr[int64](3), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("这是文件夹"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // Description: to.Ptr(""), - // ID: to.Ptr[int64](4), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("1"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](5), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("2"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](6), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("3"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](7), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("4"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](8), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("5"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](9), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("6"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](10), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("7"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](11), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("8"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](12), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("9"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](13), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("10"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](14), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("11"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](15), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("12"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](16), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("13"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](17), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("14"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](18), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("15"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](19), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("16"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](20), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("17"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](21), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("18"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](22), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("19"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](23), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("20"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](24), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("21"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](25), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("22"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](26), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("23"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](27), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("24"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](28), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("25"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](29), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("26"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](30), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("27"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](31), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("28"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](32), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("29"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](33), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("30"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](34), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("31"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](35), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("32"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](36), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("33"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](37), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("34"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](38), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("35"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](39), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("36"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](40), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("37"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](41), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("38"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](42), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("39"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](43), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("40"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](44), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("41"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](45), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("42"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](46), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("43"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](47), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("44"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](48), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("45"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](49), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("46"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](50), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("47"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](51), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("48"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](52), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("49"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](53), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("50"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](54), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("51"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](55), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("52"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](56), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("53"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](57), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("54"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](58), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("55"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](59), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("56"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](60), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("57"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](61), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("58"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](62), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("59"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](63), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("60"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](64), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("61"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](65), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("62"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](66), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("63"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](67), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("64"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](68), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("65"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](69), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("66"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](70), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("67"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](71), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("68"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](72), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("69"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](73), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("70"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](74), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("71"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](75), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("72"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](76), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("73"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](77), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("74"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](78), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("75"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](79), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("76"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](80), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("77"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](81), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("78"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](82), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("79"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](83), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("80"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](84), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("81"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](85), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("82"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](86), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("83"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](87), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("84"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](88), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("85"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](89), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("86"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](90), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("87"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](91), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("88"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](92), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("89"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](93), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("90"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](94), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("91"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](95), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("92"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](96), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("93"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](97), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("94"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](98), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("95"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](99), - // }, - // &armdatafactory.SsisFolder{ - // Name: to.Ptr("96"), - // Type: to.Ptr(armdatafactory.SsisObjectMetadataTypeFolder), - // ID: to.Ptr[int64](100), - // }}, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimes_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimes_client_example_test.go deleted file mode 100644 index aff38b458ac6..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/integrationruntimes_client_example_test.go +++ /dev/null @@ -1,638 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_ListByFactory.json -func ExampleIntegrationRuntimesClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewIntegrationRuntimesClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.IntegrationRuntimeListResponse = armdatafactory.IntegrationRuntimeListResponse{ - // Value: []*armdatafactory.IntegrationRuntimeResource{ - // { - // Name: to.Ptr("exampleIntegrationRuntime"), - // Type: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes"), - // Etag: to.Ptr("0400f1a1-0000-0000-0000-5b2188640000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/integrationruntimes/exampleIntegrationRuntime"), - // Properties: &armdatafactory.SelfHostedIntegrationRuntime{ - // Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeSelfHosted), - // Description: to.Ptr("A selfhosted integration runtime"), - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_Create.json -func ExampleIntegrationRuntimesClient_CreateOrUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", armdatafactory.IntegrationRuntimeResource{ - Properties: &armdatafactory.SelfHostedIntegrationRuntime{ - Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeSelfHosted), - Description: to.Ptr("A selfhosted integration runtime"), - }, - }, &armdatafactory.IntegrationRuntimesClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeResource = armdatafactory.IntegrationRuntimeResource{ - // Name: to.Ptr("exampleIntegrationRuntime"), - // Type: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes"), - // Etag: to.Ptr("000046c4-0000-0000-0000-5b2198bf0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/integrationruntimes/exampleIntegrationRuntime"), - // Properties: &armdatafactory.SelfHostedIntegrationRuntime{ - // Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeSelfHosted), - // Description: to.Ptr("A selfhosted integration runtime"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_Get.json -func ExampleIntegrationRuntimesClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", &armdatafactory.IntegrationRuntimesClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeResource = armdatafactory.IntegrationRuntimeResource{ - // Name: to.Ptr("exampleIntegrationRuntime"), - // Type: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes"), - // Etag: to.Ptr("15003c4f-0000-0200-0000-5cbe090b0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/integrationruntimes/exampleIntegrationRuntime"), - // Properties: &armdatafactory.SelfHostedIntegrationRuntime{ - // Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeSelfHosted), - // Description: to.Ptr("A selfhosted integration runtime"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_Update.json -func ExampleIntegrationRuntimesClient_Update() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().Update(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", armdatafactory.UpdateIntegrationRuntimeRequest{ - AutoUpdate: to.Ptr(armdatafactory.IntegrationRuntimeAutoUpdateOff), - UpdateDelayOffset: to.Ptr("\"PT3H\""), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeResource = armdatafactory.IntegrationRuntimeResource{ - // Name: to.Ptr("exampleIntegrationRuntime"), - // Type: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes"), - // Etag: to.Ptr("0400f1a1-0000-0000-0000-5b2188640000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/integrationruntimes/exampleIntegrationRuntime"), - // Properties: &armdatafactory.SelfHostedIntegrationRuntime{ - // Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeSelfHosted), - // Description: to.Ptr("A selfhosted integration runtime"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_Delete.json -func ExampleIntegrationRuntimesClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewIntegrationRuntimesClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_GetStatus.json -func ExampleIntegrationRuntimesClient_GetStatus() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().GetStatus(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeStatusResponse = armdatafactory.IntegrationRuntimeStatusResponse{ - // Name: to.Ptr("exampleIntegrationRuntime"), - // Properties: &armdatafactory.SelfHostedIntegrationRuntimeStatus{ - // Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeSelfHosted), - // State: to.Ptr(armdatafactory.IntegrationRuntimeStateOnline), - // TypeProperties: &armdatafactory.SelfHostedIntegrationRuntimeStatusTypeProperties{ - // AutoUpdate: to.Ptr(armdatafactory.IntegrationRuntimeAutoUpdateOff), - // Capabilities: map[string]*string{ - // "connectedToResourceManager": to.Ptr("True"), - // "credentialInSync": to.Ptr("True"), - // "httpsPortEnabled": to.Ptr("True"), - // "nodeEnabled": to.Ptr("True"), - // "serviceBusConnected": to.Ptr("True"), - // }, - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T09:17:45.183Z"); return t}()), - // LatestVersion: to.Ptr("3.7.6711.1"), - // LocalTimeZoneOffset: to.Ptr("PT8H"), - // Nodes: []*armdatafactory.SelfHostedIntegrationRuntimeNode{ - // { - // Capabilities: map[string]*string{ - // "connectedToResourceManager": to.Ptr("True"), - // "credentialInSync": to.Ptr("True"), - // "httpsPortEnabled": to.Ptr("True"), - // "nodeEnabled": to.Ptr("True"), - // "serviceBusConnected": to.Ptr("True"), - // }, - // HostServiceURI: to.Ptr("https://yanzhang-dt.fareast.corp.microsoft.com:8050/HostServiceRemote.svc/"), - // IsActiveDispatcher: to.Ptr(true), - // LastConnectTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T14:52:59.893Z"); return t}()), - // LastStartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T14:52:59.893Z"); return t}()), - // LastUpdateResult: to.Ptr(armdatafactory.IntegrationRuntimeUpdateResultNone), - // MachineName: to.Ptr("YANZHANG-DT"), - // MaxConcurrentJobs: to.Ptr[int32](56), - // NodeName: to.Ptr("Node_1"), - // RegisterTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-14T14:51:44.923Z"); return t}()), - // Status: to.Ptr(armdatafactory.SelfHostedIntegrationRuntimeNodeStatusOnline), - // Version: to.Ptr("3.8.6730.2"), - // VersionStatus: to.Ptr("UpToDate"), - // }}, - // ServiceUrls: []*string{ - // to.Ptr("wu.frontend.int.clouddatahub-int.net"), - // to.Ptr("*.servicebus.windows.net")}, - // TaskQueueID: to.Ptr("1a6296ab-423c-4346-9bcc-85a78c2c0582"), - // UpdateDelayOffset: to.Ptr("PT3H"), - // Version: to.Ptr("3.8.6730.2"), - // VersionStatus: to.Ptr("UpToDate"), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_ListOutboundNetworkDependenciesEndpoints.json -func ExampleIntegrationRuntimesClient_ListOutboundNetworkDependenciesEndpoints() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().ListOutboundNetworkDependenciesEndpoints(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse = armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse{ - // Value: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint{ - // { - // Category: to.Ptr("Azure Data Factory (Management)"), - // Endpoints: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpoint{ - // { - // DomainName: to.Ptr("wu.frontend.int.clouddatahub-int.net"), - // EndpointDetails: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails{ - // { - // Port: to.Ptr[int32](443), - // }}, - // }}, - // }, - // { - // Category: to.Ptr("Azure Storage (Management)"), - // Endpoints: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpoint{ - // { - // DomainName: to.Ptr("*.blob.core.windows.net"), - // EndpointDetails: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails{ - // { - // Port: to.Ptr[int32](443), - // }}, - // }, - // { - // DomainName: to.Ptr("*.table.core.windows.net"), - // EndpointDetails: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails{ - // { - // Port: to.Ptr[int32](443), - // }}, - // }}, - // }, - // { - // Category: to.Ptr("Event Hub (Logging)"), - // Endpoints: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpoint{ - // { - // DomainName: to.Ptr("*.servicebus.windows.net"), - // EndpointDetails: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails{ - // { - // Port: to.Ptr[int32](443), - // }}, - // }}, - // }, - // { - // Category: to.Ptr("Microsoft Logging service (Internal Use)"), - // Endpoints: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpoint{ - // { - // DomainName: to.Ptr("gcs.prod.monitoring.core.windows.net"), - // EndpointDetails: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails{ - // { - // Port: to.Ptr[int32](443), - // }}, - // }, - // { - // DomainName: to.Ptr("prod.warmpath.msftcloudes.com"), - // EndpointDetails: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails{ - // { - // Port: to.Ptr[int32](443), - // }}, - // }, - // { - // DomainName: to.Ptr("azurewatsonanalysis-prod.core.windows.net"), - // EndpointDetails: []*armdatafactory.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails{ - // { - // Port: to.Ptr[int32](443), - // }}, - // }}, - // }}, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_GetConnectionInfo.json -func ExampleIntegrationRuntimesClient_GetConnectionInfo() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().GetConnectionInfo(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeConnectionInfo = armdatafactory.IntegrationRuntimeConnectionInfo{ - // HostServiceURI: to.Ptr("https://yanzhang-dt.fareast.corp.microsoft.com:8050/HostServiceRemote.svc/"), - // IdentityCertThumbprint: to.Ptr("**********"), - // IsIdentityCertExprired: to.Ptr(false), - // PublicKey: to.Ptr("**********"), - // ServiceToken: to.Ptr("**********"), - // Version: to.Ptr("3.8.6730.2"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_RegenerateAuthKey.json -func ExampleIntegrationRuntimesClient_RegenerateAuthKey() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().RegenerateAuthKey(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", armdatafactory.IntegrationRuntimeRegenerateKeyParameters{ - KeyName: to.Ptr(armdatafactory.IntegrationRuntimeAuthKeyNameAuthKey2), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeAuthKeys = armdatafactory.IntegrationRuntimeAuthKeys{ - // AuthKey2: to.Ptr("**********"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_ListAuthKeys.json -func ExampleIntegrationRuntimesClient_ListAuthKeys() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().ListAuthKeys(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeAuthKeys = armdatafactory.IntegrationRuntimeAuthKeys{ - // AuthKey1: to.Ptr("**********"), - // AuthKey2: to.Ptr("**********"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_Start.json -func ExampleIntegrationRuntimesClient_BeginStart() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewIntegrationRuntimesClient().BeginStart(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleManagedIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - res, err := poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeStatusResponse = armdatafactory.IntegrationRuntimeStatusResponse{ - // Name: to.Ptr("exampleManagedIntegrationRuntime"), - // Properties: &armdatafactory.ManagedIntegrationRuntimeStatus{ - // Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeManaged), - // DataFactoryName: to.Ptr("exampleFactoryName"), - // State: to.Ptr(armdatafactory.IntegrationRuntimeStateStarted), - // TypeProperties: &armdatafactory.ManagedIntegrationRuntimeStatusTypeProperties{ - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-13T21:11:01.869Z"); return t}()), - // Nodes: []*armdatafactory.ManagedIntegrationRuntimeNode{ - // }, - // OtherErrors: []*armdatafactory.ManagedIntegrationRuntimeError{ - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_Stop.json -func ExampleIntegrationRuntimesClient_BeginStop() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewIntegrationRuntimesClient().BeginStop(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleManagedIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - _, err = poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_SyncCredentials.json -func ExampleIntegrationRuntimesClient_SyncCredentials() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewIntegrationRuntimesClient().SyncCredentials(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_GetMonitoringData.json -func ExampleIntegrationRuntimesClient_GetMonitoringData() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().GetMonitoringData(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeMonitoringData = armdatafactory.IntegrationRuntimeMonitoringData{ - // Name: to.Ptr("exampleIntegrationRuntime"), - // Nodes: []*armdatafactory.IntegrationRuntimeNodeMonitoringData{ - // { - // AvailableMemoryInMB: to.Ptr[int32](16740), - // ConcurrentJobsLimit: to.Ptr[int32](28), - // ConcurrentJobsRunning: to.Ptr[int32](0), - // CPUUtilization: to.Ptr[int32](15), - // NodeName: to.Ptr("Node_1"), - // ReceivedBytes: to.Ptr[float32](6.731423377990723), - // SentBytes: to.Ptr[float32](2.647491693496704), - // }}, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_Upgrade.json -func ExampleIntegrationRuntimesClient_Upgrade() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewIntegrationRuntimesClient().Upgrade(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_RemoveLinks.json -func ExampleIntegrationRuntimesClient_RemoveLinks() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewIntegrationRuntimesClient().RemoveLinks(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", armdatafactory.LinkedIntegrationRuntimeRequest{ - LinkedFactoryName: to.Ptr("exampleFactoryName-linked"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/IntegrationRuntimes_CreateLinkedIntegrationRuntime.json -func ExampleIntegrationRuntimesClient_CreateLinkedIntegrationRuntime() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewIntegrationRuntimesClient().CreateLinkedIntegrationRuntime(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleIntegrationRuntime", armdatafactory.CreateLinkedIntegrationRuntimeRequest{ - Name: to.Ptr("bfa92911-9fb6-4fbe-8f23-beae87bc1c83"), - DataFactoryLocation: to.Ptr("West US"), - DataFactoryName: to.Ptr("e9955d6d-56ea-4be3-841c-52a12c1a9981"), - SubscriptionID: to.Ptr("061774c7-4b5a-4159-a55b-365581830283"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.IntegrationRuntimeStatusResponse = armdatafactory.IntegrationRuntimeStatusResponse{ - // Name: to.Ptr("exampleIntegrationRuntime"), - // Properties: &armdatafactory.SelfHostedIntegrationRuntimeStatus{ - // Type: to.Ptr(armdatafactory.IntegrationRuntimeTypeSelfHosted), - // DataFactoryName: to.Ptr("exampleFactoryName"), - // State: to.Ptr(armdatafactory.IntegrationRuntimeStateOnline), - // TypeProperties: &armdatafactory.SelfHostedIntegrationRuntimeStatusTypeProperties{ - // AutoUpdate: to.Ptr(armdatafactory.IntegrationRuntimeAutoUpdateOn), - // AutoUpdateETA: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-20T19:00:00.000Z"); return t}()), - // Capabilities: map[string]*string{ - // "connectedToResourceManager": to.Ptr("True"), - // "credentialInSync": to.Ptr("True"), - // "httpsPortEnabled": to.Ptr("True"), - // "nodeEnabled": to.Ptr("True"), - // "serviceBusConnected": to.Ptr("True"), - // }, - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T03:43:25.705Z"); return t}()), - // LatestVersion: to.Ptr("3.9.6774.1"), - // Links: []*armdatafactory.LinkedIntegrationRuntime{ - // { - // Name: to.Ptr("bfa92911-9fb6-4fbe-8f23-beae87bc1c83"), - // CreateTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T06:31:04.061Z"); return t}()), - // DataFactoryLocation: to.Ptr("West US"), - // DataFactoryName: to.Ptr("e9955d6d-56ea-4be3-841c-52a12c1a9981"), - // SubscriptionID: to.Ptr("061774c7-4b5a-4159-a55b-365581830283"), - // }}, - // LocalTimeZoneOffset: to.Ptr("PT8H"), - // Nodes: []*armdatafactory.SelfHostedIntegrationRuntimeNode{ - // { - // Capabilities: map[string]*string{ - // "connectedToResourceManager": to.Ptr("True"), - // "credentialInSync": to.Ptr("True"), - // "httpsPortEnabled": to.Ptr("True"), - // "nodeEnabled": to.Ptr("True"), - // "serviceBusConnected": to.Ptr("True"), - // }, - // HostServiceURI: to.Ptr("https://yanzhang-dt.fareast.corp.microsoft.com:8050/HostServiceRemote.svc/"), - // IsActiveDispatcher: to.Ptr(true), - // LastConnectTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T06:30:46.626Z"); return t}()), - // LastStartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T03:45:30.849Z"); return t}()), - // LastUpdateResult: to.Ptr(armdatafactory.IntegrationRuntimeUpdateResultNone), - // MachineName: to.Ptr("YANZHANG-DT"), - // MaxConcurrentJobs: to.Ptr[int32](20), - // NodeName: to.Ptr("Node_1"), - // RegisterTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-17T03:44:55.801Z"); return t}()), - // Status: to.Ptr(armdatafactory.SelfHostedIntegrationRuntimeNodeStatusOnline), - // Version: to.Ptr("3.8.6743.6"), - // VersionStatus: to.Ptr("UpToDate"), - // }}, - // PushedVersion: to.Ptr("3.9.6774.1"), - // ScheduledUpdateDate: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-08-20T00:00:00.000Z"); return t}()), - // ServiceUrls: []*string{ - // to.Ptr("wu.frontend.int.clouddatahub-int.net"), - // to.Ptr("*.servicebus.windows.net")}, - // TaskQueueID: to.Ptr("823da112-f2d9-426b-a0d8-5f361b94f72a"), - // UpdateDelayOffset: to.Ptr("PT19H"), - // Version: to.Ptr("3.8.6743.6"), - // VersionStatus: to.Ptr("UpdateAvailable"), - // }, - // }, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/interfaces.go b/sdk/resourcemanager/datafactory/armdatafactory/interfaces.go index 1ab960b15cb1..fb5c0e4f0439 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/interfaces.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/interfaces.go @@ -52,7 +52,8 @@ type ControlActivityClassification interface { // - *CommonDataServiceForAppsSink, *CopySink, *CosmosDbMongoDbAPISink, *CosmosDbSQLAPISink, *DelimitedTextSink, *DocumentDbCollectionSink, // - *DynamicsCrmSink, *DynamicsSink, *FileSystemSink, *InformixSink, *JSONSink, *LakeHouseTableSink, *MicrosoftAccessSink, // - *MongoDbAtlasSink, *MongoDbV2Sink, *OdbcSink, *OracleSink, *OrcSink, *ParquetSink, *RestSink, *SQLDWSink, *SQLMISink, -// - *SQLServerSink, *SQLSink, *SalesforceServiceCloudSink, *SalesforceSink, *SapCloudForCustomerSink, *SnowflakeSink +// - *SQLServerSink, *SQLSink, *SalesforceServiceCloudSink, *SalesforceServiceCloudV2Sink, *SalesforceSink, *SalesforceV2Sink, +// - *SapCloudForCustomerSink, *SnowflakeSink, *SnowflakeV2Sink, *WarehouseSink type CopySinkClassification interface { // GetCopySink returns the CopySink content of the underlying type. GetCopySink() *CopySink @@ -72,9 +73,10 @@ type CopySinkClassification interface { // - *NetezzaSource, *ODataSource, *OdbcSource, *Office365Source, *OracleServiceCloudSource, *OracleSource, *OrcSource, *ParquetSource, // - *PaypalSource, *PhoenixSource, *PostgreSQLSource, *PrestoSource, *QuickBooksSource, *RelationalSource, *ResponsysSource, // - *RestSource, *SQLDWSource, *SQLMISource, *SQLServerSource, *SQLSource, *SalesforceMarketingCloudSource, *SalesforceServiceCloudSource, -// - *SalesforceSource, *SapBwSource, *SapCloudForCustomerSource, *SapEccSource, *SapHanaSource, *SapOdpSource, *SapOpenHubSource, -// - *SapTableSource, *ServiceNowSource, *SharePointOnlineListSource, *ShopifySource, *SnowflakeSource, *SparkSource, *SquareSource, -// - *SybaseSource, *TabularSource, *TeradataSource, *VerticaSource, *WebSource, *XMLSource, *XeroSource, *ZohoSource +// - *SalesforceServiceCloudV2Source, *SalesforceSource, *SalesforceV2Source, *SapBwSource, *SapCloudForCustomerSource, *SapEccSource, +// - *SapHanaSource, *SapOdpSource, *SapOpenHubSource, *SapTableSource, *ServiceNowSource, *SharePointOnlineListSource, *ShopifySource, +// - *SnowflakeSource, *SnowflakeV2Source, *SparkSource, *SquareSource, *SybaseSource, *TabularSource, *TeradataSource, *VerticaSource, +// - *WarehouseSource, *WebSource, *XMLSource, *XeroSource, *ZohoSource type CopySourceClassification interface { // GetCopySource returns the CopySource content of the underlying type. GetCopySource() *CopySource @@ -125,9 +127,10 @@ type DataFlowClassification interface { // - *OracleTableDataset, *OrcDataset, *ParquetDataset, *PaypalObjectDataset, *PhoenixObjectDataset, *PostgreSQLTableDataset, // - *PrestoObjectDataset, *QuickBooksObjectDataset, *RelationalTableDataset, *ResponsysObjectDataset, *RestResourceDataset, // - *SQLServerTableDataset, *SalesforceMarketingCloudObjectDataset, *SalesforceObjectDataset, *SalesforceServiceCloudObjectDataset, -// - *SapBwCubeDataset, *SapCloudForCustomerResourceDataset, *SapEccResourceDataset, *SapHanaTableDataset, *SapOdpResourceDataset, -// - *SapOpenHubTableDataset, *SapTableResourceDataset, *ServiceNowObjectDataset, *SharePointOnlineListResourceDataset, *ShopifyObjectDataset, -// - *SnowflakeDataset, *SparkObjectDataset, *SquareObjectDataset, *SybaseTableDataset, *TeradataTableDataset, *VerticaTableDataset, +// - *SalesforceServiceCloudV2ObjectDataset, *SalesforceV2ObjectDataset, *SapBwCubeDataset, *SapCloudForCustomerResourceDataset, +// - *SapEccResourceDataset, *SapHanaTableDataset, *SapOdpResourceDataset, *SapOpenHubTableDataset, *SapTableResourceDataset, +// - *ServiceNowObjectDataset, *SharePointOnlineListResourceDataset, *ShopifyObjectDataset, *SnowflakeDataset, *SnowflakeV2Dataset, +// - *SparkObjectDataset, *SquareObjectDataset, *SybaseTableDataset, *TeradataTableDataset, *VerticaTableDataset, *WarehouseTableDataset, // - *WebTableDataset, *XMLDataset, *XeroObjectDataset, *ZohoObjectDataset type DatasetClassification interface { // GetDataset returns the Dataset content of the underlying type. @@ -272,11 +275,12 @@ type LinkedIntegrationRuntimeTypeClassification interface { // - *OracleLinkedService, *OracleServiceCloudLinkedService, *PaypalLinkedService, *PhoenixLinkedService, *PostgreSQLLinkedService, // - *PrestoLinkedService, *QuickBooksLinkedService, *QuickbaseLinkedService, *ResponsysLinkedService, *RestServiceLinkedService, // - *SQLServerLinkedService, *SalesforceLinkedService, *SalesforceMarketingCloudLinkedService, *SalesforceServiceCloudLinkedService, -// - *SapBWLinkedService, *SapCloudForCustomerLinkedService, *SapEccLinkedService, *SapHanaLinkedService, *SapOdpLinkedService, -// - *SapOpenHubLinkedService, *SapTableLinkedService, *ServiceNowLinkedService, *SftpServerLinkedService, *SharePointOnlineListLinkedService, -// - *ShopifyLinkedService, *SmartsheetLinkedService, *SnowflakeLinkedService, *SparkLinkedService, *SquareLinkedService, -// - *SybaseLinkedService, *TeamDeskLinkedService, *TeradataLinkedService, *TwilioLinkedService, *VerticaLinkedService, *WebLinkedService, -// - *XeroLinkedService, *ZendeskLinkedService, *ZohoLinkedService +// - *SalesforceServiceCloudV2LinkedService, *SalesforceV2LinkedService, *SapBWLinkedService, *SapCloudForCustomerLinkedService, +// - *SapEccLinkedService, *SapHanaLinkedService, *SapOdpLinkedService, *SapOpenHubLinkedService, *SapTableLinkedService, +// - *ServiceNowLinkedService, *SftpServerLinkedService, *SharePointOnlineListLinkedService, *ShopifyLinkedService, *SmartsheetLinkedService, +// - *SnowflakeLinkedService, *SnowflakeV2LinkedService, *SparkLinkedService, *SquareLinkedService, *SybaseLinkedService, +// - *TeamDeskLinkedService, *TeradataLinkedService, *TwilioLinkedService, *VerticaLinkedService, *WarehouseLinkedService, +// - *WebLinkedService, *XeroLinkedService, *ZendeskLinkedService, *ZohoLinkedService type LinkedServiceClassification interface { // GetLinkedService returns the LinkedService content of the underlying type. GetLinkedService() *LinkedService @@ -340,9 +344,9 @@ type StoreWriteSettingsClassification interface { // - *ImpalaSource, *InformixSource, *JiraSource, *MagentoSource, *MariaDBSource, *MarketoSource, *MySQLSource, *NetezzaSource, // - *OdbcSource, *OracleServiceCloudSource, *PaypalSource, *PhoenixSource, *PostgreSQLSource, *PrestoSource, *QuickBooksSource, // - *ResponsysSource, *SQLDWSource, *SQLMISource, *SQLServerSource, *SQLSource, *SalesforceMarketingCloudSource, *SalesforceSource, -// - *SapBwSource, *SapCloudForCustomerSource, *SapEccSource, *SapHanaSource, *SapOdpSource, *SapOpenHubSource, *SapTableSource, -// - *ServiceNowSource, *ShopifySource, *SparkSource, *SquareSource, *SybaseSource, *TabularSource, *TeradataSource, *VerticaSource, -// - *XeroSource, *ZohoSource +// - *SalesforceV2Source, *SapBwSource, *SapCloudForCustomerSource, *SapEccSource, *SapHanaSource, *SapOdpSource, *SapOpenHubSource, +// - *SapTableSource, *ServiceNowSource, *ShopifySource, *SparkSource, *SquareSource, *SybaseSource, *TabularSource, *TeradataSource, +// - *VerticaSource, *WarehouseSource, *XeroSource, *ZohoSource type TabularSourceClassification interface { CopySourceClassification // GetTabularSource returns the TabularSource content of the underlying type. diff --git a/sdk/resourcemanager/datafactory/armdatafactory/linkedservices_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/linkedservices_client_example_test.go deleted file mode 100644 index 55ca6706f7dd..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/linkedservices_client_example_test.go +++ /dev/null @@ -1,212 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/LinkedServices_ListByFactory.json -func ExampleLinkedServicesClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewLinkedServicesClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.LinkedServiceListResponse = armdatafactory.LinkedServiceListResponse{ - // Value: []*armdatafactory.LinkedServiceResource{ - // { - // Name: to.Ptr("exampleLinkedService"), - // Type: to.Ptr("Microsoft.DataFactory/factories/linkedservices"), - // Etag: to.Ptr("0a0064d4-0000-0000-0000-5b245bd00000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/linkedservices/exampleLinkedService"), - // Properties: &armdatafactory.AzureStorageLinkedService{ - // Type: to.Ptr("AzureStorage"), - // Description: to.Ptr("Example description"), - // TypeProperties: &armdatafactory.AzureStorageLinkedServiceTypeProperties{ - // ConnectionString: map[string]any{ - // "type": "SecureString", - // "value": "**********", - // }, - // EncryptedCredential: to.Ptr("ew0KICAiVmVyc2lvbiI6ICIyMDE3LTExLTMwIiwNCiAgIlByb3RlY3Rpb25Nb2RlIjogIktleSIsDQogICJTZWNyZXRDb250ZW50VHlwZSI6ICJQbGFpbnRleHQiLA0KICAiQ3JlZGVudGlhbElkIjogIkRGLURPR0ZPT0QtWUFOWkhBTkctV1VfMGI2M2EyMmYtMGEzNC00NDg2LWIzMDktNzM0NTlkODUyY2Q1Ig0KfQ=="), - // }, - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/LinkedServices_Create.json -func ExampleLinkedServicesClient_CreateOrUpdate_linkedServicesCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewLinkedServicesClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleLinkedService", armdatafactory.LinkedServiceResource{ - Properties: &armdatafactory.AzureStorageLinkedService{ - Type: to.Ptr("AzureStorage"), - TypeProperties: &armdatafactory.AzureStorageLinkedServiceTypeProperties{ - ConnectionString: map[string]any{ - "type": "SecureString", - "value": "DefaultEndpointsProtocol=https;AccountName=examplestorageaccount;AccountKey=", - }, - }, - }, - }, &armdatafactory.LinkedServicesClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.LinkedServiceResource = armdatafactory.LinkedServiceResource{ - // Name: to.Ptr("exampleLinkedService"), - // Type: to.Ptr("Microsoft.DataFactory/factories/linkedservices"), - // Etag: to.Ptr("0a0062d4-0000-0000-0000-5b245bcf0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/linkedservices/exampleLinkedService"), - // Properties: &armdatafactory.AzureStorageLinkedService{ - // Type: to.Ptr("AzureStorage"), - // TypeProperties: &armdatafactory.AzureStorageLinkedServiceTypeProperties{ - // ConnectionString: map[string]any{ - // "type": "SecureString", - // "value": "**********", - // }, - // EncryptedCredential: to.Ptr("ew0KICAiVmVyc2lvbiI6ICIyMDE3LTExLTMwIiwNCiAgIlByb3RlY3Rpb25Nb2RlIjogIktleSIsDQogICJTZWNyZXRDb250ZW50VHlwZSI6ICJQbGFpbnRleHQiLA0KICAiQ3JlZGVudGlhbElkIjogIkRGLURPR0ZPT0QtWUFOWkhBTkctV1VfM2FiMTk0NjYtNWUxNi00NzU1LWJlNzktMjI2ZTVmZWU3YzY0Ig0KfQ=="), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/LinkedServices_Update.json -func ExampleLinkedServicesClient_CreateOrUpdate_linkedServicesUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewLinkedServicesClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleLinkedService", armdatafactory.LinkedServiceResource{ - Properties: &armdatafactory.AzureStorageLinkedService{ - Type: to.Ptr("AzureStorage"), - Description: to.Ptr("Example description"), - TypeProperties: &armdatafactory.AzureStorageLinkedServiceTypeProperties{ - ConnectionString: map[string]any{ - "type": "SecureString", - "value": "DefaultEndpointsProtocol=https;AccountName=examplestorageaccount;AccountKey=", - }, - }, - }, - }, &armdatafactory.LinkedServicesClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.LinkedServiceResource = armdatafactory.LinkedServiceResource{ - // Name: to.Ptr("exampleLinkedService"), - // Type: to.Ptr("Microsoft.DataFactory/factories/linkedservices"), - // Etag: to.Ptr("0a0064d4-0000-0000-0000-5b245bd00000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/linkedservices/exampleLinkedService"), - // Properties: &armdatafactory.AzureStorageLinkedService{ - // Type: to.Ptr("AzureStorage"), - // Description: to.Ptr("Example description"), - // TypeProperties: &armdatafactory.AzureStorageLinkedServiceTypeProperties{ - // ConnectionString: map[string]any{ - // "type": "SecureString", - // "value": "**********", - // }, - // EncryptedCredential: to.Ptr("ew0KICAiVmVyc2lvbiI6ICIyMDE3LTExLTMwIiwNCiAgIlByb3RlY3Rpb25Nb2RlIjogIktleSIsDQogICJTZWNyZXRDb250ZW50VHlwZSI6ICJQbGFpbnRleHQiLA0KICAiQ3JlZGVudGlhbElkIjogIkRGLURPR0ZPT0QtWUFOWkhBTkctV1VfMGI2M2EyMmYtMGEzNC00NDg2LWIzMDktNzM0NTlkODUyY2Q1Ig0KfQ=="), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/LinkedServices_Get.json -func ExampleLinkedServicesClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewLinkedServicesClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleLinkedService", &armdatafactory.LinkedServicesClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.LinkedServiceResource = armdatafactory.LinkedServiceResource{ - // Name: to.Ptr("exampleLinkedService"), - // Type: to.Ptr("Microsoft.DataFactory/factories/linkedservices"), - // Etag: to.Ptr("1500474f-0000-0200-0000-5cbe090d0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/linkedservices/exampleLinkedService"), - // Properties: &armdatafactory.AzureStorageLinkedService{ - // Type: to.Ptr("AzureStorage"), - // Description: to.Ptr("Example description"), - // TypeProperties: &armdatafactory.AzureStorageLinkedServiceTypeProperties{ - // ConnectionString: map[string]any{ - // "type": "SecureString", - // "value": "**********", - // }, - // EncryptedCredential: to.Ptr("ew0KICAiVmVyc2lvbiI6ICIyMDE3LTExLTMwIiwNCiAgIlByb3RlY3Rpb25Nb2RlIjogIktleSIsDQogICJTZWNyZXRDb250ZW50VHlwZSI6ICJQbGFpbnRleHQiLA0KICAiQ3JlZGVudGlhbElkIjogIkhWVEVTVEVYQU1QTEVTXzg5MjU1NGY0LTViNGItNDFhOS1hYWYxLTg4ZWI5ZDBjOWIzMyINCn0="), - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/LinkedServices_Delete.json -func ExampleLinkedServicesClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewLinkedServicesClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleLinkedService", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/managedprivateendpoints_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/managedprivateendpoints_client_example_test.go deleted file mode 100644 index 6dc077352022..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/managedprivateendpoints_client_example_test.go +++ /dev/null @@ -1,164 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ManagedPrivateEndpoints_ListByFactory.json -func ExampleManagedPrivateEndpointsClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewManagedPrivateEndpointsClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", "exampleManagedVirtualNetworkName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.ManagedPrivateEndpointListResponse = armdatafactory.ManagedPrivateEndpointListResponse{ - // Value: []*armdatafactory.ManagedPrivateEndpointResource{ - // { - // Name: to.Ptr("exampleManagedPrivateEndpointName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/managedVirtualNetworks/managedPrivateEndpoints"), - // Etag: to.Ptr("000046c4-0000-0000-0000-5b2198bf0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/managedVirtualNetworks/exampleManagedVirtualNetworkName/managedPrivateEndpoints/exampleManagedPrivateEndpointName"), - // Properties: &armdatafactory.ManagedPrivateEndpoint{ - // ConnectionState: &armdatafactory.ConnectionStateProperties{ - // Description: to.Ptr(""), - // ActionsRequired: to.Ptr("None"), - // Status: to.Ptr("Pending"), - // }, - // Fqdns: []*string{ - // }, - // GroupID: to.Ptr("blob"), - // PrivateLinkResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.Storage/storageAccounts/exampleBlobStorage"), - // ProvisioningState: to.Ptr("Succeeded"), - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ManagedPrivateEndpoints_Create.json -func ExampleManagedPrivateEndpointsClient_CreateOrUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewManagedPrivateEndpointsClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleManagedVirtualNetworkName", "exampleManagedPrivateEndpointName", armdatafactory.ManagedPrivateEndpointResource{ - Properties: &armdatafactory.ManagedPrivateEndpoint{ - Fqdns: []*string{}, - GroupID: to.Ptr("blob"), - PrivateLinkResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.Storage/storageAccounts/exampleBlobStorage"), - }, - }, &armdatafactory.ManagedPrivateEndpointsClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ManagedPrivateEndpointResource = armdatafactory.ManagedPrivateEndpointResource{ - // Name: to.Ptr("exampleManagedPrivateEndpointName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/managedVirtualNetworks/managedPrivateEndpoints"), - // Etag: to.Ptr("000046c4-0000-0000-0000-5b2198bf0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/managedVirtualNetworks/exampleManagedVirtualNetworkName/managedPrivateEndpoints/exampleManagedPrivateEndpointName"), - // Properties: &armdatafactory.ManagedPrivateEndpoint{ - // ConnectionState: &armdatafactory.ConnectionStateProperties{ - // Description: to.Ptr(""), - // ActionsRequired: to.Ptr("None"), - // Status: to.Ptr("Pending"), - // }, - // Fqdns: []*string{ - // }, - // GroupID: to.Ptr("blob"), - // PrivateLinkResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.Storage/storageAccounts/exampleBlobStorage"), - // ProvisioningState: to.Ptr("Succeeded"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ManagedPrivateEndpoints_Get.json -func ExampleManagedPrivateEndpointsClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewManagedPrivateEndpointsClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleManagedVirtualNetworkName", "exampleManagedPrivateEndpointName", &armdatafactory.ManagedPrivateEndpointsClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ManagedPrivateEndpointResource = armdatafactory.ManagedPrivateEndpointResource{ - // Name: to.Ptr("exampleManagedPrivateEndpointName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/managedVirtualNetworks/managedPrivateEndpoints"), - // Etag: to.Ptr("000046c4-0000-0000-0000-5b2198bf0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/managedVirtualNetworks/exampleManagedVirtualNetworkName/managedPrivateEndpoints/exampleManagedPrivateEndpointName"), - // Properties: &armdatafactory.ManagedPrivateEndpoint{ - // ConnectionState: &armdatafactory.ConnectionStateProperties{ - // Description: to.Ptr(""), - // ActionsRequired: to.Ptr("None"), - // Status: to.Ptr("Pending"), - // }, - // Fqdns: []*string{ - // }, - // GroupID: to.Ptr("blob"), - // PrivateLinkResourceID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.Storage/storageAccounts/exampleBlobStorage"), - // ProvisioningState: to.Ptr("Succeeded"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ManagedPrivateEndpoints_Delete.json -func ExampleManagedPrivateEndpointsClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewManagedPrivateEndpointsClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleManagedVirtualNetworkName", "exampleManagedPrivateEndpointName", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/managedvirtualnetworks_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/managedvirtualnetworks_client_example_test.go deleted file mode 100644 index 9da1628a05bd..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/managedvirtualnetworks_client_example_test.go +++ /dev/null @@ -1,118 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ManagedVirtualNetworks_ListByFactory.json -func ExampleManagedVirtualNetworksClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewManagedVirtualNetworksClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.ManagedVirtualNetworkListResponse = armdatafactory.ManagedVirtualNetworkListResponse{ - // Value: []*armdatafactory.ManagedVirtualNetworkResource{ - // { - // Name: to.Ptr("exampleManagedVirtualNetworkName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/managedVirtualNetworks"), - // Etag: to.Ptr("0400f1a1-0000-0000-0000-5b2188640000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/managedVirtualNetworks/exampleManagedVirtualNetworkName"), - // Properties: &armdatafactory.ManagedVirtualNetwork{ - // Alias: to.Ptr("exampleFactoryName"), - // VNetID: to.Ptr("5a7bd944-87e6-454a-8d4d-9fba446514fd"), - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ManagedVirtualNetworks_Create.json -func ExampleManagedVirtualNetworksClient_CreateOrUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewManagedVirtualNetworksClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleManagedVirtualNetworkName", armdatafactory.ManagedVirtualNetworkResource{ - Properties: &armdatafactory.ManagedVirtualNetwork{}, - }, &armdatafactory.ManagedVirtualNetworksClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ManagedVirtualNetworkResource = armdatafactory.ManagedVirtualNetworkResource{ - // Name: to.Ptr("exampleManagedVirtualNetworkName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/managedVirtualNetworks"), - // Etag: to.Ptr("000046c4-0000-0000-0000-5b2198bf0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/managedVirtualNetworks/exampleManagedVirtualNetworkName"), - // Properties: &armdatafactory.ManagedVirtualNetwork{ - // Alias: to.Ptr("exampleFactoryName"), - // VNetID: to.Ptr("12345678-1234-1234-1234-12345678123"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ManagedVirtualNetworks_Get.json -func ExampleManagedVirtualNetworksClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewManagedVirtualNetworksClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleManagedVirtualNetworkName", &armdatafactory.ManagedVirtualNetworksClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.ManagedVirtualNetworkResource = armdatafactory.ManagedVirtualNetworkResource{ - // Name: to.Ptr("exampleManagedVirtualNetworkName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/managedVirtualNetworks"), - // Etag: to.Ptr("15003c4f-0000-0200-0000-5cbe090b0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/managedVirtualNetworks/exampleManagedVirtualNetworkName"), - // Properties: &armdatafactory.ManagedVirtualNetwork{ - // Alias: to.Ptr("exampleFactoryName"), - // VNetID: to.Ptr("5a7bd944-87e6-454a-8d4d-9fba446514fd"), - // }, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/models.go b/sdk/resourcemanager/datafactory/armdatafactory/models.go index 624c569ef625..468d0746c310 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/models.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/models.go @@ -2116,6 +2116,9 @@ type AzureBlobFSWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem } // GetStoreWriteSettings implements the StoreWriteSettingsClassification interface for type AzureBlobFSWriteSettings. @@ -2125,6 +2128,7 @@ func (a *AzureBlobFSWriteSettings) GetStoreWriteSettings() *StoreWriteSettings { CopyBehavior: a.CopyBehavior, DisableMetricsCollection: a.DisableMetricsCollection, MaxConcurrentConnections: a.MaxConcurrentConnections, + Metadata: a.Metadata, Type: a.Type, } } @@ -2323,6 +2327,9 @@ type AzureBlobStorageWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem } // GetStoreWriteSettings implements the StoreWriteSettingsClassification interface for type AzureBlobStorageWriteSettings. @@ -2332,6 +2339,7 @@ func (a *AzureBlobStorageWriteSettings) GetStoreWriteSettings() *StoreWriteSetti CopyBehavior: a.CopyBehavior, DisableMetricsCollection: a.DisableMetricsCollection, MaxConcurrentConnections: a.MaxConcurrentConnections, + Metadata: a.Metadata, Type: a.Type, } } @@ -3026,6 +3034,9 @@ type AzureDataLakeStoreWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem } // GetStoreWriteSettings implements the StoreWriteSettingsClassification interface for type AzureDataLakeStoreWriteSettings. @@ -3035,6 +3046,7 @@ func (a *AzureDataLakeStoreWriteSettings) GetStoreWriteSettings() *StoreWriteSet CopyBehavior: a.CopyBehavior, DisableMetricsCollection: a.DisableMetricsCollection, MaxConcurrentConnections: a.MaxConcurrentConnections, + Metadata: a.Metadata, Type: a.Type, } } @@ -3577,6 +3589,9 @@ type AzureFileStorageWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem } // GetStoreWriteSettings implements the StoreWriteSettingsClassification interface for type AzureFileStorageWriteSettings. @@ -3586,6 +3601,7 @@ func (a *AzureFileStorageWriteSettings) GetStoreWriteSettings() *StoreWriteSetti CopyBehavior: a.CopyBehavior, DisableMetricsCollection: a.DisableMetricsCollection, MaxConcurrentConnections: a.MaxConcurrentConnections, + Metadata: a.Metadata, Type: a.Type, } } @@ -3672,7 +3688,7 @@ type AzureFunctionActivityTypeProperties struct { // Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" // : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: - // string (or Expression with resultType string). + // dictionary (or Expression with resultType dictionary). Headers any } @@ -3714,7 +3730,8 @@ func (a *AzureFunctionLinkedService) GetLinkedService() *LinkedService { // AzureFunctionLinkedServiceTypeProperties - Azure Function linked service properties. type AzureFunctionLinkedServiceTypeProperties struct { - // REQUIRED; The endpoint of the Azure Function App. URL will be in the format https://.azurewebsites.net. + // REQUIRED; The endpoint of the Azure Function App. URL will be in the format https://.azurewebsites.net. Type: string (or + // Expression with resultType string). FunctionAppURL any // Type of authentication (Required to specify MSI) used to connect to AzureFunction. Type: string (or Expression with resultType @@ -3731,7 +3748,7 @@ type AzureFunctionLinkedServiceTypeProperties struct { // Function or Host key for Azure Function App. FunctionKey SecretBaseClassification - // Allowed token audiences for azure function. + // Allowed token audiences for azure function. Type: string (or Expression with resultType string). ResourceID any } @@ -3964,8 +3981,8 @@ type AzureMLExecutePipelineActivityTypeProperties struct { ContinueOnStepFailure any // Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments - // property of the published pipeline execution request. Type: object with key - // value pairs (or Expression with resultType object). + // property of the published pipeline execution request. Type: object (or + // Expression with resultType object). DataPathAssignments any // Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the @@ -5258,7 +5275,7 @@ type AzureSQLSource struct { MaxConcurrentConnections any // The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", - // "DynamicRange". + // "DynamicRange". Type: string (or Expression with resultType string). PartitionOption any // The settings that will be leveraged for Sql source partitioning. @@ -6546,10 +6563,10 @@ type CmdkeySetupTypeProperties struct { // REQUIRED; The password of data source access. Password SecretBaseClassification - // REQUIRED; The server name of data source access. + // REQUIRED; The server name of data source access. Type: string. TargetName any - // REQUIRED; The user name of data source access. + // REQUIRED; The user name of data source access. Type: string. UserName any } @@ -11778,6 +11795,9 @@ type FileServerWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem } // GetStoreWriteSettings implements the StoreWriteSettingsClassification interface for type FileServerWriteSettings. @@ -11787,6 +11807,7 @@ func (f *FileServerWriteSettings) GetStoreWriteSettings() *StoreWriteSettings { CopyBehavior: f.CopyBehavior, DisableMetricsCollection: f.DisableMetricsCollection, MaxConcurrentConnections: f.MaxConcurrentConnections, + Metadata: f.Metadata, Type: f.Type, } } @@ -13713,8 +13734,7 @@ type HDInsightOnDemandLinkedServiceTypeProperties struct { // REQUIRED; The resource group where the cluster belongs. Type: string (or Expression with resultType string). ClusterResourceGroup any - // REQUIRED; Number of worker/data nodes in the cluster. Suggestion value: 4. Type: string (or Expression with resultType - // string). + // REQUIRED; Number of worker/data nodes in the cluster. Suggestion value: 4. Type: int (or Expression with resultType int). ClusterSize any // REQUIRED; The customer’s subscription to host the cluster. Type: string (or Expression with resultType string). @@ -14226,8 +14246,8 @@ type HTTPLinkedServiceTypeProperties struct { // string). URL any - // The additional HTTP headers in the request to RESTful API used for authorization. Type: object (or Expression with resultType - // object). + // The additional HTTP headers in the request to RESTful API used for authorization. Type: key value pairs (value should be + // string type). AuthHeaders any // The authentication type to be used to connect to the HTTP server. @@ -16526,6 +16546,9 @@ type LakeHouseWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem } // GetStoreWriteSettings implements the StoreWriteSettingsClassification interface for type LakeHouseWriteSettings. @@ -16535,6 +16558,7 @@ func (l *LakeHouseWriteSettings) GetStoreWriteSettings() *StoreWriteSettings { CopyBehavior: l.CopyBehavior, DisableMetricsCollection: l.DisableMetricsCollection, MaxConcurrentConnections: l.MaxConcurrentConnections, + Metadata: l.Metadata, Type: l.Type, } } @@ -17498,12 +17522,28 @@ type MariaDBLinkedServiceTypeProperties struct { // An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. ConnectionString any + // Database name for connection. Type: string. + Database any + + // The version of the MariaDB driver. Type: string. V1 or empty for legacy driver, V2 for new driver. V1 can support connection + // string and property bag, V2 can only support connection string. + DriverVersion any + // The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. // Type: string. EncryptedCredential *string // The Azure key vault secret reference of password in connection string. - Pwd *AzureKeyVaultSecretReference + Password *AzureKeyVaultSecretReference + + // The port for the connection. Type: integer. + Port any + + // Server name for connection. Type: string. + Server any + + // Username for authentication. Type: string. + Username any } // MariaDBSource - A copy activity MariaDB server source. @@ -18654,15 +18694,37 @@ func (m *MySQLLinkedService) GetLinkedService() *LinkedService { // MySQLLinkedServiceTypeProperties - MySQL linked service properties. type MySQLLinkedServiceTypeProperties struct { - // REQUIRED; The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + // The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. ConnectionString any + // Database name for connection. Type: string. + Database any + + // The version of the MySQL driver. Type: string. V1 or empty for legacy driver, V2 for new driver. V1 can support connection + // string and property bag, V2 can only support connection string. + DriverVersion any + // The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. // Type: string. EncryptedCredential *string // The Azure key vault secret reference of password in connection string. Password *AzureKeyVaultSecretReference + + // The port for the connection. Type: integer. + Port any + + // SSL mode for connection. Type: integer. 0: disable, 1: prefer, 2: require, 3: verify-ca, 4: verify-full. + SSLMode any + + // Server name for connection. Type: string. + Server any + + // Use system trust store for connection. Type: integer. 0: enable, 1: disable. + UseSystemTrustStore any + + // Username for authentication. Type: string. + Username any } // MySQLSource - A copy activity source for MySQL databases. @@ -19021,8 +19083,8 @@ type ODataLinkedServiceTypeProperties struct { // Specify the credential type (key or cert) is used for service principal. AADServicePrincipalCredentialType *ODataAADServicePrincipalCredentialType - // The additional HTTP headers in the request to RESTful API used for authorization. Type: object (or Expression with resultType - // object). + // The additional HTTP headers in the request to RESTful API used for authorization. Type: key value pairs (value should be + // string type). AuthHeaders any // Type of authentication used to connect to the OData service. @@ -22552,7 +22614,7 @@ type RestSink struct { // REQUIRED; Copy sink type. Type *string - // The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). + // The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type). AdditionalHeaders any // OPTIONAL; Contains additional key/value pairs not defined in the schema. @@ -22562,7 +22624,7 @@ type RestSink struct { DisableMetricsCollection any // Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported - // option is Gzip. + // option is Gzip. Type: string (or Expression with resultType string). HTTPCompressionType any // The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. @@ -22611,8 +22673,7 @@ type RestSource struct { // REQUIRED; Copy source type. Type *string - // Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with - // resultType array of objects). + // Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type). AdditionalColumns any // The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). @@ -22825,7 +22886,7 @@ type SQLDWSource struct { MaxConcurrentConnections any // The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", - // "DynamicRange". + // "DynamicRange". Type: string (or Expression with resultType string). PartitionOption any // The settings that will be leveraged for Sql source partitioning. @@ -22938,7 +22999,7 @@ type SQLMISink struct { // Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). WriteBatchTimeout any - // White behavior when copying data into azure SQL MI. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) + // White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string) WriteBehavior any } @@ -22980,7 +23041,7 @@ type SQLMISource struct { MaxConcurrentConnections any // The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", - // "DynamicRange". + // "DynamicRange". Type: string (or Expression with resultType string). PartitionOption any // The settings that will be leveraged for Sql source partitioning. @@ -23159,7 +23220,7 @@ type SQLServerSink struct { // Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). WriteBatchTimeout any - // Write behavior when copying data into sql server. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) + // Write behavior when copying data into sql server. Type: string (or Expression with resultType string). WriteBehavior any } @@ -23201,7 +23262,7 @@ type SQLServerSource struct { MaxConcurrentConnections any // The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", - // "DynamicRange". + // "DynamicRange". Type: string (or Expression with resultType string). PartitionOption any // The settings that will be leveraged for Sql source partitioning. @@ -23444,7 +23505,7 @@ type SQLSink struct { // Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). WriteBatchTimeout any - // Write behavior when copying data into sql. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) + // Write behavior when copying data into sql. Type: string (or Expression with resultType string). WriteBehavior any } @@ -23486,7 +23547,7 @@ type SQLSource struct { MaxConcurrentConnections any // The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", - // "DynamicRange". + // "DynamicRange". Type: string (or Expression with resultType string). PartitionOption any // The settings that will be leveraged for Sql source partitioning. @@ -24171,6 +24232,226 @@ func (s *SalesforceServiceCloudSource) GetCopySource() *CopySource { } } +// SalesforceServiceCloudV2LinkedService - Linked service for Salesforce Service Cloud V2. +type SalesforceServiceCloudV2LinkedService struct { + // REQUIRED; Type of linked service. + Type *string + + // REQUIRED; Salesforce Service Cloud V2 linked service properties. + TypeProperties *SalesforceServiceCloudV2LinkedServiceTypeProperties + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the linked service. + Annotations []any + + // The integration runtime reference. + ConnectVia *IntegrationRuntimeReference + + // Linked service description. + Description *string + + // Parameters for linked service. + Parameters map[string]*ParameterSpecification +} + +// GetLinkedService implements the LinkedServiceClassification interface for type SalesforceServiceCloudV2LinkedService. +func (s *SalesforceServiceCloudV2LinkedService) GetLinkedService() *LinkedService { + return &LinkedService{ + AdditionalProperties: s.AdditionalProperties, + Annotations: s.Annotations, + ConnectVia: s.ConnectVia, + Description: s.Description, + Parameters: s.Parameters, + Type: s.Type, + } +} + +// SalesforceServiceCloudV2LinkedServiceTypeProperties - Salesforce Service Cloud V2 linked service properties. +type SalesforceServiceCloudV2LinkedServiceTypeProperties struct { + // The Salesforce API version used in ADF. The version must be larger than or equal to 47.0 which is required by Salesforce + // BULK API 2.0. Type: string (or Expression with resultType string). + APIVersion any + + // The authentication type to be used to connect to the Salesforce. Currently, we only support OAuth2ClientCredentials, it + // is also the default value + AuthenticationType any + + // The client Id for OAuth 2.0 Client Credentials Flow authentication of the Salesforce instance. Type: string (or Expression + // with resultType string). + ClientID any + + // The client secret for OAuth 2.0 Client Credentials Flow authentication of the Salesforce instance. + ClientSecret SecretBaseClassification + + // The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. + // Type: string. + EncryptedCredential *string + + // The URL of Salesforce Service Cloud instance. For example, 'https://[domain].my.salesforce.com'. Type: string (or Expression + // with resultType string). + EnvironmentURL any +} + +// SalesforceServiceCloudV2ObjectDataset - The Salesforce Service Cloud V2 object dataset. +type SalesforceServiceCloudV2ObjectDataset struct { + // REQUIRED; Linked service reference. + LinkedServiceName *LinkedServiceReference + + // REQUIRED; Type of dataset. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the Dataset. + Annotations []any + + // Dataset description. + Description *string + + // The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + Folder *DatasetFolder + + // Parameters for dataset. + Parameters map[string]*ParameterSpecification + + // Columns that define the physical type schema of the dataset. Type: array (or Expression with resultType array), itemType: + // DatasetSchemaDataElement. + Schema any + + // Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. + Structure any + + // Salesforce Service Cloud V2 object dataset properties. + TypeProperties *SalesforceServiceCloudV2ObjectDatasetTypeProperties +} + +// GetDataset implements the DatasetClassification interface for type SalesforceServiceCloudV2ObjectDataset. +func (s *SalesforceServiceCloudV2ObjectDataset) GetDataset() *Dataset { + return &Dataset{ + AdditionalProperties: s.AdditionalProperties, + Annotations: s.Annotations, + Description: s.Description, + Folder: s.Folder, + LinkedServiceName: s.LinkedServiceName, + Parameters: s.Parameters, + Schema: s.Schema, + Structure: s.Structure, + Type: s.Type, + } +} + +// SalesforceServiceCloudV2ObjectDatasetTypeProperties - Salesforce Service Cloud V2 object dataset properties. +type SalesforceServiceCloudV2ObjectDatasetTypeProperties struct { + // The Salesforce Service Cloud V2 object API name. Type: string (or Expression with resultType string). + ObjectAPIName any + + // The Salesforce Service Cloud V2 reportId. Type: string (or Expression with resultType string). + ReportID any +} + +// SalesforceServiceCloudV2Sink - A copy activity Salesforce Service Cloud V2 sink. +type SalesforceServiceCloudV2Sink struct { + // REQUIRED; Copy sink type. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with + // resultType string). + ExternalIDFieldName any + + // The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. + // Default value is false. If set it to true, it means ADF will leave the data in + // the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert + // operation, versus ADF will update the data in the destination object to NULL when + // doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType + // boolean). + IgnoreNullValues any + + // The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // Sink retry count. Type: integer (or Expression with resultType integer). + SinkRetryCount any + + // Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SinkRetryWait any + + // Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. + WriteBatchSize any + + // Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + WriteBatchTimeout any + + // The write behavior for the operation. Default is Insert. + WriteBehavior *SalesforceV2SinkWriteBehavior +} + +// GetCopySink implements the CopySinkClassification interface for type SalesforceServiceCloudV2Sink. +func (s *SalesforceServiceCloudV2Sink) GetCopySink() *CopySink { + return &CopySink{ + AdditionalProperties: s.AdditionalProperties, + DisableMetricsCollection: s.DisableMetricsCollection, + MaxConcurrentConnections: s.MaxConcurrentConnections, + SinkRetryCount: s.SinkRetryCount, + SinkRetryWait: s.SinkRetryWait, + Type: s.Type, + WriteBatchSize: s.WriteBatchSize, + WriteBatchTimeout: s.WriteBatchTimeout, + } +} + +// SalesforceServiceCloudV2Source - A copy activity Salesforce Service Cloud V2 source. +type SalesforceServiceCloudV2Source struct { + // REQUIRED; Copy source type. + Type *string + + // Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with + // resultType array of objects). + AdditionalColumns any + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with + // resultType boolean). + IncludeDeletedObjects any + + // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // Database query. Type: string (or Expression with resultType string). + SOQLQuery any + + // Source retry count. Type: integer (or Expression with resultType integer). + SourceRetryCount any + + // Source retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SourceRetryWait any +} + +// GetCopySource implements the CopySourceClassification interface for type SalesforceServiceCloudV2Source. +func (s *SalesforceServiceCloudV2Source) GetCopySource() *CopySource { + return &CopySource{ + AdditionalProperties: s.AdditionalProperties, + DisableMetricsCollection: s.DisableMetricsCollection, + MaxConcurrentConnections: s.MaxConcurrentConnections, + SourceRetryCount: s.SourceRetryCount, + SourceRetryWait: s.SourceRetryWait, + Type: s.Type, + } +} + // SalesforceSink - A copy activity Salesforce sink. type SalesforceSink struct { // REQUIRED; Copy sink type. @@ -24288,13 +24569,13 @@ func (s *SalesforceSource) GetTabularSource() *TabularSource { } } -// SapBWLinkedService - SAP Business Warehouse Linked Service. -type SapBWLinkedService struct { +// SalesforceV2LinkedService - Linked service for Salesforce V2. +type SalesforceV2LinkedService struct { // REQUIRED; Type of linked service. Type *string - // REQUIRED; Properties specific to this linked service type. - TypeProperties *SapBWLinkedServiceTypeProperties + // REQUIRED; Salesforce V2 linked service properties. + TypeProperties *SalesforceV2LinkedServiceTypeProperties // OPTIONAL; Contains additional key/value pairs not defined in the schema. AdditionalProperties map[string]any @@ -24312,8 +24593,8 @@ type SapBWLinkedService struct { Parameters map[string]*ParameterSpecification } -// GetLinkedService implements the LinkedServiceClassification interface for type SapBWLinkedService. -func (s *SapBWLinkedService) GetLinkedService() *LinkedService { +// GetLinkedService implements the LinkedServiceClassification interface for type SalesforceV2LinkedService. +func (s *SalesforceV2LinkedService) GetLinkedService() *LinkedService { return &LinkedService{ AdditionalProperties: s.AdditionalProperties, Annotations: s.Annotations, @@ -24324,14 +24605,251 @@ func (s *SapBWLinkedService) GetLinkedService() *LinkedService { } } -// SapBWLinkedServiceTypeProperties - Properties specific to this linked service type. -type SapBWLinkedServiceTypeProperties struct { - // REQUIRED; Client ID of the client on the BW system. (Usually a three-digit decimal number represented as a string) Type: - // string (or Expression with resultType string). - ClientID any +// SalesforceV2LinkedServiceTypeProperties - Salesforce V2 linked service properties. +type SalesforceV2LinkedServiceTypeProperties struct { + // The Salesforce API version used in ADF. The version must be larger than or equal to 47.0 which is required by Salesforce + // BULK API 2.0. Type: string (or Expression with resultType string). + APIVersion any - // REQUIRED; Host name of the SAP BW instance. Type: string (or Expression with resultType string). - Server any + // The authentication type to be used to connect to the Salesforce. Currently, we only support OAuth2ClientCredentials, it + // is also the default value + AuthenticationType any + + // The client Id for OAuth 2.0 Client Credentials Flow authentication of the Salesforce instance. Type: string (or Expression + // with resultType string). + ClientID any + + // The client secret for OAuth 2.0 Client Credentials Flow authentication of the Salesforce instance. + ClientSecret SecretBaseClassification + + // The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. + // Type: string. + EncryptedCredential *string + + // The URL of Salesforce instance. For example, 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType + // string). + EnvironmentURL any +} + +// SalesforceV2ObjectDataset - The Salesforce V2 object dataset. +type SalesforceV2ObjectDataset struct { + // REQUIRED; Linked service reference. + LinkedServiceName *LinkedServiceReference + + // REQUIRED; Type of dataset. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the Dataset. + Annotations []any + + // Dataset description. + Description *string + + // The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + Folder *DatasetFolder + + // Parameters for dataset. + Parameters map[string]*ParameterSpecification + + // Columns that define the physical type schema of the dataset. Type: array (or Expression with resultType array), itemType: + // DatasetSchemaDataElement. + Schema any + + // Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. + Structure any + + // Salesforce V2 object dataset properties. + TypeProperties *SalesforceV2ObjectDatasetTypeProperties +} + +// GetDataset implements the DatasetClassification interface for type SalesforceV2ObjectDataset. +func (s *SalesforceV2ObjectDataset) GetDataset() *Dataset { + return &Dataset{ + AdditionalProperties: s.AdditionalProperties, + Annotations: s.Annotations, + Description: s.Description, + Folder: s.Folder, + LinkedServiceName: s.LinkedServiceName, + Parameters: s.Parameters, + Schema: s.Schema, + Structure: s.Structure, + Type: s.Type, + } +} + +// SalesforceV2ObjectDatasetTypeProperties - Salesforce V2 object dataset properties. +type SalesforceV2ObjectDatasetTypeProperties struct { + // The Salesforce V2 object API name. Type: string (or Expression with resultType string). + ObjectAPIName any + + // The Salesforce V2 report Id. Type: string (or Expression with resultType string). + ReportID any +} + +// SalesforceV2Sink - A copy activity Salesforce V2 sink. +type SalesforceV2Sink struct { + // REQUIRED; Copy sink type. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with + // resultType string). + ExternalIDFieldName any + + // The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. + // Default value is false. If set it to true, it means ADF will leave the data in + // the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert + // operation, versus ADF will update the data in the destination object to NULL when + // doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType + // boolean). + IgnoreNullValues any + + // The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // Sink retry count. Type: integer (or Expression with resultType integer). + SinkRetryCount any + + // Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SinkRetryWait any + + // Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. + WriteBatchSize any + + // Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + WriteBatchTimeout any + + // The write behavior for the operation. Default is Insert. + WriteBehavior *SalesforceV2SinkWriteBehavior +} + +// GetCopySink implements the CopySinkClassification interface for type SalesforceV2Sink. +func (s *SalesforceV2Sink) GetCopySink() *CopySink { + return &CopySink{ + AdditionalProperties: s.AdditionalProperties, + DisableMetricsCollection: s.DisableMetricsCollection, + MaxConcurrentConnections: s.MaxConcurrentConnections, + SinkRetryCount: s.SinkRetryCount, + SinkRetryWait: s.SinkRetryWait, + Type: s.Type, + WriteBatchSize: s.WriteBatchSize, + WriteBatchTimeout: s.WriteBatchTimeout, + } +} + +// SalesforceV2Source - A copy activity Salesforce V2 source. +type SalesforceV2Source struct { + // REQUIRED; Copy source type. + Type *string + + // Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with + // resultType array of objects). + AdditionalColumns any + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with + // resultType boolean). + IncludeDeletedObjects any + + // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + QueryTimeout any + + // Database query. Type: string (or Expression with resultType string). + SOQLQuery any + + // Source retry count. Type: integer (or Expression with resultType integer). + SourceRetryCount any + + // Source retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SourceRetryWait any +} + +// GetCopySource implements the CopySourceClassification interface for type SalesforceV2Source. +func (s *SalesforceV2Source) GetCopySource() *CopySource { + return &CopySource{ + AdditionalProperties: s.AdditionalProperties, + DisableMetricsCollection: s.DisableMetricsCollection, + MaxConcurrentConnections: s.MaxConcurrentConnections, + SourceRetryCount: s.SourceRetryCount, + SourceRetryWait: s.SourceRetryWait, + Type: s.Type, + } +} + +// GetTabularSource implements the TabularSourceClassification interface for type SalesforceV2Source. +func (s *SalesforceV2Source) GetTabularSource() *TabularSource { + return &TabularSource{ + AdditionalColumns: s.AdditionalColumns, + AdditionalProperties: s.AdditionalProperties, + DisableMetricsCollection: s.DisableMetricsCollection, + MaxConcurrentConnections: s.MaxConcurrentConnections, + QueryTimeout: s.QueryTimeout, + SourceRetryCount: s.SourceRetryCount, + SourceRetryWait: s.SourceRetryWait, + Type: s.Type, + } +} + +// SapBWLinkedService - SAP Business Warehouse Linked Service. +type SapBWLinkedService struct { + // REQUIRED; Type of linked service. + Type *string + + // REQUIRED; Properties specific to this linked service type. + TypeProperties *SapBWLinkedServiceTypeProperties + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the linked service. + Annotations []any + + // The integration runtime reference. + ConnectVia *IntegrationRuntimeReference + + // Linked service description. + Description *string + + // Parameters for linked service. + Parameters map[string]*ParameterSpecification +} + +// GetLinkedService implements the LinkedServiceClassification interface for type SapBWLinkedService. +func (s *SapBWLinkedService) GetLinkedService() *LinkedService { + return &LinkedService{ + AdditionalProperties: s.AdditionalProperties, + Annotations: s.Annotations, + ConnectVia: s.ConnectVia, + Description: s.Description, + Parameters: s.Parameters, + Type: s.Type, + } +} + +// SapBWLinkedServiceTypeProperties - Properties specific to this linked service type. +type SapBWLinkedServiceTypeProperties struct { + // REQUIRED; Client ID of the client on the BW system. (Usually a three-digit decimal number represented as a string) Type: + // string (or Expression with resultType string). + ClientID any + + // REQUIRED; Host name of the SAP BW instance. Type: string (or Expression with resultType string). + Server any // REQUIRED; System number of the BW system. (Usually a two-digit decimal number represented as a string.) Type: string (or // Expression with resultType string). @@ -26642,6 +27160,9 @@ type SftpWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem + // Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression // with resultType string). OperationTimeout any @@ -26658,6 +27179,7 @@ func (s *SftpWriteSettings) GetStoreWriteSettings() *StoreWriteSettings { CopyBehavior: s.CopyBehavior, DisableMetricsCollection: s.DisableMetricsCollection, MaxConcurrentConnections: s.MaxConcurrentConnections, + Metadata: s.Metadata, Type: s.Type, } } @@ -27201,6 +27723,49 @@ type SnowflakeLinkedServiceTypeProperties struct { Password *AzureKeyVaultSecretReference } +// SnowflakeLinkedV2ServiceTypeProperties - Snowflake linked service properties. +type SnowflakeLinkedV2ServiceTypeProperties struct { + // REQUIRED; The account identifier of your Snowflake account, e.g. xy12345.east-us-2.azure + AccountIdentifier any + + // REQUIRED; The name of the Snowflake database. + Database any + + // REQUIRED; The name of the Snowflake warehouse. + Warehouse any + + // The type used for authentication. Type: string. + AuthenticationType *SnowflakeAuthenticationType + + // The client ID of the application registered in Azure Active Directory for AADServicePrincipal authentication. + ClientID any + + // The Azure key vault secret reference of client secret for AADServicePrincipal authentication. + ClientSecret SecretBaseClassification + + // The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. + // Type: string. + EncryptedCredential *string + + // The Azure key vault secret reference of password in connection string. + Password SecretBaseClassification + + // The Azure key vault secret reference of privateKey for KeyPair auth. + PrivateKey SecretBaseClassification + + // The Azure key vault secret reference of private key password for KeyPair auth with encrypted private key. + PrivateKeyPassphrase SecretBaseClassification + + // The scope of the application registered in Azure Active Directory for AADServicePrincipal authentication. + Scope any + + // The tenant ID of the application registered in Azure Active Directory for AADServicePrincipal authentication. + TenantID any + + // The name of the Snowflake user. + User any +} + // SnowflakeSink - A copy activity snowflake sink. type SnowflakeSink struct { // REQUIRED; Copy sink type. @@ -27287,21 +27852,192 @@ func (s *SnowflakeSource) GetCopySource() *CopySource { } } -// SparkConfigurationParametrizationReference - Spark configuration reference. -type SparkConfigurationParametrizationReference struct { - // REQUIRED; Reference spark configuration name. Type: string (or Expression with resultType string). - ReferenceName any +// SnowflakeV2Dataset - The snowflake dataset. +type SnowflakeV2Dataset struct { + // REQUIRED; Linked service reference. + LinkedServiceName *LinkedServiceReference - // REQUIRED; Spark configuration reference type. - Type *SparkConfigurationReferenceType -} + // REQUIRED; Type of dataset. + Type *string -// SparkDatasetTypeProperties - Spark Properties -type SparkDatasetTypeProperties struct { - // The schema name of the Spark. Type: string (or Expression with resultType string). - Schema any + // REQUIRED; Snowflake dataset properties. + TypeProperties *SnowflakeDatasetTypeProperties - // The table name of the Spark. Type: string (or Expression with resultType string). + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the Dataset. + Annotations []any + + // Dataset description. + Description *string + + // The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + Folder *DatasetFolder + + // Parameters for dataset. + Parameters map[string]*ParameterSpecification + + // Columns that define the physical type schema of the dataset. Type: array (or Expression with resultType array), itemType: + // DatasetSchemaDataElement. + Schema any + + // Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. + Structure any +} + +// GetDataset implements the DatasetClassification interface for type SnowflakeV2Dataset. +func (s *SnowflakeV2Dataset) GetDataset() *Dataset { + return &Dataset{ + AdditionalProperties: s.AdditionalProperties, + Annotations: s.Annotations, + Description: s.Description, + Folder: s.Folder, + LinkedServiceName: s.LinkedServiceName, + Parameters: s.Parameters, + Schema: s.Schema, + Structure: s.Structure, + Type: s.Type, + } +} + +// SnowflakeV2LinkedService - Snowflake linked service. +type SnowflakeV2LinkedService struct { + // REQUIRED; Type of linked service. + Type *string + + // REQUIRED; Snowflake linked service properties. + TypeProperties *SnowflakeLinkedV2ServiceTypeProperties + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the linked service. + Annotations []any + + // The integration runtime reference. + ConnectVia *IntegrationRuntimeReference + + // Linked service description. + Description *string + + // Parameters for linked service. + Parameters map[string]*ParameterSpecification +} + +// GetLinkedService implements the LinkedServiceClassification interface for type SnowflakeV2LinkedService. +func (s *SnowflakeV2LinkedService) GetLinkedService() *LinkedService { + return &LinkedService{ + AdditionalProperties: s.AdditionalProperties, + Annotations: s.Annotations, + ConnectVia: s.ConnectVia, + Description: s.Description, + Parameters: s.Parameters, + Type: s.Type, + } +} + +// SnowflakeV2Sink - A copy activity snowflake sink. +type SnowflakeV2Sink struct { + // REQUIRED; Copy sink type. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // Snowflake import settings. + ImportSettings *SnowflakeImportCopyCommand + + // The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // SQL pre-copy script. Type: string (or Expression with resultType string). + PreCopyScript any + + // Sink retry count. Type: integer (or Expression with resultType integer). + SinkRetryCount any + + // Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SinkRetryWait any + + // Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. + WriteBatchSize any + + // Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + WriteBatchTimeout any +} + +// GetCopySink implements the CopySinkClassification interface for type SnowflakeV2Sink. +func (s *SnowflakeV2Sink) GetCopySink() *CopySink { + return &CopySink{ + AdditionalProperties: s.AdditionalProperties, + DisableMetricsCollection: s.DisableMetricsCollection, + MaxConcurrentConnections: s.MaxConcurrentConnections, + SinkRetryCount: s.SinkRetryCount, + SinkRetryWait: s.SinkRetryWait, + Type: s.Type, + WriteBatchSize: s.WriteBatchSize, + WriteBatchTimeout: s.WriteBatchTimeout, + } +} + +// SnowflakeV2Source - A copy activity snowflake source. +type SnowflakeV2Source struct { + // REQUIRED; Snowflake export settings. + ExportSettings *SnowflakeExportCopyCommand + + // REQUIRED; Copy source type. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // Snowflake Sql query. Type: string (or Expression with resultType string). + Query any + + // Source retry count. Type: integer (or Expression with resultType integer). + SourceRetryCount any + + // Source retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SourceRetryWait any +} + +// GetCopySource implements the CopySourceClassification interface for type SnowflakeV2Source. +func (s *SnowflakeV2Source) GetCopySource() *CopySource { + return &CopySource{ + AdditionalProperties: s.AdditionalProperties, + DisableMetricsCollection: s.DisableMetricsCollection, + MaxConcurrentConnections: s.MaxConcurrentConnections, + SourceRetryCount: s.SourceRetryCount, + SourceRetryWait: s.SourceRetryWait, + Type: s.Type, + } +} + +// SparkConfigurationParametrizationReference - Spark configuration reference. +type SparkConfigurationParametrizationReference struct { + // REQUIRED; Reference spark configuration name. Type: string (or Expression with resultType string). + ReferenceName any + + // REQUIRED; Spark configuration reference type. + Type *SparkConfigurationReferenceType +} + +// SparkDatasetTypeProperties - Spark Properties +type SparkDatasetTypeProperties struct { + // The schema name of the Spark. Type: string (or Expression with resultType string). + Schema any + + // The table name of the Spark. Type: string (or Expression with resultType string). Table any // This property will be retired. Please consider using schema + table properties instead. @@ -27974,6 +28710,9 @@ type StoreWriteSettings struct { // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). MaxConcurrentConnections any + + // Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). + Metadata []*MetadataItem } // GetStoreWriteSettings implements the StoreWriteSettingsClassification interface for type StoreWriteSettings. @@ -29647,6 +30386,272 @@ type WaitActivityTypeProperties struct { WaitTimeInSeconds any } +// WarehouseLinkedService - Microsoft Fabric Warehouse linked service. +type WarehouseLinkedService struct { + // REQUIRED; Type of linked service. + Type *string + + // REQUIRED; Microsoft Fabric Warehouse linked service properties. + TypeProperties *WarehouseLinkedServiceTypeProperties + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the linked service. + Annotations []any + + // The integration runtime reference. + ConnectVia *IntegrationRuntimeReference + + // Linked service description. + Description *string + + // Parameters for linked service. + Parameters map[string]*ParameterSpecification +} + +// GetLinkedService implements the LinkedServiceClassification interface for type WarehouseLinkedService. +func (w *WarehouseLinkedService) GetLinkedService() *LinkedService { + return &LinkedService{ + AdditionalProperties: w.AdditionalProperties, + Annotations: w.Annotations, + ConnectVia: w.ConnectVia, + Description: w.Description, + Parameters: w.Parameters, + Type: w.Type, + } +} + +// WarehouseLinkedServiceTypeProperties - Microsoft Fabric Warehouse linked service properties. +type WarehouseLinkedServiceTypeProperties struct { + // REQUIRED; The ID of Microsoft Fabric Warehouse artifact. Type: string (or Expression with resultType string). + ArtifactID any + + // REQUIRED; The endpoint of Microsoft Fabric Warehouse server. Type: string (or Expression with resultType string). + Endpoint any + + // The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. + // Type: string. + EncryptedCredential *string + + // The credential of the service principal object in Azure Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', + // servicePrincipalCredential can be SecureString or + // AzureKeyVaultSecretReference. If servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can + // only be AzureKeyVaultSecretReference. + ServicePrincipalCredential SecretBaseClassification + + // The service principal credential type to use in Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, + // 'ServicePrincipalCert' for certificate. Type: string (or Expression with + // resultType string). + ServicePrincipalCredentialType any + + // The ID of the application used to authenticate against Microsoft Fabric Warehouse. Type: string (or Expression with resultType + // string). + ServicePrincipalID any + + // The Key of the application used to authenticate against Microsoft Fabric Warehouse. + ServicePrincipalKey SecretBaseClassification + + // The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). + Tenant any + + // The ID of Microsoft Fabric workspace. Type: string (or Expression with resultType string). + WorkspaceID any +} + +// WarehouseSink - A copy activity Microsoft Fabric Warehouse sink. +type WarehouseSink struct { + // REQUIRED; Copy sink type. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). + AllowCopyCommand any + + // Specifies Copy Command related settings when allowCopyCommand is true. + CopyCommandSettings *DWCopyCommandSettings + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // SQL pre-copy script. Type: string (or Expression with resultType string). + PreCopyScript any + + // Sink retry count. Type: integer (or Expression with resultType integer). + SinkRetryCount any + + // Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SinkRetryWait any + + // The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression + // with resultType string). + TableOption any + + // Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. + WriteBatchSize any + + // Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + WriteBatchTimeout any + + // Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with + // resultType DWWriteBehaviorEnum) + WriteBehavior any +} + +// GetCopySink implements the CopySinkClassification interface for type WarehouseSink. +func (w *WarehouseSink) GetCopySink() *CopySink { + return &CopySink{ + AdditionalProperties: w.AdditionalProperties, + DisableMetricsCollection: w.DisableMetricsCollection, + MaxConcurrentConnections: w.MaxConcurrentConnections, + SinkRetryCount: w.SinkRetryCount, + SinkRetryWait: w.SinkRetryWait, + Type: w.Type, + WriteBatchSize: w.WriteBatchSize, + WriteBatchTimeout: w.WriteBatchTimeout, + } +} + +// WarehouseSource - A copy activity Microsoft Fabric Warehouse source. +type WarehouseSource struct { + // REQUIRED; Copy source type. + Type *string + + // Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with + // resultType array of objects). + AdditionalColumns any + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). + DisableMetricsCollection any + + // Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. + // The default value is + // ReadCommitted. Type: string (or Expression with resultType string). + IsolationLevel any + + // The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). + MaxConcurrentConnections any + + // The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", + // "DynamicRange". + PartitionOption any + + // The settings that will be leveraged for Sql source partitioning. + PartitionSettings *SQLPartitionSettings + + // Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + QueryTimeout any + + // Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string). + SQLReaderQuery any + + // Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. + // Type: string (or Expression with resultType string). + SQLReaderStoredProcedureName any + + // Source retry count. Type: integer (or Expression with resultType integer). + SourceRetryCount any + + // Source retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + SourceRetryWait any + + // Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object + // (or Expression with resultType object), itemType: StoredProcedureParameter. + StoredProcedureParameters any +} + +// GetCopySource implements the CopySourceClassification interface for type WarehouseSource. +func (w *WarehouseSource) GetCopySource() *CopySource { + return &CopySource{ + AdditionalProperties: w.AdditionalProperties, + DisableMetricsCollection: w.DisableMetricsCollection, + MaxConcurrentConnections: w.MaxConcurrentConnections, + SourceRetryCount: w.SourceRetryCount, + SourceRetryWait: w.SourceRetryWait, + Type: w.Type, + } +} + +// GetTabularSource implements the TabularSourceClassification interface for type WarehouseSource. +func (w *WarehouseSource) GetTabularSource() *TabularSource { + return &TabularSource{ + AdditionalColumns: w.AdditionalColumns, + AdditionalProperties: w.AdditionalProperties, + DisableMetricsCollection: w.DisableMetricsCollection, + MaxConcurrentConnections: w.MaxConcurrentConnections, + QueryTimeout: w.QueryTimeout, + SourceRetryCount: w.SourceRetryCount, + SourceRetryWait: w.SourceRetryWait, + Type: w.Type, + } +} + +// WarehouseTableDataset - Microsoft Fabric Warehouse dataset. +type WarehouseTableDataset struct { + // REQUIRED; Linked service reference. + LinkedServiceName *LinkedServiceReference + + // REQUIRED; Type of dataset. + Type *string + + // OPTIONAL; Contains additional key/value pairs not defined in the schema. + AdditionalProperties map[string]any + + // List of tags that can be used for describing the Dataset. + Annotations []any + + // Dataset description. + Description *string + + // The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + Folder *DatasetFolder + + // Parameters for dataset. + Parameters map[string]*ParameterSpecification + + // Columns that define the physical type schema of the dataset. Type: array (or Expression with resultType array), itemType: + // DatasetSchemaDataElement. + Schema any + + // Columns that define the structure of the dataset. Type: array (or Expression with resultType array), itemType: DatasetDataElement. + Structure any + + // Microsoft Fabric Warehouse dataset properties. + TypeProperties *WarehouseTableDatasetTypeProperties +} + +// GetDataset implements the DatasetClassification interface for type WarehouseTableDataset. +func (w *WarehouseTableDataset) GetDataset() *Dataset { + return &Dataset{ + AdditionalProperties: w.AdditionalProperties, + Annotations: w.Annotations, + Description: w.Description, + Folder: w.Folder, + LinkedServiceName: w.LinkedServiceName, + Parameters: w.Parameters, + Schema: w.Schema, + Structure: w.Structure, + Type: w.Type, + } +} + +// WarehouseTableDatasetTypeProperties - Microsoft Fabric Warehouse dataset properties. +type WarehouseTableDatasetTypeProperties struct { + // The schema name of the Microsoft Fabric Warehouse. Type: string (or Expression with resultType string). + Schema any + + // The table name of the Microsoft Fabric Warehouse. Type: string (or Expression with resultType string). + Table any +} + // WebActivity - Web activity. type WebActivity struct { // REQUIRED; Activity name. @@ -29765,13 +30770,23 @@ type WebActivityTypeProperties struct { // When set to true, Certificate validation will be disabled. DisableCertValidation *bool + // Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, + // not the activity timeout. The default value is 00:01:00 (1 minute). The + // range is from 1 to 10 minutes + HTTPRequestTimeout any + // Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" // : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: - // string (or Expression with resultType string). + // dictionary (or Expression with resultType dictionary). Headers any // List of linked services passed to web endpoint. LinkedServices []*LinkedServiceReference + + // Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops + // invoking HTTP GET on http location given in response header. If set false then + // continues to invoke HTTP GET call on location given in http response headers. + TurnOffAsync *bool } // WebAnonymousAuthentication - A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. @@ -29921,7 +30936,7 @@ type WebHookActivityTypeProperties struct { // Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" // : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: - // string (or Expression with resultType string). + // dictionary (or Expression with resultType dictionary). Headers any // When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can diff --git a/sdk/resourcemanager/datafactory/armdatafactory/models_serde.go b/sdk/resourcemanager/datafactory/armdatafactory/models_serde.go index d9f3171250f2..eccdb123ba10 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/models_serde.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/models_serde.go @@ -3699,6 +3699,7 @@ func (a AzureBlobFSWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "copyBehavior", a.CopyBehavior) populateAny(objectMap, "disableMetricsCollection", a.DisableMetricsCollection) populateAny(objectMap, "maxConcurrentConnections", a.MaxConcurrentConnections) + populate(objectMap, "metadata", a.Metadata) objectMap["type"] = "AzureBlobFSWriteSettings" if a.AdditionalProperties != nil { for key, val := range a.AdditionalProperties { @@ -3729,6 +3730,9 @@ func (a *AzureBlobFSWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &a.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &a.Metadata) + delete(rawMsg, key) case "type": err = unpopulate(val, "Type", &a.Type) delete(rawMsg, key) @@ -4042,6 +4046,7 @@ func (a AzureBlobStorageWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "copyBehavior", a.CopyBehavior) populateAny(objectMap, "disableMetricsCollection", a.DisableMetricsCollection) populateAny(objectMap, "maxConcurrentConnections", a.MaxConcurrentConnections) + populate(objectMap, "metadata", a.Metadata) objectMap["type"] = "AzureBlobStorageWriteSettings" if a.AdditionalProperties != nil { for key, val := range a.AdditionalProperties { @@ -4072,6 +4077,9 @@ func (a *AzureBlobStorageWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &a.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &a.Metadata) + delete(rawMsg, key) case "type": err = unpopulate(val, "Type", &a.Type) delete(rawMsg, key) @@ -5206,6 +5214,7 @@ func (a AzureDataLakeStoreWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "disableMetricsCollection", a.DisableMetricsCollection) populateAny(objectMap, "expiryDateTime", a.ExpiryDateTime) populateAny(objectMap, "maxConcurrentConnections", a.MaxConcurrentConnections) + populate(objectMap, "metadata", a.Metadata) objectMap["type"] = "AzureDataLakeStoreWriteSettings" if a.AdditionalProperties != nil { for key, val := range a.AdditionalProperties { @@ -5236,6 +5245,9 @@ func (a *AzureDataLakeStoreWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &a.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &a.Metadata) + delete(rawMsg, key) case "type": err = unpopulate(val, "Type", &a.Type) delete(rawMsg, key) @@ -6143,6 +6155,7 @@ func (a AzureFileStorageWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "copyBehavior", a.CopyBehavior) populateAny(objectMap, "disableMetricsCollection", a.DisableMetricsCollection) populateAny(objectMap, "maxConcurrentConnections", a.MaxConcurrentConnections) + populate(objectMap, "metadata", a.Metadata) objectMap["type"] = "AzureFileStorageWriteSettings" if a.AdditionalProperties != nil { for key, val := range a.AdditionalProperties { @@ -6170,6 +6183,9 @@ func (a *AzureFileStorageWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &a.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &a.Metadata) + delete(rawMsg, key) case "type": err = unpopulate(val, "Type", &a.Type) delete(rawMsg, key) @@ -20449,6 +20465,7 @@ func (f FileServerWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "copyBehavior", f.CopyBehavior) populateAny(objectMap, "disableMetricsCollection", f.DisableMetricsCollection) populateAny(objectMap, "maxConcurrentConnections", f.MaxConcurrentConnections) + populate(objectMap, "metadata", f.Metadata) objectMap["type"] = "FileServerWriteSettings" if f.AdditionalProperties != nil { for key, val := range f.AdditionalProperties { @@ -20476,6 +20493,9 @@ func (f *FileServerWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &f.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &f.Metadata) + delete(rawMsg, key) case "type": err = unpopulate(val, "Type", &f.Type) delete(rawMsg, key) @@ -28657,6 +28677,7 @@ func (l LakeHouseWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "copyBehavior", l.CopyBehavior) populateAny(objectMap, "disableMetricsCollection", l.DisableMetricsCollection) populateAny(objectMap, "maxConcurrentConnections", l.MaxConcurrentConnections) + populate(objectMap, "metadata", l.Metadata) objectMap["type"] = "LakeHouseWriteSettings" if l.AdditionalProperties != nil { for key, val := range l.AdditionalProperties { @@ -28684,6 +28705,9 @@ func (l *LakeHouseWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &l.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &l.Metadata) + delete(rawMsg, key) case "type": err = unpopulate(val, "Type", &l.Type) delete(rawMsg, key) @@ -30937,8 +30961,13 @@ func (m *MariaDBLinkedService) UnmarshalJSON(data []byte) error { func (m MariaDBLinkedServiceTypeProperties) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) populateAny(objectMap, "connectionString", m.ConnectionString) + populateAny(objectMap, "database", m.Database) + populateAny(objectMap, "driverVersion", m.DriverVersion) populate(objectMap, "encryptedCredential", m.EncryptedCredential) - populate(objectMap, "pwd", m.Pwd) + populate(objectMap, "password", m.Password) + populateAny(objectMap, "port", m.Port) + populateAny(objectMap, "server", m.Server) + populateAny(objectMap, "username", m.Username) return json.Marshal(objectMap) } @@ -30954,11 +30983,26 @@ func (m *MariaDBLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { case "connectionString": err = unpopulate(val, "ConnectionString", &m.ConnectionString) delete(rawMsg, key) + case "database": + err = unpopulate(val, "Database", &m.Database) + delete(rawMsg, key) + case "driverVersion": + err = unpopulate(val, "DriverVersion", &m.DriverVersion) + delete(rawMsg, key) case "encryptedCredential": err = unpopulate(val, "EncryptedCredential", &m.EncryptedCredential) delete(rawMsg, key) - case "pwd": - err = unpopulate(val, "Pwd", &m.Pwd) + case "password": + err = unpopulate(val, "Password", &m.Password) + delete(rawMsg, key) + case "port": + err = unpopulate(val, "Port", &m.Port) + delete(rawMsg, key) + case "server": + err = unpopulate(val, "Server", &m.Server) + delete(rawMsg, key) + case "username": + err = unpopulate(val, "Username", &m.Username) delete(rawMsg, key) } if err != nil { @@ -32904,8 +32948,15 @@ func (m *MySQLLinkedService) UnmarshalJSON(data []byte) error { func (m MySQLLinkedServiceTypeProperties) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) populateAny(objectMap, "connectionString", m.ConnectionString) + populateAny(objectMap, "database", m.Database) + populateAny(objectMap, "driverVersion", m.DriverVersion) populate(objectMap, "encryptedCredential", m.EncryptedCredential) populate(objectMap, "password", m.Password) + populateAny(objectMap, "port", m.Port) + populateAny(objectMap, "sslMode", m.SSLMode) + populateAny(objectMap, "server", m.Server) + populateAny(objectMap, "useSystemTrustStore", m.UseSystemTrustStore) + populateAny(objectMap, "username", m.Username) return json.Marshal(objectMap) } @@ -32921,12 +32972,33 @@ func (m *MySQLLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { case "connectionString": err = unpopulate(val, "ConnectionString", &m.ConnectionString) delete(rawMsg, key) + case "database": + err = unpopulate(val, "Database", &m.Database) + delete(rawMsg, key) + case "driverVersion": + err = unpopulate(val, "DriverVersion", &m.DriverVersion) + delete(rawMsg, key) case "encryptedCredential": err = unpopulate(val, "EncryptedCredential", &m.EncryptedCredential) delete(rawMsg, key) case "password": err = unpopulate(val, "Password", &m.Password) delete(rawMsg, key) + case "port": + err = unpopulate(val, "Port", &m.Port) + delete(rawMsg, key) + case "sslMode": + err = unpopulate(val, "SSLMode", &m.SSLMode) + delete(rawMsg, key) + case "server": + err = unpopulate(val, "Server", &m.Server) + delete(rawMsg, key) + case "useSystemTrustStore": + err = unpopulate(val, "UseSystemTrustStore", &m.UseSystemTrustStore) + delete(rawMsg, key) + case "username": + err = unpopulate(val, "Username", &m.Username) + delete(rawMsg, key) } if err != nil { return fmt.Errorf("unmarshalling type %T: %v", m, err) @@ -42760,166 +42832,14 @@ func (s *SalesforceServiceCloudSource) UnmarshalJSON(data []byte) error { return nil } -// MarshalJSON implements the json.Marshaller interface for type SalesforceSink. -func (s SalesforceSink) MarshalJSON() ([]byte, error) { - objectMap := make(map[string]any) - populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) - populateAny(objectMap, "externalIdFieldName", s.ExternalIDFieldName) - populateAny(objectMap, "ignoreNullValues", s.IgnoreNullValues) - populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) - populateAny(objectMap, "sinkRetryCount", s.SinkRetryCount) - populateAny(objectMap, "sinkRetryWait", s.SinkRetryWait) - objectMap["type"] = "SalesforceSink" - populateAny(objectMap, "writeBatchSize", s.WriteBatchSize) - populateAny(objectMap, "writeBatchTimeout", s.WriteBatchTimeout) - populate(objectMap, "writeBehavior", s.WriteBehavior) - if s.AdditionalProperties != nil { - for key, val := range s.AdditionalProperties { - objectMap[key] = val - } - } - return json.Marshal(objectMap) -} - -// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceSink. -func (s *SalesforceSink) UnmarshalJSON(data []byte) error { - var rawMsg map[string]json.RawMessage - if err := json.Unmarshal(data, &rawMsg); err != nil { - return fmt.Errorf("unmarshalling type %T: %v", s, err) - } - for key, val := range rawMsg { - var err error - switch key { - case "disableMetricsCollection": - err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) - delete(rawMsg, key) - case "externalIdFieldName": - err = unpopulate(val, "ExternalIDFieldName", &s.ExternalIDFieldName) - delete(rawMsg, key) - case "ignoreNullValues": - err = unpopulate(val, "IgnoreNullValues", &s.IgnoreNullValues) - delete(rawMsg, key) - case "maxConcurrentConnections": - err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) - delete(rawMsg, key) - case "sinkRetryCount": - err = unpopulate(val, "SinkRetryCount", &s.SinkRetryCount) - delete(rawMsg, key) - case "sinkRetryWait": - err = unpopulate(val, "SinkRetryWait", &s.SinkRetryWait) - delete(rawMsg, key) - case "type": - err = unpopulate(val, "Type", &s.Type) - delete(rawMsg, key) - case "writeBatchSize": - err = unpopulate(val, "WriteBatchSize", &s.WriteBatchSize) - delete(rawMsg, key) - case "writeBatchTimeout": - err = unpopulate(val, "WriteBatchTimeout", &s.WriteBatchTimeout) - delete(rawMsg, key) - case "writeBehavior": - err = unpopulate(val, "WriteBehavior", &s.WriteBehavior) - delete(rawMsg, key) - default: - if s.AdditionalProperties == nil { - s.AdditionalProperties = map[string]any{} - } - if val != nil { - var aux any - err = json.Unmarshal(val, &aux) - s.AdditionalProperties[key] = aux - } - delete(rawMsg, key) - } - if err != nil { - return fmt.Errorf("unmarshalling type %T: %v", s, err) - } - } - return nil -} - -// MarshalJSON implements the json.Marshaller interface for type SalesforceSource. -func (s SalesforceSource) MarshalJSON() ([]byte, error) { - objectMap := make(map[string]any) - populateAny(objectMap, "additionalColumns", s.AdditionalColumns) - populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) - populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) - populateAny(objectMap, "query", s.Query) - populateAny(objectMap, "queryTimeout", s.QueryTimeout) - populateAny(objectMap, "readBehavior", s.ReadBehavior) - populateAny(objectMap, "sourceRetryCount", s.SourceRetryCount) - populateAny(objectMap, "sourceRetryWait", s.SourceRetryWait) - objectMap["type"] = "SalesforceSource" - if s.AdditionalProperties != nil { - for key, val := range s.AdditionalProperties { - objectMap[key] = val - } - } - return json.Marshal(objectMap) -} - -// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceSource. -func (s *SalesforceSource) UnmarshalJSON(data []byte) error { - var rawMsg map[string]json.RawMessage - if err := json.Unmarshal(data, &rawMsg); err != nil { - return fmt.Errorf("unmarshalling type %T: %v", s, err) - } - for key, val := range rawMsg { - var err error - switch key { - case "additionalColumns": - err = unpopulate(val, "AdditionalColumns", &s.AdditionalColumns) - delete(rawMsg, key) - case "disableMetricsCollection": - err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) - delete(rawMsg, key) - case "maxConcurrentConnections": - err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) - delete(rawMsg, key) - case "query": - err = unpopulate(val, "Query", &s.Query) - delete(rawMsg, key) - case "queryTimeout": - err = unpopulate(val, "QueryTimeout", &s.QueryTimeout) - delete(rawMsg, key) - case "readBehavior": - err = unpopulate(val, "ReadBehavior", &s.ReadBehavior) - delete(rawMsg, key) - case "sourceRetryCount": - err = unpopulate(val, "SourceRetryCount", &s.SourceRetryCount) - delete(rawMsg, key) - case "sourceRetryWait": - err = unpopulate(val, "SourceRetryWait", &s.SourceRetryWait) - delete(rawMsg, key) - case "type": - err = unpopulate(val, "Type", &s.Type) - delete(rawMsg, key) - default: - if s.AdditionalProperties == nil { - s.AdditionalProperties = map[string]any{} - } - if val != nil { - var aux any - err = json.Unmarshal(val, &aux) - s.AdditionalProperties[key] = aux - } - delete(rawMsg, key) - } - if err != nil { - return fmt.Errorf("unmarshalling type %T: %v", s, err) - } - } - return nil -} - -// MarshalJSON implements the json.Marshaller interface for type SapBWLinkedService. -func (s SapBWLinkedService) MarshalJSON() ([]byte, error) { +// MarshalJSON implements the json.Marshaller interface for type SalesforceServiceCloudV2LinkedService. +func (s SalesforceServiceCloudV2LinkedService) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) populate(objectMap, "annotations", s.Annotations) populate(objectMap, "connectVia", s.ConnectVia) populate(objectMap, "description", s.Description) populate(objectMap, "parameters", s.Parameters) - objectMap["type"] = "SapBW" + objectMap["type"] = "SalesforceServiceCloudV2" populate(objectMap, "typeProperties", s.TypeProperties) if s.AdditionalProperties != nil { for key, val := range s.AdditionalProperties { @@ -42929,8 +42849,8 @@ func (s SapBWLinkedService) MarshalJSON() ([]byte, error) { return json.Marshal(objectMap) } -// UnmarshalJSON implements the json.Unmarshaller interface for type SapBWLinkedService. -func (s *SapBWLinkedService) UnmarshalJSON(data []byte) error { +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceServiceCloudV2LinkedService. +func (s *SalesforceServiceCloudV2LinkedService) UnmarshalJSON(data []byte) error { var rawMsg map[string]json.RawMessage if err := json.Unmarshal(data, &rawMsg); err != nil { return fmt.Errorf("unmarshalling type %T: %v", s, err) @@ -42974,20 +42894,20 @@ func (s *SapBWLinkedService) UnmarshalJSON(data []byte) error { return nil } -// MarshalJSON implements the json.Marshaller interface for type SapBWLinkedServiceTypeProperties. -func (s SapBWLinkedServiceTypeProperties) MarshalJSON() ([]byte, error) { +// MarshalJSON implements the json.Marshaller interface for type SalesforceServiceCloudV2LinkedServiceTypeProperties. +func (s SalesforceServiceCloudV2LinkedServiceTypeProperties) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) + populateAny(objectMap, "apiVersion", s.APIVersion) + populateAny(objectMap, "authenticationType", s.AuthenticationType) populateAny(objectMap, "clientId", s.ClientID) + populate(objectMap, "clientSecret", s.ClientSecret) populate(objectMap, "encryptedCredential", s.EncryptedCredential) - populate(objectMap, "password", s.Password) - populateAny(objectMap, "server", s.Server) - populateAny(objectMap, "systemNumber", s.SystemNumber) - populateAny(objectMap, "userName", s.UserName) + populateAny(objectMap, "environmentUrl", s.EnvironmentURL) return json.Marshal(objectMap) } -// UnmarshalJSON implements the json.Unmarshaller interface for type SapBWLinkedServiceTypeProperties. -func (s *SapBWLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceServiceCloudV2LinkedServiceTypeProperties. +func (s *SalesforceServiceCloudV2LinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { var rawMsg map[string]json.RawMessage if err := json.Unmarshal(data, &rawMsg); err != nil { return fmt.Errorf("unmarshalling type %T: %v", s, err) @@ -42995,23 +42915,23 @@ func (s *SapBWLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { for key, val := range rawMsg { var err error switch key { + case "apiVersion": + err = unpopulate(val, "APIVersion", &s.APIVersion) + delete(rawMsg, key) + case "authenticationType": + err = unpopulate(val, "AuthenticationType", &s.AuthenticationType) + delete(rawMsg, key) case "clientId": err = unpopulate(val, "ClientID", &s.ClientID) delete(rawMsg, key) + case "clientSecret": + s.ClientSecret, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) case "encryptedCredential": err = unpopulate(val, "EncryptedCredential", &s.EncryptedCredential) delete(rawMsg, key) - case "password": - s.Password, err = unmarshalSecretBaseClassification(val) - delete(rawMsg, key) - case "server": - err = unpopulate(val, "Server", &s.Server) - delete(rawMsg, key) - case "systemNumber": - err = unpopulate(val, "SystemNumber", &s.SystemNumber) - delete(rawMsg, key) - case "userName": - err = unpopulate(val, "UserName", &s.UserName) + case "environmentUrl": + err = unpopulate(val, "EnvironmentURL", &s.EnvironmentURL) delete(rawMsg, key) } if err != nil { @@ -43021,8 +42941,8 @@ func (s *SapBWLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { return nil } -// MarshalJSON implements the json.Marshaller interface for type SapBwCubeDataset. -func (s SapBwCubeDataset) MarshalJSON() ([]byte, error) { +// MarshalJSON implements the json.Marshaller interface for type SalesforceServiceCloudV2ObjectDataset. +func (s SalesforceServiceCloudV2ObjectDataset) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) populate(objectMap, "annotations", s.Annotations) populate(objectMap, "description", s.Description) @@ -43031,7 +42951,8 @@ func (s SapBwCubeDataset) MarshalJSON() ([]byte, error) { populate(objectMap, "parameters", s.Parameters) populateAny(objectMap, "schema", s.Schema) populateAny(objectMap, "structure", s.Structure) - objectMap["type"] = "SapBwCube" + objectMap["type"] = "SalesforceServiceCloudV2Object" + populate(objectMap, "typeProperties", s.TypeProperties) if s.AdditionalProperties != nil { for key, val := range s.AdditionalProperties { objectMap[key] = val @@ -43040,8 +42961,887 @@ func (s SapBwCubeDataset) MarshalJSON() ([]byte, error) { return json.Marshal(objectMap) } -// UnmarshalJSON implements the json.Unmarshaller interface for type SapBwCubeDataset. -func (s *SapBwCubeDataset) UnmarshalJSON(data []byte) error { +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceServiceCloudV2ObjectDataset. +func (s *SalesforceServiceCloudV2ObjectDataset) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &s.Annotations) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &s.Description) + delete(rawMsg, key) + case "folder": + err = unpopulate(val, "Folder", &s.Folder) + delete(rawMsg, key) + case "linkedServiceName": + err = unpopulate(val, "LinkedServiceName", &s.LinkedServiceName) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &s.Parameters) + delete(rawMsg, key) + case "schema": + err = unpopulate(val, "Schema", &s.Schema) + delete(rawMsg, key) + case "structure": + err = unpopulate(val, "Structure", &s.Structure) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &s.TypeProperties) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceServiceCloudV2ObjectDatasetTypeProperties. +func (s SalesforceServiceCloudV2ObjectDatasetTypeProperties) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "objectApiName", s.ObjectAPIName) + populateAny(objectMap, "reportId", s.ReportID) + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceServiceCloudV2ObjectDatasetTypeProperties. +func (s *SalesforceServiceCloudV2ObjectDatasetTypeProperties) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "objectApiName": + err = unpopulate(val, "ObjectAPIName", &s.ObjectAPIName) + delete(rawMsg, key) + case "reportId": + err = unpopulate(val, "ReportID", &s.ReportID) + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceServiceCloudV2Sink. +func (s SalesforceServiceCloudV2Sink) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populateAny(objectMap, "externalIdFieldName", s.ExternalIDFieldName) + populateAny(objectMap, "ignoreNullValues", s.IgnoreNullValues) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "sinkRetryCount", s.SinkRetryCount) + populateAny(objectMap, "sinkRetryWait", s.SinkRetryWait) + objectMap["type"] = "SalesforceServiceCloudV2Sink" + populateAny(objectMap, "writeBatchSize", s.WriteBatchSize) + populateAny(objectMap, "writeBatchTimeout", s.WriteBatchTimeout) + populate(objectMap, "writeBehavior", s.WriteBehavior) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceServiceCloudV2Sink. +func (s *SalesforceServiceCloudV2Sink) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "externalIdFieldName": + err = unpopulate(val, "ExternalIDFieldName", &s.ExternalIDFieldName) + delete(rawMsg, key) + case "ignoreNullValues": + err = unpopulate(val, "IgnoreNullValues", &s.IgnoreNullValues) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "sinkRetryCount": + err = unpopulate(val, "SinkRetryCount", &s.SinkRetryCount) + delete(rawMsg, key) + case "sinkRetryWait": + err = unpopulate(val, "SinkRetryWait", &s.SinkRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "writeBatchSize": + err = unpopulate(val, "WriteBatchSize", &s.WriteBatchSize) + delete(rawMsg, key) + case "writeBatchTimeout": + err = unpopulate(val, "WriteBatchTimeout", &s.WriteBatchTimeout) + delete(rawMsg, key) + case "writeBehavior": + err = unpopulate(val, "WriteBehavior", &s.WriteBehavior) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceServiceCloudV2Source. +func (s SalesforceServiceCloudV2Source) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "additionalColumns", s.AdditionalColumns) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populateAny(objectMap, "includeDeletedObjects", s.IncludeDeletedObjects) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "SOQLQuery", s.SOQLQuery) + populateAny(objectMap, "sourceRetryCount", s.SourceRetryCount) + populateAny(objectMap, "sourceRetryWait", s.SourceRetryWait) + objectMap["type"] = "SalesforceServiceCloudV2Source" + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceServiceCloudV2Source. +func (s *SalesforceServiceCloudV2Source) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "additionalColumns": + err = unpopulate(val, "AdditionalColumns", &s.AdditionalColumns) + delete(rawMsg, key) + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "includeDeletedObjects": + err = unpopulate(val, "IncludeDeletedObjects", &s.IncludeDeletedObjects) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "SOQLQuery": + err = unpopulate(val, "SOQLQuery", &s.SOQLQuery) + delete(rawMsg, key) + case "sourceRetryCount": + err = unpopulate(val, "SourceRetryCount", &s.SourceRetryCount) + delete(rawMsg, key) + case "sourceRetryWait": + err = unpopulate(val, "SourceRetryWait", &s.SourceRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceSink. +func (s SalesforceSink) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populateAny(objectMap, "externalIdFieldName", s.ExternalIDFieldName) + populateAny(objectMap, "ignoreNullValues", s.IgnoreNullValues) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "sinkRetryCount", s.SinkRetryCount) + populateAny(objectMap, "sinkRetryWait", s.SinkRetryWait) + objectMap["type"] = "SalesforceSink" + populateAny(objectMap, "writeBatchSize", s.WriteBatchSize) + populateAny(objectMap, "writeBatchTimeout", s.WriteBatchTimeout) + populate(objectMap, "writeBehavior", s.WriteBehavior) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceSink. +func (s *SalesforceSink) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "externalIdFieldName": + err = unpopulate(val, "ExternalIDFieldName", &s.ExternalIDFieldName) + delete(rawMsg, key) + case "ignoreNullValues": + err = unpopulate(val, "IgnoreNullValues", &s.IgnoreNullValues) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "sinkRetryCount": + err = unpopulate(val, "SinkRetryCount", &s.SinkRetryCount) + delete(rawMsg, key) + case "sinkRetryWait": + err = unpopulate(val, "SinkRetryWait", &s.SinkRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "writeBatchSize": + err = unpopulate(val, "WriteBatchSize", &s.WriteBatchSize) + delete(rawMsg, key) + case "writeBatchTimeout": + err = unpopulate(val, "WriteBatchTimeout", &s.WriteBatchTimeout) + delete(rawMsg, key) + case "writeBehavior": + err = unpopulate(val, "WriteBehavior", &s.WriteBehavior) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceSource. +func (s SalesforceSource) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "additionalColumns", s.AdditionalColumns) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "query", s.Query) + populateAny(objectMap, "queryTimeout", s.QueryTimeout) + populateAny(objectMap, "readBehavior", s.ReadBehavior) + populateAny(objectMap, "sourceRetryCount", s.SourceRetryCount) + populateAny(objectMap, "sourceRetryWait", s.SourceRetryWait) + objectMap["type"] = "SalesforceSource" + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceSource. +func (s *SalesforceSource) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "additionalColumns": + err = unpopulate(val, "AdditionalColumns", &s.AdditionalColumns) + delete(rawMsg, key) + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "query": + err = unpopulate(val, "Query", &s.Query) + delete(rawMsg, key) + case "queryTimeout": + err = unpopulate(val, "QueryTimeout", &s.QueryTimeout) + delete(rawMsg, key) + case "readBehavior": + err = unpopulate(val, "ReadBehavior", &s.ReadBehavior) + delete(rawMsg, key) + case "sourceRetryCount": + err = unpopulate(val, "SourceRetryCount", &s.SourceRetryCount) + delete(rawMsg, key) + case "sourceRetryWait": + err = unpopulate(val, "SourceRetryWait", &s.SourceRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceV2LinkedService. +func (s SalesforceV2LinkedService) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", s.Annotations) + populate(objectMap, "connectVia", s.ConnectVia) + populate(objectMap, "description", s.Description) + populate(objectMap, "parameters", s.Parameters) + objectMap["type"] = "SalesforceV2" + populate(objectMap, "typeProperties", s.TypeProperties) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceV2LinkedService. +func (s *SalesforceV2LinkedService) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &s.Annotations) + delete(rawMsg, key) + case "connectVia": + err = unpopulate(val, "ConnectVia", &s.ConnectVia) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &s.Description) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &s.Parameters) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &s.TypeProperties) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceV2LinkedServiceTypeProperties. +func (s SalesforceV2LinkedServiceTypeProperties) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "apiVersion", s.APIVersion) + populateAny(objectMap, "authenticationType", s.AuthenticationType) + populateAny(objectMap, "clientId", s.ClientID) + populate(objectMap, "clientSecret", s.ClientSecret) + populate(objectMap, "encryptedCredential", s.EncryptedCredential) + populateAny(objectMap, "environmentUrl", s.EnvironmentURL) + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceV2LinkedServiceTypeProperties. +func (s *SalesforceV2LinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "apiVersion": + err = unpopulate(val, "APIVersion", &s.APIVersion) + delete(rawMsg, key) + case "authenticationType": + err = unpopulate(val, "AuthenticationType", &s.AuthenticationType) + delete(rawMsg, key) + case "clientId": + err = unpopulate(val, "ClientID", &s.ClientID) + delete(rawMsg, key) + case "clientSecret": + s.ClientSecret, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "encryptedCredential": + err = unpopulate(val, "EncryptedCredential", &s.EncryptedCredential) + delete(rawMsg, key) + case "environmentUrl": + err = unpopulate(val, "EnvironmentURL", &s.EnvironmentURL) + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceV2ObjectDataset. +func (s SalesforceV2ObjectDataset) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", s.Annotations) + populate(objectMap, "description", s.Description) + populate(objectMap, "folder", s.Folder) + populate(objectMap, "linkedServiceName", s.LinkedServiceName) + populate(objectMap, "parameters", s.Parameters) + populateAny(objectMap, "schema", s.Schema) + populateAny(objectMap, "structure", s.Structure) + objectMap["type"] = "SalesforceV2Object" + populate(objectMap, "typeProperties", s.TypeProperties) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceV2ObjectDataset. +func (s *SalesforceV2ObjectDataset) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &s.Annotations) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &s.Description) + delete(rawMsg, key) + case "folder": + err = unpopulate(val, "Folder", &s.Folder) + delete(rawMsg, key) + case "linkedServiceName": + err = unpopulate(val, "LinkedServiceName", &s.LinkedServiceName) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &s.Parameters) + delete(rawMsg, key) + case "schema": + err = unpopulate(val, "Schema", &s.Schema) + delete(rawMsg, key) + case "structure": + err = unpopulate(val, "Structure", &s.Structure) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &s.TypeProperties) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceV2ObjectDatasetTypeProperties. +func (s SalesforceV2ObjectDatasetTypeProperties) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "objectApiName", s.ObjectAPIName) + populateAny(objectMap, "reportId", s.ReportID) + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceV2ObjectDatasetTypeProperties. +func (s *SalesforceV2ObjectDatasetTypeProperties) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "objectApiName": + err = unpopulate(val, "ObjectAPIName", &s.ObjectAPIName) + delete(rawMsg, key) + case "reportId": + err = unpopulate(val, "ReportID", &s.ReportID) + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceV2Sink. +func (s SalesforceV2Sink) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populateAny(objectMap, "externalIdFieldName", s.ExternalIDFieldName) + populateAny(objectMap, "ignoreNullValues", s.IgnoreNullValues) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "sinkRetryCount", s.SinkRetryCount) + populateAny(objectMap, "sinkRetryWait", s.SinkRetryWait) + objectMap["type"] = "SalesforceV2Sink" + populateAny(objectMap, "writeBatchSize", s.WriteBatchSize) + populateAny(objectMap, "writeBatchTimeout", s.WriteBatchTimeout) + populate(objectMap, "writeBehavior", s.WriteBehavior) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceV2Sink. +func (s *SalesforceV2Sink) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "externalIdFieldName": + err = unpopulate(val, "ExternalIDFieldName", &s.ExternalIDFieldName) + delete(rawMsg, key) + case "ignoreNullValues": + err = unpopulate(val, "IgnoreNullValues", &s.IgnoreNullValues) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "sinkRetryCount": + err = unpopulate(val, "SinkRetryCount", &s.SinkRetryCount) + delete(rawMsg, key) + case "sinkRetryWait": + err = unpopulate(val, "SinkRetryWait", &s.SinkRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "writeBatchSize": + err = unpopulate(val, "WriteBatchSize", &s.WriteBatchSize) + delete(rawMsg, key) + case "writeBatchTimeout": + err = unpopulate(val, "WriteBatchTimeout", &s.WriteBatchTimeout) + delete(rawMsg, key) + case "writeBehavior": + err = unpopulate(val, "WriteBehavior", &s.WriteBehavior) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SalesforceV2Source. +func (s SalesforceV2Source) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "additionalColumns", s.AdditionalColumns) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populateAny(objectMap, "includeDeletedObjects", s.IncludeDeletedObjects) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "queryTimeout", s.QueryTimeout) + populateAny(objectMap, "SOQLQuery", s.SOQLQuery) + populateAny(objectMap, "sourceRetryCount", s.SourceRetryCount) + populateAny(objectMap, "sourceRetryWait", s.SourceRetryWait) + objectMap["type"] = "SalesforceV2Source" + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SalesforceV2Source. +func (s *SalesforceV2Source) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "additionalColumns": + err = unpopulate(val, "AdditionalColumns", &s.AdditionalColumns) + delete(rawMsg, key) + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "includeDeletedObjects": + err = unpopulate(val, "IncludeDeletedObjects", &s.IncludeDeletedObjects) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "queryTimeout": + err = unpopulate(val, "QueryTimeout", &s.QueryTimeout) + delete(rawMsg, key) + case "SOQLQuery": + err = unpopulate(val, "SOQLQuery", &s.SOQLQuery) + delete(rawMsg, key) + case "sourceRetryCount": + err = unpopulate(val, "SourceRetryCount", &s.SourceRetryCount) + delete(rawMsg, key) + case "sourceRetryWait": + err = unpopulate(val, "SourceRetryWait", &s.SourceRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SapBWLinkedService. +func (s SapBWLinkedService) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", s.Annotations) + populate(objectMap, "connectVia", s.ConnectVia) + populate(objectMap, "description", s.Description) + populate(objectMap, "parameters", s.Parameters) + objectMap["type"] = "SapBW" + populate(objectMap, "typeProperties", s.TypeProperties) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SapBWLinkedService. +func (s *SapBWLinkedService) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &s.Annotations) + delete(rawMsg, key) + case "connectVia": + err = unpopulate(val, "ConnectVia", &s.ConnectVia) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &s.Description) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &s.Parameters) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &s.TypeProperties) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SapBWLinkedServiceTypeProperties. +func (s SapBWLinkedServiceTypeProperties) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "clientId", s.ClientID) + populate(objectMap, "encryptedCredential", s.EncryptedCredential) + populate(objectMap, "password", s.Password) + populateAny(objectMap, "server", s.Server) + populateAny(objectMap, "systemNumber", s.SystemNumber) + populateAny(objectMap, "userName", s.UserName) + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SapBWLinkedServiceTypeProperties. +func (s *SapBWLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "clientId": + err = unpopulate(val, "ClientID", &s.ClientID) + delete(rawMsg, key) + case "encryptedCredential": + err = unpopulate(val, "EncryptedCredential", &s.EncryptedCredential) + delete(rawMsg, key) + case "password": + s.Password, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "server": + err = unpopulate(val, "Server", &s.Server) + delete(rawMsg, key) + case "systemNumber": + err = unpopulate(val, "SystemNumber", &s.SystemNumber) + delete(rawMsg, key) + case "userName": + err = unpopulate(val, "UserName", &s.UserName) + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SapBwCubeDataset. +func (s SapBwCubeDataset) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", s.Annotations) + populate(objectMap, "description", s.Description) + populate(objectMap, "folder", s.Folder) + populate(objectMap, "linkedServiceName", s.LinkedServiceName) + populate(objectMap, "parameters", s.Parameters) + populateAny(objectMap, "schema", s.Schema) + populateAny(objectMap, "structure", s.Structure) + objectMap["type"] = "SapBwCube" + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SapBwCubeDataset. +func (s *SapBwCubeDataset) UnmarshalJSON(data []byte) error { var rawMsg map[string]json.RawMessage if err := json.Unmarshal(data, &rawMsg); err != nil { return fmt.Errorf("unmarshalling type %T: %v", s, err) @@ -46779,6 +47579,7 @@ func (s SftpWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "copyBehavior", s.CopyBehavior) populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populate(objectMap, "metadata", s.Metadata) populateAny(objectMap, "operationTimeout", s.OperationTimeout) objectMap["type"] = "SftpWriteSettings" populateAny(objectMap, "useTempFileRename", s.UseTempFileRename) @@ -46808,6 +47609,9 @@ func (s *SftpWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &s.Metadata) + delete(rawMsg, key) case "operationTimeout": err = unpopulate(val, "OperationTimeout", &s.OperationTimeout) delete(rawMsg, key) @@ -47786,6 +48590,81 @@ func (s *SnowflakeLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error return nil } +// MarshalJSON implements the json.Marshaller interface for type SnowflakeLinkedV2ServiceTypeProperties. +func (s SnowflakeLinkedV2ServiceTypeProperties) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "accountIdentifier", s.AccountIdentifier) + populate(objectMap, "authenticationType", s.AuthenticationType) + populateAny(objectMap, "clientId", s.ClientID) + populate(objectMap, "clientSecret", s.ClientSecret) + populateAny(objectMap, "database", s.Database) + populate(objectMap, "encryptedCredential", s.EncryptedCredential) + populate(objectMap, "password", s.Password) + populate(objectMap, "privateKey", s.PrivateKey) + populate(objectMap, "privateKeyPassphrase", s.PrivateKeyPassphrase) + populateAny(objectMap, "scope", s.Scope) + populateAny(objectMap, "tenantId", s.TenantID) + populateAny(objectMap, "user", s.User) + populateAny(objectMap, "warehouse", s.Warehouse) + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SnowflakeLinkedV2ServiceTypeProperties. +func (s *SnowflakeLinkedV2ServiceTypeProperties) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "accountIdentifier": + err = unpopulate(val, "AccountIdentifier", &s.AccountIdentifier) + delete(rawMsg, key) + case "authenticationType": + err = unpopulate(val, "AuthenticationType", &s.AuthenticationType) + delete(rawMsg, key) + case "clientId": + err = unpopulate(val, "ClientID", &s.ClientID) + delete(rawMsg, key) + case "clientSecret": + s.ClientSecret, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "database": + err = unpopulate(val, "Database", &s.Database) + delete(rawMsg, key) + case "encryptedCredential": + err = unpopulate(val, "EncryptedCredential", &s.EncryptedCredential) + delete(rawMsg, key) + case "password": + s.Password, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "privateKey": + s.PrivateKey, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "privateKeyPassphrase": + s.PrivateKeyPassphrase, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "scope": + err = unpopulate(val, "Scope", &s.Scope) + delete(rawMsg, key) + case "tenantId": + err = unpopulate(val, "TenantID", &s.TenantID) + delete(rawMsg, key) + case "user": + err = unpopulate(val, "User", &s.User) + delete(rawMsg, key) + case "warehouse": + err = unpopulate(val, "Warehouse", &s.Warehouse) + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + // MarshalJSON implements the json.Marshaller interface for type SnowflakeSink. func (s SnowflakeSink) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) @@ -47926,6 +48805,282 @@ func (s *SnowflakeSource) UnmarshalJSON(data []byte) error { return nil } +// MarshalJSON implements the json.Marshaller interface for type SnowflakeV2Dataset. +func (s SnowflakeV2Dataset) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", s.Annotations) + populate(objectMap, "description", s.Description) + populate(objectMap, "folder", s.Folder) + populate(objectMap, "linkedServiceName", s.LinkedServiceName) + populate(objectMap, "parameters", s.Parameters) + populateAny(objectMap, "schema", s.Schema) + populateAny(objectMap, "structure", s.Structure) + objectMap["type"] = "SnowflakeV2Table" + populate(objectMap, "typeProperties", s.TypeProperties) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SnowflakeV2Dataset. +func (s *SnowflakeV2Dataset) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &s.Annotations) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &s.Description) + delete(rawMsg, key) + case "folder": + err = unpopulate(val, "Folder", &s.Folder) + delete(rawMsg, key) + case "linkedServiceName": + err = unpopulate(val, "LinkedServiceName", &s.LinkedServiceName) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &s.Parameters) + delete(rawMsg, key) + case "schema": + err = unpopulate(val, "Schema", &s.Schema) + delete(rawMsg, key) + case "structure": + err = unpopulate(val, "Structure", &s.Structure) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &s.TypeProperties) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SnowflakeV2LinkedService. +func (s SnowflakeV2LinkedService) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", s.Annotations) + populate(objectMap, "connectVia", s.ConnectVia) + populate(objectMap, "description", s.Description) + populate(objectMap, "parameters", s.Parameters) + objectMap["type"] = "SnowflakeV2" + populate(objectMap, "typeProperties", s.TypeProperties) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SnowflakeV2LinkedService. +func (s *SnowflakeV2LinkedService) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &s.Annotations) + delete(rawMsg, key) + case "connectVia": + err = unpopulate(val, "ConnectVia", &s.ConnectVia) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &s.Description) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &s.Parameters) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &s.TypeProperties) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SnowflakeV2Sink. +func (s SnowflakeV2Sink) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populate(objectMap, "importSettings", s.ImportSettings) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "preCopyScript", s.PreCopyScript) + populateAny(objectMap, "sinkRetryCount", s.SinkRetryCount) + populateAny(objectMap, "sinkRetryWait", s.SinkRetryWait) + objectMap["type"] = "SnowflakeV2Sink" + populateAny(objectMap, "writeBatchSize", s.WriteBatchSize) + populateAny(objectMap, "writeBatchTimeout", s.WriteBatchTimeout) + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SnowflakeV2Sink. +func (s *SnowflakeV2Sink) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "importSettings": + err = unpopulate(val, "ImportSettings", &s.ImportSettings) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "preCopyScript": + err = unpopulate(val, "PreCopyScript", &s.PreCopyScript) + delete(rawMsg, key) + case "sinkRetryCount": + err = unpopulate(val, "SinkRetryCount", &s.SinkRetryCount) + delete(rawMsg, key) + case "sinkRetryWait": + err = unpopulate(val, "SinkRetryWait", &s.SinkRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + case "writeBatchSize": + err = unpopulate(val, "WriteBatchSize", &s.WriteBatchSize) + delete(rawMsg, key) + case "writeBatchTimeout": + err = unpopulate(val, "WriteBatchTimeout", &s.WriteBatchTimeout) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type SnowflakeV2Source. +func (s SnowflakeV2Source) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) + populate(objectMap, "exportSettings", s.ExportSettings) + populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populateAny(objectMap, "query", s.Query) + populateAny(objectMap, "sourceRetryCount", s.SourceRetryCount) + populateAny(objectMap, "sourceRetryWait", s.SourceRetryWait) + objectMap["type"] = "SnowflakeV2Source" + if s.AdditionalProperties != nil { + for key, val := range s.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type SnowflakeV2Source. +func (s *SnowflakeV2Source) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &s.DisableMetricsCollection) + delete(rawMsg, key) + case "exportSettings": + err = unpopulate(val, "ExportSettings", &s.ExportSettings) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) + delete(rawMsg, key) + case "query": + err = unpopulate(val, "Query", &s.Query) + delete(rawMsg, key) + case "sourceRetryCount": + err = unpopulate(val, "SourceRetryCount", &s.SourceRetryCount) + delete(rawMsg, key) + case "sourceRetryWait": + err = unpopulate(val, "SourceRetryWait", &s.SourceRetryWait) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &s.Type) + delete(rawMsg, key) + default: + if s.AdditionalProperties == nil { + s.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + s.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", s, err) + } + } + return nil +} + // MarshalJSON implements the json.Marshaller interface for type SparkConfigurationParametrizationReference. func (s SparkConfigurationParametrizationReference) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) @@ -49114,6 +50269,7 @@ func (s StoreWriteSettings) MarshalJSON() ([]byte, error) { populateAny(objectMap, "copyBehavior", s.CopyBehavior) populateAny(objectMap, "disableMetricsCollection", s.DisableMetricsCollection) populateAny(objectMap, "maxConcurrentConnections", s.MaxConcurrentConnections) + populate(objectMap, "metadata", s.Metadata) objectMap["type"] = s.Type if s.AdditionalProperties != nil { for key, val := range s.AdditionalProperties { @@ -49141,6 +50297,9 @@ func (s *StoreWriteSettings) UnmarshalJSON(data []byte) error { case "maxConcurrentConnections": err = unpopulate(val, "MaxConcurrentConnections", &s.MaxConcurrentConnections) delete(rawMsg, key) + case "metadata": + err = unpopulate(val, "Metadata", &s.Metadata) + delete(rawMsg, key) case "type": err = unpopulate(val, "Type", &s.Type) delete(rawMsg, key) @@ -52099,6 +53258,408 @@ func (w *WaitActivityTypeProperties) UnmarshalJSON(data []byte) error { return nil } +// MarshalJSON implements the json.Marshaller interface for type WarehouseLinkedService. +func (w WarehouseLinkedService) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", w.Annotations) + populate(objectMap, "connectVia", w.ConnectVia) + populate(objectMap, "description", w.Description) + populate(objectMap, "parameters", w.Parameters) + objectMap["type"] = "Warehouse" + populate(objectMap, "typeProperties", w.TypeProperties) + if w.AdditionalProperties != nil { + for key, val := range w.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type WarehouseLinkedService. +func (w *WarehouseLinkedService) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &w.Annotations) + delete(rawMsg, key) + case "connectVia": + err = unpopulate(val, "ConnectVia", &w.ConnectVia) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &w.Description) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &w.Parameters) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &w.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &w.TypeProperties) + delete(rawMsg, key) + default: + if w.AdditionalProperties == nil { + w.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + w.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type WarehouseLinkedServiceTypeProperties. +func (w WarehouseLinkedServiceTypeProperties) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "artifactId", w.ArtifactID) + populate(objectMap, "encryptedCredential", w.EncryptedCredential) + populateAny(objectMap, "endpoint", w.Endpoint) + populate(objectMap, "servicePrincipalCredential", w.ServicePrincipalCredential) + populateAny(objectMap, "servicePrincipalCredentialType", w.ServicePrincipalCredentialType) + populateAny(objectMap, "servicePrincipalId", w.ServicePrincipalID) + populate(objectMap, "servicePrincipalKey", w.ServicePrincipalKey) + populateAny(objectMap, "tenant", w.Tenant) + populateAny(objectMap, "workspaceId", w.WorkspaceID) + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type WarehouseLinkedServiceTypeProperties. +func (w *WarehouseLinkedServiceTypeProperties) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "artifactId": + err = unpopulate(val, "ArtifactID", &w.ArtifactID) + delete(rawMsg, key) + case "encryptedCredential": + err = unpopulate(val, "EncryptedCredential", &w.EncryptedCredential) + delete(rawMsg, key) + case "endpoint": + err = unpopulate(val, "Endpoint", &w.Endpoint) + delete(rawMsg, key) + case "servicePrincipalCredential": + w.ServicePrincipalCredential, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "servicePrincipalCredentialType": + err = unpopulate(val, "ServicePrincipalCredentialType", &w.ServicePrincipalCredentialType) + delete(rawMsg, key) + case "servicePrincipalId": + err = unpopulate(val, "ServicePrincipalID", &w.ServicePrincipalID) + delete(rawMsg, key) + case "servicePrincipalKey": + w.ServicePrincipalKey, err = unmarshalSecretBaseClassification(val) + delete(rawMsg, key) + case "tenant": + err = unpopulate(val, "Tenant", &w.Tenant) + delete(rawMsg, key) + case "workspaceId": + err = unpopulate(val, "WorkspaceID", &w.WorkspaceID) + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type WarehouseSink. +func (w WarehouseSink) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "allowCopyCommand", w.AllowCopyCommand) + populate(objectMap, "copyCommandSettings", w.CopyCommandSettings) + populateAny(objectMap, "disableMetricsCollection", w.DisableMetricsCollection) + populateAny(objectMap, "maxConcurrentConnections", w.MaxConcurrentConnections) + populateAny(objectMap, "preCopyScript", w.PreCopyScript) + populateAny(objectMap, "sinkRetryCount", w.SinkRetryCount) + populateAny(objectMap, "sinkRetryWait", w.SinkRetryWait) + populateAny(objectMap, "tableOption", w.TableOption) + objectMap["type"] = "WarehouseSink" + populateAny(objectMap, "writeBatchSize", w.WriteBatchSize) + populateAny(objectMap, "writeBatchTimeout", w.WriteBatchTimeout) + populateAny(objectMap, "writeBehavior", w.WriteBehavior) + if w.AdditionalProperties != nil { + for key, val := range w.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type WarehouseSink. +func (w *WarehouseSink) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "allowCopyCommand": + err = unpopulate(val, "AllowCopyCommand", &w.AllowCopyCommand) + delete(rawMsg, key) + case "copyCommandSettings": + err = unpopulate(val, "CopyCommandSettings", &w.CopyCommandSettings) + delete(rawMsg, key) + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &w.DisableMetricsCollection) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &w.MaxConcurrentConnections) + delete(rawMsg, key) + case "preCopyScript": + err = unpopulate(val, "PreCopyScript", &w.PreCopyScript) + delete(rawMsg, key) + case "sinkRetryCount": + err = unpopulate(val, "SinkRetryCount", &w.SinkRetryCount) + delete(rawMsg, key) + case "sinkRetryWait": + err = unpopulate(val, "SinkRetryWait", &w.SinkRetryWait) + delete(rawMsg, key) + case "tableOption": + err = unpopulate(val, "TableOption", &w.TableOption) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &w.Type) + delete(rawMsg, key) + case "writeBatchSize": + err = unpopulate(val, "WriteBatchSize", &w.WriteBatchSize) + delete(rawMsg, key) + case "writeBatchTimeout": + err = unpopulate(val, "WriteBatchTimeout", &w.WriteBatchTimeout) + delete(rawMsg, key) + case "writeBehavior": + err = unpopulate(val, "WriteBehavior", &w.WriteBehavior) + delete(rawMsg, key) + default: + if w.AdditionalProperties == nil { + w.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + w.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type WarehouseSource. +func (w WarehouseSource) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "additionalColumns", w.AdditionalColumns) + populateAny(objectMap, "disableMetricsCollection", w.DisableMetricsCollection) + populateAny(objectMap, "isolationLevel", w.IsolationLevel) + populateAny(objectMap, "maxConcurrentConnections", w.MaxConcurrentConnections) + populateAny(objectMap, "partitionOption", w.PartitionOption) + populate(objectMap, "partitionSettings", w.PartitionSettings) + populateAny(objectMap, "queryTimeout", w.QueryTimeout) + populateAny(objectMap, "sqlReaderQuery", w.SQLReaderQuery) + populateAny(objectMap, "sqlReaderStoredProcedureName", w.SQLReaderStoredProcedureName) + populateAny(objectMap, "sourceRetryCount", w.SourceRetryCount) + populateAny(objectMap, "sourceRetryWait", w.SourceRetryWait) + populateAny(objectMap, "storedProcedureParameters", w.StoredProcedureParameters) + objectMap["type"] = "WarehouseSource" + if w.AdditionalProperties != nil { + for key, val := range w.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type WarehouseSource. +func (w *WarehouseSource) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "additionalColumns": + err = unpopulate(val, "AdditionalColumns", &w.AdditionalColumns) + delete(rawMsg, key) + case "disableMetricsCollection": + err = unpopulate(val, "DisableMetricsCollection", &w.DisableMetricsCollection) + delete(rawMsg, key) + case "isolationLevel": + err = unpopulate(val, "IsolationLevel", &w.IsolationLevel) + delete(rawMsg, key) + case "maxConcurrentConnections": + err = unpopulate(val, "MaxConcurrentConnections", &w.MaxConcurrentConnections) + delete(rawMsg, key) + case "partitionOption": + err = unpopulate(val, "PartitionOption", &w.PartitionOption) + delete(rawMsg, key) + case "partitionSettings": + err = unpopulate(val, "PartitionSettings", &w.PartitionSettings) + delete(rawMsg, key) + case "queryTimeout": + err = unpopulate(val, "QueryTimeout", &w.QueryTimeout) + delete(rawMsg, key) + case "sqlReaderQuery": + err = unpopulate(val, "SQLReaderQuery", &w.SQLReaderQuery) + delete(rawMsg, key) + case "sqlReaderStoredProcedureName": + err = unpopulate(val, "SQLReaderStoredProcedureName", &w.SQLReaderStoredProcedureName) + delete(rawMsg, key) + case "sourceRetryCount": + err = unpopulate(val, "SourceRetryCount", &w.SourceRetryCount) + delete(rawMsg, key) + case "sourceRetryWait": + err = unpopulate(val, "SourceRetryWait", &w.SourceRetryWait) + delete(rawMsg, key) + case "storedProcedureParameters": + err = unpopulate(val, "StoredProcedureParameters", &w.StoredProcedureParameters) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &w.Type) + delete(rawMsg, key) + default: + if w.AdditionalProperties == nil { + w.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + w.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type WarehouseTableDataset. +func (w WarehouseTableDataset) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populate(objectMap, "annotations", w.Annotations) + populate(objectMap, "description", w.Description) + populate(objectMap, "folder", w.Folder) + populate(objectMap, "linkedServiceName", w.LinkedServiceName) + populate(objectMap, "parameters", w.Parameters) + populateAny(objectMap, "schema", w.Schema) + populateAny(objectMap, "structure", w.Structure) + objectMap["type"] = "WarehouseTable" + populate(objectMap, "typeProperties", w.TypeProperties) + if w.AdditionalProperties != nil { + for key, val := range w.AdditionalProperties { + objectMap[key] = val + } + } + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type WarehouseTableDataset. +func (w *WarehouseTableDataset) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "annotations": + err = unpopulate(val, "Annotations", &w.Annotations) + delete(rawMsg, key) + case "description": + err = unpopulate(val, "Description", &w.Description) + delete(rawMsg, key) + case "folder": + err = unpopulate(val, "Folder", &w.Folder) + delete(rawMsg, key) + case "linkedServiceName": + err = unpopulate(val, "LinkedServiceName", &w.LinkedServiceName) + delete(rawMsg, key) + case "parameters": + err = unpopulate(val, "Parameters", &w.Parameters) + delete(rawMsg, key) + case "schema": + err = unpopulate(val, "Schema", &w.Schema) + delete(rawMsg, key) + case "structure": + err = unpopulate(val, "Structure", &w.Structure) + delete(rawMsg, key) + case "type": + err = unpopulate(val, "Type", &w.Type) + delete(rawMsg, key) + case "typeProperties": + err = unpopulate(val, "TypeProperties", &w.TypeProperties) + delete(rawMsg, key) + default: + if w.AdditionalProperties == nil { + w.AdditionalProperties = map[string]any{} + } + if val != nil { + var aux any + err = json.Unmarshal(val, &aux) + w.AdditionalProperties[key] = aux + } + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + } + return nil +} + +// MarshalJSON implements the json.Marshaller interface for type WarehouseTableDatasetTypeProperties. +func (w WarehouseTableDatasetTypeProperties) MarshalJSON() ([]byte, error) { + objectMap := make(map[string]any) + populateAny(objectMap, "schema", w.Schema) + populateAny(objectMap, "table", w.Table) + return json.Marshal(objectMap) +} + +// UnmarshalJSON implements the json.Unmarshaller interface for type WarehouseTableDatasetTypeProperties. +func (w *WarehouseTableDatasetTypeProperties) UnmarshalJSON(data []byte) error { + var rawMsg map[string]json.RawMessage + if err := json.Unmarshal(data, &rawMsg); err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + for key, val := range rawMsg { + var err error + switch key { + case "schema": + err = unpopulate(val, "Schema", &w.Schema) + delete(rawMsg, key) + case "table": + err = unpopulate(val, "Table", &w.Table) + delete(rawMsg, key) + } + if err != nil { + return fmt.Errorf("unmarshalling type %T: %v", w, err) + } + } + return nil +} + // MarshalJSON implements the json.Marshaller interface for type WebActivity. func (w WebActivity) MarshalJSON() ([]byte, error) { objectMap := make(map[string]any) @@ -52236,9 +53797,11 @@ func (w WebActivityTypeProperties) MarshalJSON() ([]byte, error) { populate(objectMap, "connectVia", w.ConnectVia) populate(objectMap, "datasets", w.Datasets) populate(objectMap, "disableCertValidation", w.DisableCertValidation) + populateAny(objectMap, "httpRequestTimeout", w.HTTPRequestTimeout) populateAny(objectMap, "headers", w.Headers) populate(objectMap, "linkedServices", w.LinkedServices) populate(objectMap, "method", w.Method) + populate(objectMap, "turnOffAsync", w.TurnOffAsync) populateAny(objectMap, "url", w.URL) return json.Marshal(objectMap) } @@ -52267,6 +53830,9 @@ func (w *WebActivityTypeProperties) UnmarshalJSON(data []byte) error { case "disableCertValidation": err = unpopulate(val, "DisableCertValidation", &w.DisableCertValidation) delete(rawMsg, key) + case "httpRequestTimeout": + err = unpopulate(val, "HTTPRequestTimeout", &w.HTTPRequestTimeout) + delete(rawMsg, key) case "headers": err = unpopulate(val, "Headers", &w.Headers) delete(rawMsg, key) @@ -52276,6 +53842,9 @@ func (w *WebActivityTypeProperties) UnmarshalJSON(data []byte) error { case "method": err = unpopulate(val, "Method", &w.Method) delete(rawMsg, key) + case "turnOffAsync": + err = unpopulate(val, "TurnOffAsync", &w.TurnOffAsync) + delete(rawMsg, key) case "url": err = unpopulate(val, "URL", &w.URL) delete(rawMsg, key) diff --git a/sdk/resourcemanager/datafactory/armdatafactory/operations_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/operations_client_example_test.go deleted file mode 100644 index c21490de28cb..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/operations_client_example_test.go +++ /dev/null @@ -1,1044 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Operations_List.json -func ExampleOperationsClient_NewListPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewOperationsClient().NewListPager(nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.OperationListResponse = armdatafactory.OperationListResponse{ - // Value: []*armdatafactory.Operation{ - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/providers/Microsoft.Insights/metricDefinitions/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Gets the available metrics for datafactories"), - // Operation: to.Ptr("Read datafactories metric definitions"), - // Provider: to.Ptr("Microsoft.DataFactory"), - // Resource: to.Ptr("datafactories"), - // }, - // Origin: to.Ptr("system"), - // Properties: &armdatafactory.OperationProperties{ - // ServiceSpecification: &armdatafactory.OperationServiceSpecification{ - // MetricSpecifications: []*armdatafactory.OperationMetricSpecification{ - // { - // Name: to.Ptr("FailedRuns"), - // AggregationType: to.Ptr("Total"), - // Availabilities: []*armdatafactory.OperationMetricAvailability{ - // { - // BlobDuration: to.Ptr("P1D"), - // TimeGrain: to.Ptr("PT1H"), - // }}, - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("pipelineName"), - // DisplayName: to.Ptr("Pipeline"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("activityName"), - // DisplayName: to.Ptr("Activity"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("windowEnd"), - // DisplayName: to.Ptr("Window End"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("windowStart"), - // DisplayName: to.Ptr("Window Start"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Failed Runs"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetricsV1"), - // Unit: to.Ptr("Count"), - // }, - // { - // Name: to.Ptr("SuccessfulRuns"), - // AggregationType: to.Ptr("Total"), - // Availabilities: []*armdatafactory.OperationMetricAvailability{ - // { - // BlobDuration: to.Ptr("P1D"), - // TimeGrain: to.Ptr("PT1H"), - // }}, - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("pipelineName"), - // DisplayName: to.Ptr("Pipeline"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("activityName"), - // DisplayName: to.Ptr("Activity"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("windowEnd"), - // DisplayName: to.Ptr("Window End"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("windowStart"), - // DisplayName: to.Ptr("Window Start"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Successful Runs"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetricsV1"), - // Unit: to.Ptr("Count"), - // }}, - // }, - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/providers/Microsoft.Insights/diagnosticSettings/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Gets the diagnostic setting for the resource"), - // Operation: to.Ptr("Read diagnostic setting"), - // Provider: to.Ptr("Microsoft.DataFactory"), - // Resource: to.Ptr("datafactories"), - // }, - // Origin: to.Ptr("system"), - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/providers/Microsoft.Insights/diagnosticSettings/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or updates the diagnostic setting for the resource"), - // Operation: to.Ptr("Write diagnostic setting"), - // Provider: to.Ptr("Microsoft.DataFactory"), - // Resource: to.Ptr("datafactories"), - // }, - // Origin: to.Ptr("system"), - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/providers/Microsoft.Insights/logDefinitions/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Gets the available logs for factories"), - // Operation: to.Ptr("Read factories log definitions"), - // Provider: to.Ptr("Microsoft.DataFactory"), - // Resource: to.Ptr("The log definition of factories"), - // }, - // Origin: to.Ptr("system"), - // Properties: &armdatafactory.OperationProperties{ - // ServiceSpecification: &armdatafactory.OperationServiceSpecification{ - // LogSpecifications: []*armdatafactory.OperationLogSpecification{ - // { - // Name: to.Ptr("ActivityRuns"), - // BlobDuration: to.Ptr("PT1H"), - // DisplayName: to.Ptr("Pipeline activity runs log"), - // }, - // { - // Name: to.Ptr("PipelineRuns"), - // BlobDuration: to.Ptr("PT1H"), - // DisplayName: to.Ptr("Pipeline runs log"), - // }, - // { - // Name: to.Ptr("TriggerRuns"), - // BlobDuration: to.Ptr("PT1H"), - // DisplayName: to.Ptr("Trigger runs log"), - // }}, - // }, - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/providers/Microsoft.Insights/diagnosticSettings/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Gets the diagnostic setting for the resource"), - // Operation: to.Ptr("Read diagnostic setting"), - // Provider: to.Ptr("Microsoft.DataFactory"), - // Resource: to.Ptr("factories"), - // }, - // Origin: to.Ptr("system"), - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/providers/Microsoft.Insights/diagnosticSettings/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or updates the diagnostic setting for the resource"), - // Operation: to.Ptr("Write diagnostic setting"), - // Provider: to.Ptr("Microsoft.DataFactory"), - // Resource: to.Ptr("factories"), - // }, - // Origin: to.Ptr("system"), - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/providers/Microsoft.Insights/metricDefinitions/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Gets the available metrics for factories"), - // Operation: to.Ptr("Read metric definitions"), - // Provider: to.Ptr("Microsoft.DataFactory"), - // Resource: to.Ptr("factories"), - // }, - // Origin: to.Ptr("system"), - // Properties: &armdatafactory.OperationProperties{ - // ServiceSpecification: &armdatafactory.OperationServiceSpecification{ - // MetricSpecifications: []*armdatafactory.OperationMetricSpecification{ - // { - // Name: to.Ptr("PipelineFailedRuns"), - // AggregationType: to.Ptr("Total"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("FailureType"), - // DisplayName: to.Ptr("Failure Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("Name"), - // DisplayName: to.Ptr("Pipeline"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Failed pipeline runs metrics"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Count"), - // }, - // { - // Name: to.Ptr("PipelineSucceededRuns"), - // AggregationType: to.Ptr("Total"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("FailureType"), - // DisplayName: to.Ptr("Failure Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("Name"), - // DisplayName: to.Ptr("Pipeline"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Succeeded pipeline runs metrics"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Count"), - // }, - // { - // Name: to.Ptr("ActivityFailedRuns"), - // AggregationType: to.Ptr("Total"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("ActivityType"), - // DisplayName: to.Ptr("Activity Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("PipelineName"), - // DisplayName: to.Ptr("Pipeline"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("FailureType"), - // DisplayName: to.Ptr("Failure Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("Name"), - // DisplayName: to.Ptr("Activity"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Failed activity runs metrics"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Count"), - // }, - // { - // Name: to.Ptr("ActivitySucceededRuns"), - // AggregationType: to.Ptr("Total"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("ActivityType"), - // DisplayName: to.Ptr("Activity Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("PipelineName"), - // DisplayName: to.Ptr("Pipeline"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("FailureType"), - // DisplayName: to.Ptr("Failure Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("Name"), - // DisplayName: to.Ptr("Activity"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Succeeded activity runs metrics"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Count"), - // }, - // { - // Name: to.Ptr("TriggerFailedRuns"), - // AggregationType: to.Ptr("Total"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("Name"), - // DisplayName: to.Ptr("Trigger"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("FailureType"), - // DisplayName: to.Ptr("Failure Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Failed trigger runs metrics"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Count"), - // }, - // { - // Name: to.Ptr("TriggerSucceededRuns"), - // AggregationType: to.Ptr("Total"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("Name"), - // DisplayName: to.Ptr("Trigger"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("FailureType"), - // DisplayName: to.Ptr("Failure Type"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Succeeded trigger runs metrics"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Count"), - // }, - // { - // Name: to.Ptr("IntegrationRuntimeCpuPercentage"), - // AggregationType: to.Ptr("Average"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("IntegrationRuntimeName"), - // DisplayName: to.Ptr("Integration Runtime"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("NodeName"), - // DisplayName: to.Ptr("Integration Runtime Node"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Integration runtime CPU utilization"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Percent"), - // }, - // { - // Name: to.Ptr("IntegrationRuntimeAvailableMemory"), - // AggregationType: to.Ptr("Average"), - // Dimensions: []*armdatafactory.OperationMetricDimension{ - // { - // Name: to.Ptr("IntegrationRuntimeName"), - // DisplayName: to.Ptr("Integration Runtime"), - // ToBeExportedForShoebox: to.Ptr(true), - // }, - // { - // Name: to.Ptr("NodeName"), - // DisplayName: to.Ptr("Integration Runtime Node"), - // ToBeExportedForShoebox: to.Ptr(true), - // }}, - // DisplayName: to.Ptr("Integration runtime available memory"), - // EnableRegionalMdmAccount: to.Ptr("false"), - // SourceMdmAccount: to.Ptr("MicrosoftDataFactoryProdShoebox"), - // SourceMdmNamespace: to.Ptr("ADFMetrics"), - // Unit: to.Ptr("Bytes"), - // }}, - // }, - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/register/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Registers the subscription for the Data Factory Resource Provider."), - // Operation: to.Ptr("Register Data Factory Resource Provider"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory Resource Provider"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/unregister/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Unregisters the subscription for the Data Factory Resource Provider."), - // Operation: to.Ptr("Unregister Data Factory Resource Provider"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory Resource Provider"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads the Data Factory."), - // Operation: to.Ptr("Read Data Factory"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates the Data Factory."), - // Operation: to.Ptr("Create or Update Data Factory"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes the Data Factory."), - // Operation: to.Ptr("Delete Data Factory"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/linkedServices/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Linked Service."), - // Operation: to.Ptr("Read Linked Service"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Linked Service"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/linkedServices/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Linked Service."), - // Operation: to.Ptr("Delete Linked Service"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Linked Service"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/linkedServices/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Linked Service."), - // Operation: to.Ptr("Create or Update Linked Service"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Linked Service"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Pipeline."), - // Operation: to.Ptr("Read Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Pipeline."), - // Operation: to.Ptr("Delete Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/pause/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Pauses any Pipeline."), - // Operation: to.Ptr("Pause Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/resume/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Resumes any Pipeline."), - // Operation: to.Ptr("Resume Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/update/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Updates any Pipeline."), - // Operation: to.Ptr("Update Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Pipeline."), - // Operation: to.Ptr("Create or Update Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datasets/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Dataset."), - // Operation: to.Ptr("Read Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datasets/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Dataset."), - // Operation: to.Ptr("Delete Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datasets/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Dataset."), - // Operation: to.Ptr("Create or Update Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/tables/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Dataset."), - // Operation: to.Ptr("Read Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/tables/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Dataset."), - // Operation: to.Ptr("Delete Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/tables/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Dataset."), - // Operation: to.Ptr("Create or Update Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datasets/slices/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Gets the Data Slices in the given period."), - // Operation: to.Ptr("Read Data Slices"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Slice"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datasets/slices/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Update the Status of the Data Slice."), - // Operation: to.Ptr("Update Data Slice Status"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Slice"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datasets/sliceruns/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads the Data Slice Run for the given dataset with the given start time."), - // Operation: to.Ptr("Read Data Slice Run"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Slice"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/runs/loginfo/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads a SAS URI to a blob container containing the logs."), - // Operation: to.Ptr("Read Run Log Info"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Run Log"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/activitywindows/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Activity Windows in the Data Factory with specified parameters."), - // Operation: to.Ptr("Read Activity Windows in any Data Factory"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Activity Windows"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/activitywindows/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Activity Windows for the Pipeline with specified parameters."), - // Operation: to.Ptr("Read Activity Windows for any Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Activity Windows"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datapipelines/activities/activitywindows/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Activity Windows for the Pipeline Activity with specified parameters."), - // Operation: to.Ptr("Read Activity Windows for any Pipeline Activity"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Activity Windows"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/datasets/activitywindows/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Activity Windows for the Dataset with specified parameters."), - // Operation: to.Ptr("Read Activity Windows for any Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Activity Windows"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/gateways/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Gateway."), - // Operation: to.Ptr("Read Gateway"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Gateway"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/gateways/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Gateway."), - // Operation: to.Ptr("Create or Update Gateway"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Gateway"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/gateways/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Gateway."), - // Operation: to.Ptr("Delete Gateway"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Gateway"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/gateways/connectioninfo/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads the Connection Info for any Gateway."), - // Operation: to.Ptr("Read Gateway Connection Info"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Gateway"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/gateways/listauthkeys/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Lists the Authentication Keys for any Gateway."), - // Operation: to.Ptr("List Gateway Authentication Keys"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Gateway"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/datafactories/gateways/regenerateauthkey/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Regenerates the Authentication Keys for any Gateway."), - // Operation: to.Ptr("Regenerate Gateway Authentication Keys"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Gateway"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Data Factory."), - // Operation: to.Ptr("Read Data Factory"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Create or Update Data Factory"), - // Operation: to.Ptr("Create or Update any Data Factory."), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes Data Factory."), - // Operation: to.Ptr("Delete Data Factory"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/linkedServices/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Linked Service."), - // Operation: to.Ptr("Read Linked Service"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Linked Service"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/linkedServices/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes Linked Service."), - // Operation: to.Ptr("Delete Linked Service"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Linked Service"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/linkedServices/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Create or Update Linked Service"), - // Operation: to.Ptr("Create or Update any Linked Service."), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Linked Service"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/pipelines/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Pipeline."), - // Operation: to.Ptr("Read Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/pipelines/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes Pipeline."), - // Operation: to.Ptr("Delete Pipeline"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/pipelines/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Create or Update Pipeline"), - // Operation: to.Ptr("Create or Update any Pipeline."), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/pipelines/createrun/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates a run for the Pipeline."), - // Operation: to.Ptr("Create Pipeline Run"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Pipeline"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/cancelpipelinerun/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Cancels the pipeline run specified by the run ID."), - // Operation: to.Ptr("Cancel any Pipeline Run"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/pipelineruns/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads the Pipeline Runs."), - // Operation: to.Ptr("Read Pipeline Runs"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/pipelineruns/activityruns/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads the activity runs for the specified pipeline run ID."), - // Operation: to.Ptr("Read Activity Runs"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/datasets/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Dataset."), - // Operation: to.Ptr("Read Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/datasets/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Dataset."), - // Operation: to.Ptr("Delete Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/datasets/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Dataset."), - // Operation: to.Ptr("Create or Update Dataset"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Dataset"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Integration Runtime."), - // Operation: to.Ptr("Read Integration Runtime"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Integration Runtime."), - // Operation: to.Ptr("Create or Update Integration Runtime"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Integration Runtime."), - // Operation: to.Ptr("Delete Integration Runtime"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/start/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Starts any Integration Runtime."), - // Operation: to.Ptr("Start Integration Runtime"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/stop/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Stops any Integration Runtime."), - // Operation: to.Ptr("Stop Integration Runtime"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/getconnectioninfo/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Integration Runtime Connection Info."), - // Operation: to.Ptr("Read Integration Runtime Connection Info"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/getstatus/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads Integration Runtime Status."), - // Operation: to.Ptr("Read Integration Runtime Status"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/listauthkeys/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Lists the Authentication Keys for any Integration Runtime."), - // Operation: to.Ptr("List Integration Runtime Authentication Keys"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/monitoringdata/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Gets the Monitoring Data for any Integration Runtime."), - // Operation: to.Ptr("Get Integration Runtime Monitoring Data"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/nodes/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes the Node for the specified Integration Runtime."), - // Operation: to.Ptr("Delete Integration Runtime Node"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/nodes/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Updates a self-hosted Integration Runtime Node."), - // Operation: to.Ptr("Update Integration Runtime Node"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/nodes/ipAddress/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Returns the IP Address for the specified node of the Integration Runtime."), - // Operation: to.Ptr("Read Integration Runtime Node IP Address"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/synccredentials/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Syncs the Credentials for the specified Integration Runtime."), - // Operation: to.Ptr("Sync Credentials for Integration Runtime"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/upgrade/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Upgrades the specified Integration Runtime."), - // Operation: to.Ptr("Upgrade Integration Runtime"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/integrationruntimes/regenerateauthkey/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Regenerates the Authentication Keys for the specified Integration Runtime."), - // Operation: to.Ptr("Regenerate Integration Runtime Authentication Keys"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Integration Runtime"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/triggers/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads any Trigger."), - // Operation: to.Ptr("Read Trigger"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Trigger"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/triggers/write"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Creates or Updates any Trigger."), - // Operation: to.Ptr("Create or Update Trigger"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Trigger"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/triggers/delete"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Deletes any Trigger."), - // Operation: to.Ptr("Delete Trigger"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Trigger"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/triggers/start/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Starts any Trigger."), - // Operation: to.Ptr("Start Trigger"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Trigger"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/triggers/stop/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Stops any Trigger."), - // Operation: to.Ptr("Stop Trigger"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Trigger"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/factories/triggers/triggerruns/read"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Reads the Trigger Runs."), - // Operation: to.Ptr("Read Trigger Runs"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }, - // { - // Name: to.Ptr("Microsoft.DataFactory/locations/configureFactoryRepo/action"), - // Display: &armdatafactory.OperationDisplay{ - // Description: to.Ptr("Configures the repository for the factory."), - // Operation: to.Ptr("Configure Factory Repository"), - // Provider: to.Ptr("Microsoft Data Factory"), - // Resource: to.Ptr("Data Factory"), - // }, - // }}, - // } - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/pipelineruns_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/pipelineruns_client_example_test.go deleted file mode 100644 index 04c0a9f3f2a6..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/pipelineruns_client_example_test.go +++ /dev/null @@ -1,158 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "time" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/PipelineRuns_QueryByFactory.json -func ExamplePipelineRunsClient_QueryByFactory() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPipelineRunsClient().QueryByFactory(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.RunFilterParameters{ - Filters: []*armdatafactory.RunQueryFilter{ - { - Operand: to.Ptr(armdatafactory.RunQueryFilterOperandPipelineName), - Operator: to.Ptr(armdatafactory.RunQueryFilterOperatorEquals), - Values: []*string{ - to.Ptr("examplePipeline")}, - }}, - LastUpdatedAfter: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:36:44.334Z"); return t }()), - LastUpdatedBefore: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:49:48.368Z"); return t }()), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PipelineRunsQueryResponse = armdatafactory.PipelineRunsQueryResponse{ - // Value: []*armdatafactory.PipelineRun{ - // { - // AdditionalProperties: map[string]any{ - // "annotations": []any{ - // }, - // "runDimension": map[string]any{ - // "JobId": "79c1cc52-265f-41a5-9553-be65e736fbd3", - // }, - // }, - // DurationInMs: to.Ptr[int32](28105), - // InvokedBy: &armdatafactory.PipelineRunInvokedBy{ - // Name: to.Ptr("Manual"), - // ID: to.Ptr("80a01654a9d34ad18b3fcac5d5d76b67"), - // }, - // LastUpdated: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:38:12.731Z"); return t}()), - // Message: to.Ptr(""), - // Parameters: map[string]*string{ - // "OutputBlobNameList": to.Ptr("[\"exampleoutput.csv\"]"), - // }, - // PipelineName: to.Ptr("examplePipeline"), - // RunEnd: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:38:12.731Z"); return t}()), - // RunID: to.Ptr("2f7fdb90-5df1-4b8e-ac2f-064cfa58202b"), - // RunStart: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:37:44.625Z"); return t}()), - // Status: to.Ptr("Succeeded"), - // }, - // { - // AdditionalProperties: map[string]any{ - // "annotations": []any{ - // }, - // "runDimension": map[string]any{ - // "JobId": "84a3c493-0628-4b44-852f-ef5b3a11bdab", - // }, - // }, - // InvokedBy: &armdatafactory.PipelineRunInvokedBy{ - // Name: to.Ptr("Manual"), - // ID: to.Ptr("7c5fd7ef7e8a464b98b931cf15fcac66"), - // }, - // LastUpdated: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:51.216Z"); return t}()), - // Message: to.Ptr(""), - // Parameters: map[string]*string{ - // "OutputBlobNameList": to.Ptr("[\"exampleoutput.csv\"]"), - // }, - // PipelineName: to.Ptr("examplePipeline"), - // RunID: to.Ptr("16ac5348-ff82-4f95-a80d-638c1d47b721"), - // RunStart: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:49.274Z"); return t}()), - // Status: to.Ptr("Cancelled"), - // }}, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/PipelineRuns_Get.json -func ExamplePipelineRunsClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPipelineRunsClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "2f7fdb90-5df1-4b8e-ac2f-064cfa58202b", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PipelineRun = armdatafactory.PipelineRun{ - // AdditionalProperties: map[string]any{ - // "annotations": []any{ - // }, - // }, - // DurationInMs: to.Ptr[int32](28105), - // InvokedBy: &armdatafactory.PipelineRunInvokedBy{ - // Name: to.Ptr("Manual"), - // ID: to.Ptr("80a01654a9d34ad18b3fcac5d5d76b67"), - // }, - // LastUpdated: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:38:12.731Z"); return t}()), - // Message: to.Ptr(""), - // Parameters: map[string]*string{ - // "OutputBlobNameList": to.Ptr("[\"exampleoutput.csv\"]"), - // }, - // PipelineName: to.Ptr("examplePipeline"), - // RunEnd: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:38:12.731Z"); return t}()), - // RunID: to.Ptr("2f7fdb90-5df1-4b8e-ac2f-064cfa58202b"), - // RunStart: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:37:44.625Z"); return t}()), - // Status: to.Ptr("Succeeded"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/PipelineRuns_Cancel.json -func ExamplePipelineRunsClient_Cancel() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewPipelineRunsClient().Cancel(ctx, "exampleResourceGroup", "exampleFactoryName", "16ac5348-ff82-4f95-a80d-638c1d47b721", &armdatafactory.PipelineRunsClientCancelOptions{IsRecursive: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/pipelines_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/pipelines_client_example_test.go deleted file mode 100644 index 78958739af42..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/pipelines_client_example_test.go +++ /dev/null @@ -1,558 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Pipelines_ListByFactory.json -func ExamplePipelinesClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewPipelinesClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.PipelineListResponse = armdatafactory.PipelineListResponse{ - // Value: []*armdatafactory.PipelineResource{ - // { - // Name: to.Ptr("examplePipeline"), - // Type: to.Ptr("Microsoft.DataFactory/factories/pipelines"), - // Etag: to.Ptr("0a006cd4-0000-0000-0000-5b245bd60000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/pipelines/examplePipeline"), - // Properties: &armdatafactory.Pipeline{ - // Description: to.Ptr("Example description"), - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.ForEachActivity{ - // Name: to.Ptr("ExampleForeachActivity"), - // Type: to.Ptr("ForEach"), - // TypeProperties: &armdatafactory.ForEachActivityTypeProperties{ - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.CopyActivity{ - // Name: to.Ptr("ExampleCopyActivity"), - // Type: to.Ptr("Copy"), - // Inputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": "examplecontainer.csv", - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // Outputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": map[string]any{ - // "type": "Expression", - // "value": "@item()", - // }, - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // TypeProperties: &armdatafactory.CopyActivityTypeProperties{ - // DataIntegrationUnits: float64(32), - // Sink: &armdatafactory.BlobSink{ - // Type: to.Ptr("BlobSink"), - // }, - // Source: &armdatafactory.BlobSource{ - // Type: to.Ptr("BlobSource"), - // }, - // }, - // }}, - // IsSequential: to.Ptr(true), - // Items: &armdatafactory.Expression{ - // Type: to.Ptr(armdatafactory.ExpressionTypeExpression), - // Value: to.Ptr("@pipeline().parameters.OutputBlobNameList"), - // }, - // }, - // }}, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "OutputBlobNameList": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeArray), - // }, - // }, - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Pipelines_Create.json -func ExamplePipelinesClient_CreateOrUpdate_pipelinesCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPipelinesClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "examplePipeline", armdatafactory.PipelineResource{ - Properties: &armdatafactory.Pipeline{ - Activities: []armdatafactory.ActivityClassification{ - &armdatafactory.ForEachActivity{ - Name: to.Ptr("ExampleForeachActivity"), - Type: to.Ptr("ForEach"), - TypeProperties: &armdatafactory.ForEachActivityTypeProperties{ - Activities: []armdatafactory.ActivityClassification{ - &armdatafactory.CopyActivity{ - Name: to.Ptr("ExampleCopyActivity"), - Type: to.Ptr("Copy"), - Inputs: []*armdatafactory.DatasetReference{ - { - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - Parameters: map[string]any{ - "MyFileName": "examplecontainer.csv", - "MyFolderPath": "examplecontainer", - }, - ReferenceName: to.Ptr("exampleDataset"), - }}, - Outputs: []*armdatafactory.DatasetReference{ - { - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - Parameters: map[string]any{ - "MyFileName": map[string]any{ - "type": "Expression", - "value": "@item()", - }, - "MyFolderPath": "examplecontainer", - }, - ReferenceName: to.Ptr("exampleDataset"), - }}, - TypeProperties: &armdatafactory.CopyActivityTypeProperties{ - DataIntegrationUnits: float64(32), - Sink: &armdatafactory.BlobSink{ - Type: to.Ptr("BlobSink"), - }, - Source: &armdatafactory.BlobSource{ - Type: to.Ptr("BlobSource"), - }, - }, - }}, - IsSequential: to.Ptr(true), - Items: &armdatafactory.Expression{ - Type: to.Ptr(armdatafactory.ExpressionTypeExpression), - Value: to.Ptr("@pipeline().parameters.OutputBlobNameList"), - }, - }, - }}, - Parameters: map[string]*armdatafactory.ParameterSpecification{ - "JobId": { - Type: to.Ptr(armdatafactory.ParameterTypeString), - }, - "OutputBlobNameList": { - Type: to.Ptr(armdatafactory.ParameterTypeArray), - }, - }, - Policy: &armdatafactory.PipelinePolicy{ - ElapsedTimeMetric: &armdatafactory.PipelineElapsedTimeMetricPolicy{ - Duration: "0.00:10:00", - }, - }, - RunDimensions: map[string]any{ - "JobId": map[string]any{ - "type": "Expression", - "value": "@pipeline().parameters.JobId", - }, - }, - Variables: map[string]*armdatafactory.VariableSpecification{ - "TestVariableArray": { - Type: to.Ptr(armdatafactory.VariableTypeArray), - }, - }, - }, - }, &armdatafactory.PipelinesClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PipelineResource = armdatafactory.PipelineResource{ - // Name: to.Ptr("examplePipeline"), - // Type: to.Ptr("Microsoft.DataFactory/factories/pipelines"), - // Etag: to.Ptr("0a0069d4-0000-0000-0000-5b245bd50000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/pipelines/examplePipeline"), - // Properties: &armdatafactory.Pipeline{ - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.ForEachActivity{ - // Name: to.Ptr("ExampleForeachActivity"), - // Type: to.Ptr("ForEach"), - // TypeProperties: &armdatafactory.ForEachActivityTypeProperties{ - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.CopyActivity{ - // Name: to.Ptr("ExampleCopyActivity"), - // Type: to.Ptr("Copy"), - // Inputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": "examplecontainer.csv", - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // Outputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": map[string]any{ - // "type": "Expression", - // "value": "@item()", - // }, - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // TypeProperties: &armdatafactory.CopyActivityTypeProperties{ - // DataIntegrationUnits: float64(32), - // Sink: &armdatafactory.BlobSink{ - // Type: to.Ptr("BlobSink"), - // }, - // Source: &armdatafactory.BlobSource{ - // Type: to.Ptr("BlobSource"), - // }, - // }, - // }}, - // IsSequential: to.Ptr(true), - // Items: &armdatafactory.Expression{ - // Type: to.Ptr(armdatafactory.ExpressionTypeExpression), - // Value: to.Ptr("@pipeline().parameters.OutputBlobNameList"), - // }, - // }, - // }}, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "JobId": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeString), - // }, - // "OutputBlobNameList": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeArray), - // }, - // }, - // RunDimensions: map[string]any{ - // "JobId": map[string]any{ - // "type": "Expression", - // "value": "@pipeline().parameters.JobId", - // }, - // }, - // Variables: map[string]*armdatafactory.VariableSpecification{ - // "TestVariableArray": &armdatafactory.VariableSpecification{ - // Type: to.Ptr(armdatafactory.VariableTypeArray), - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Pipelines_Update.json -func ExamplePipelinesClient_CreateOrUpdate_pipelinesUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPipelinesClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "examplePipeline", armdatafactory.PipelineResource{ - Properties: &armdatafactory.Pipeline{ - Description: to.Ptr("Example description"), - Activities: []armdatafactory.ActivityClassification{ - &armdatafactory.ForEachActivity{ - Name: to.Ptr("ExampleForeachActivity"), - Type: to.Ptr("ForEach"), - TypeProperties: &armdatafactory.ForEachActivityTypeProperties{ - Activities: []armdatafactory.ActivityClassification{ - &armdatafactory.CopyActivity{ - Name: to.Ptr("ExampleCopyActivity"), - Type: to.Ptr("Copy"), - Inputs: []*armdatafactory.DatasetReference{ - { - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - Parameters: map[string]any{ - "MyFileName": "examplecontainer.csv", - "MyFolderPath": "examplecontainer", - }, - ReferenceName: to.Ptr("exampleDataset"), - }}, - Outputs: []*armdatafactory.DatasetReference{ - { - Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - Parameters: map[string]any{ - "MyFileName": map[string]any{ - "type": "Expression", - "value": "@item()", - }, - "MyFolderPath": "examplecontainer", - }, - ReferenceName: to.Ptr("exampleDataset"), - }}, - TypeProperties: &armdatafactory.CopyActivityTypeProperties{ - DataIntegrationUnits: float64(32), - Sink: &armdatafactory.BlobSink{ - Type: to.Ptr("BlobSink"), - }, - Source: &armdatafactory.BlobSource{ - Type: to.Ptr("BlobSource"), - }, - }, - }}, - IsSequential: to.Ptr(true), - Items: &armdatafactory.Expression{ - Type: to.Ptr(armdatafactory.ExpressionTypeExpression), - Value: to.Ptr("@pipeline().parameters.OutputBlobNameList"), - }, - }, - }}, - Parameters: map[string]*armdatafactory.ParameterSpecification{ - "OutputBlobNameList": { - Type: to.Ptr(armdatafactory.ParameterTypeArray), - }, - }, - Policy: &armdatafactory.PipelinePolicy{ - ElapsedTimeMetric: &armdatafactory.PipelineElapsedTimeMetricPolicy{ - Duration: "0.00:10:00", - }, - }, - }, - }, &armdatafactory.PipelinesClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PipelineResource = armdatafactory.PipelineResource{ - // Name: to.Ptr("examplePipeline"), - // Type: to.Ptr("Microsoft.DataFactory/factories/pipelines"), - // Etag: to.Ptr("0a006cd4-0000-0000-0000-5b245bd60000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/pipelines/examplePipeline"), - // Properties: &armdatafactory.Pipeline{ - // Description: to.Ptr("Example description"), - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.ForEachActivity{ - // Name: to.Ptr("ExampleForeachActivity"), - // Type: to.Ptr("ForEach"), - // TypeProperties: &armdatafactory.ForEachActivityTypeProperties{ - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.CopyActivity{ - // Name: to.Ptr("ExampleCopyActivity"), - // Type: to.Ptr("Copy"), - // Inputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": "examplecontainer.csv", - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // Outputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": map[string]any{ - // "type": "Expression", - // "value": "@item()", - // }, - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // TypeProperties: &armdatafactory.CopyActivityTypeProperties{ - // DataIntegrationUnits: float64(32), - // Sink: &armdatafactory.BlobSink{ - // Type: to.Ptr("BlobSink"), - // }, - // Source: &armdatafactory.BlobSource{ - // Type: to.Ptr("BlobSource"), - // }, - // }, - // }}, - // IsSequential: to.Ptr(true), - // Items: &armdatafactory.Expression{ - // Type: to.Ptr(armdatafactory.ExpressionTypeExpression), - // Value: to.Ptr("@pipeline().parameters.OutputBlobNameList"), - // }, - // }, - // }}, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "OutputBlobNameList": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeArray), - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Pipelines_Get.json -func ExamplePipelinesClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPipelinesClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "examplePipeline", &armdatafactory.PipelinesClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PipelineResource = armdatafactory.PipelineResource{ - // Name: to.Ptr("examplePipeline"), - // Type: to.Ptr("Microsoft.DataFactory/factories/pipelines"), - // Etag: to.Ptr("1500504f-0000-0200-0000-5cbe090f0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/pipelines/examplePipeline"), - // Properties: &armdatafactory.Pipeline{ - // Description: to.Ptr("Example description"), - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.ForEachActivity{ - // Name: to.Ptr("ExampleForeachActivity"), - // Type: to.Ptr("ForEach"), - // TypeProperties: &armdatafactory.ForEachActivityTypeProperties{ - // Activities: []armdatafactory.ActivityClassification{ - // &armdatafactory.CopyActivity{ - // Name: to.Ptr("ExampleCopyActivity"), - // Type: to.Ptr("Copy"), - // Inputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": "examplecontainer.csv", - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // Outputs: []*armdatafactory.DatasetReference{ - // { - // Type: to.Ptr(armdatafactory.DatasetReferenceTypeDatasetReference), - // Parameters: map[string]any{ - // "MyFileName": map[string]any{ - // "type": "Expression", - // "value": "@item()", - // }, - // "MyFolderPath": "examplecontainer", - // }, - // ReferenceName: to.Ptr("exampleDataset"), - // }}, - // TypeProperties: &armdatafactory.CopyActivityTypeProperties{ - // DataIntegrationUnits: float64(32), - // Sink: &armdatafactory.BlobSink{ - // Type: to.Ptr("BlobSink"), - // }, - // Source: &armdatafactory.BlobSource{ - // Type: to.Ptr("BlobSource"), - // }, - // }, - // }}, - // IsSequential: to.Ptr(true), - // Items: &armdatafactory.Expression{ - // Type: to.Ptr(armdatafactory.ExpressionTypeExpression), - // Value: to.Ptr("@pipeline().parameters.OutputBlobNameList"), - // }, - // }, - // }}, - // Parameters: map[string]*armdatafactory.ParameterSpecification{ - // "OutputBlobNameList": &armdatafactory.ParameterSpecification{ - // Type: to.Ptr(armdatafactory.ParameterTypeArray), - // }, - // }, - // Policy: &armdatafactory.PipelinePolicy{ - // ElapsedTimeMetric: &armdatafactory.PipelineElapsedTimeMetricPolicy{ - // Duration: "0.00:10:00", - // }, - // }, - // Variables: map[string]*armdatafactory.VariableSpecification{ - // "TestVariableArray": &armdatafactory.VariableSpecification{ - // Type: to.Ptr(armdatafactory.VariableTypeArray), - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Pipelines_Delete.json -func ExamplePipelinesClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewPipelinesClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "examplePipeline", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Pipelines_CreateRun.json -func ExamplePipelinesClient_CreateRun() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPipelinesClient().CreateRun(ctx, "exampleResourceGroup", "exampleFactoryName", "examplePipeline", &armdatafactory.PipelinesClientCreateRunOptions{ReferencePipelineRunID: nil, - IsRecovery: nil, - StartActivityName: nil, - StartFromFailure: nil, - Parameters: map[string]any{ - "OutputBlobNameList": []any{ - "exampleoutput.csv", - }, - }, - }) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.CreateRunResponse = armdatafactory.CreateRunResponse{ - // RunID: to.Ptr("2f7fdb90-5df1-4b8e-ac2f-064cfa58202b"), - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/polymorphic_helpers.go b/sdk/resourcemanager/datafactory/armdatafactory/polymorphic_helpers.go index 2a937742dc25..23b1f3fa341a 100644 --- a/sdk/resourcemanager/datafactory/armdatafactory/polymorphic_helpers.go +++ b/sdk/resourcemanager/datafactory/armdatafactory/polymorphic_helpers.go @@ -231,12 +231,18 @@ func unmarshalCopySinkClassification(rawMsg json.RawMessage) (CopySinkClassifica b = &RestSink{} case "SalesforceServiceCloudSink": b = &SalesforceServiceCloudSink{} + case "SalesforceServiceCloudV2Sink": + b = &SalesforceServiceCloudV2Sink{} case "SalesforceSink": b = &SalesforceSink{} + case "SalesforceV2Sink": + b = &SalesforceV2Sink{} case "SapCloudForCustomerSink": b = &SapCloudForCustomerSink{} case "SnowflakeSink": b = &SnowflakeSink{} + case "SnowflakeV2Sink": + b = &SnowflakeV2Sink{} case "SqlDWSink": b = &SQLDWSink{} case "SqlMISink": @@ -245,6 +251,8 @@ func unmarshalCopySinkClassification(rawMsg json.RawMessage) (CopySinkClassifica b = &SQLServerSink{} case "SqlSink": b = &SQLSink{} + case "WarehouseSink": + b = &WarehouseSink{} default: b = &CopySink{} } @@ -406,8 +414,12 @@ func unmarshalCopySourceClassification(rawMsg json.RawMessage) (CopySourceClassi b = &SalesforceMarketingCloudSource{} case "SalesforceServiceCloudSource": b = &SalesforceServiceCloudSource{} + case "SalesforceServiceCloudV2Source": + b = &SalesforceServiceCloudV2Source{} case "SalesforceSource": b = &SalesforceSource{} + case "SalesforceV2Source": + b = &SalesforceV2Source{} case "SapBwSource": b = &SapBwSource{} case "SapCloudForCustomerSource": @@ -430,6 +442,8 @@ func unmarshalCopySourceClassification(rawMsg json.RawMessage) (CopySourceClassi b = &ShopifySource{} case "SnowflakeSource": b = &SnowflakeSource{} + case "SnowflakeV2Source": + b = &SnowflakeV2Source{} case "SparkSource": b = &SparkSource{} case "SqlDWSource": @@ -450,6 +464,8 @@ func unmarshalCopySourceClassification(rawMsg json.RawMessage) (CopySourceClassi b = &TeradataSource{} case "VerticaSource": b = &VerticaSource{} + case "WarehouseSource": + b = &WarehouseSource{} case "WebSource": b = &WebSource{} case "XeroSource": @@ -700,6 +716,10 @@ func unmarshalDatasetClassification(rawMsg json.RawMessage) (DatasetClassificati b = &SalesforceObjectDataset{} case "SalesforceServiceCloudObject": b = &SalesforceServiceCloudObjectDataset{} + case "SalesforceServiceCloudV2Object": + b = &SalesforceServiceCloudV2ObjectDataset{} + case "SalesforceV2Object": + b = &SalesforceV2ObjectDataset{} case "SapBwCube": b = &SapBwCubeDataset{} case "SapCloudForCustomerResource": @@ -722,6 +742,8 @@ func unmarshalDatasetClassification(rawMsg json.RawMessage) (DatasetClassificati b = &ShopifyObjectDataset{} case "SnowflakeTable": b = &SnowflakeDataset{} + case "SnowflakeV2Table": + b = &SnowflakeV2Dataset{} case "SparkObject": b = &SparkObjectDataset{} case "SqlServerTable": @@ -734,6 +756,8 @@ func unmarshalDatasetClassification(rawMsg json.RawMessage) (DatasetClassificati b = &TeradataTableDataset{} case "VerticaTable": b = &VerticaTableDataset{} + case "WarehouseTable": + b = &WarehouseTableDataset{} case "WebTable": b = &WebTableDataset{} case "XeroObject": @@ -1180,6 +1204,10 @@ func unmarshalLinkedServiceClassification(rawMsg json.RawMessage) (LinkedService b = &SalesforceMarketingCloudLinkedService{} case "SalesforceServiceCloud": b = &SalesforceServiceCloudLinkedService{} + case "SalesforceServiceCloudV2": + b = &SalesforceServiceCloudV2LinkedService{} + case "SalesforceV2": + b = &SalesforceV2LinkedService{} case "SapBW": b = &SapBWLinkedService{} case "SapCloudForCustomer": @@ -1206,6 +1234,8 @@ func unmarshalLinkedServiceClassification(rawMsg json.RawMessage) (LinkedService b = &SmartsheetLinkedService{} case "Snowflake": b = &SnowflakeLinkedService{} + case "SnowflakeV2": + b = &SnowflakeV2LinkedService{} case "Spark": b = &SparkLinkedService{} case "SqlServer": @@ -1222,6 +1252,8 @@ func unmarshalLinkedServiceClassification(rawMsg json.RawMessage) (LinkedService b = &TwilioLinkedService{} case "Vertica": b = &VerticaLinkedService{} + case "Warehouse": + b = &WarehouseLinkedService{} case "Web": b = &WebLinkedService{} case "Xero": diff --git a/sdk/resourcemanager/datafactory/armdatafactory/privateendpointconnection_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/privateendpointconnection_client_example_test.go deleted file mode 100644 index 49c1b5161b1d..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/privateendpointconnection_client_example_test.go +++ /dev/null @@ -1,119 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/ApproveRejectPrivateEndpointConnection.json -func ExamplePrivateEndpointConnectionClient_CreateOrUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPrivateEndpointConnectionClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "connection", armdatafactory.PrivateLinkConnectionApprovalRequestResource{ - Properties: &armdatafactory.PrivateLinkConnectionApprovalRequest{ - PrivateEndpoint: &armdatafactory.PrivateEndpoint{ - ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/privateEndpoints/myPrivateEndpoint"), - }, - PrivateLinkServiceConnectionState: &armdatafactory.PrivateLinkConnectionState{ - Description: to.Ptr("Approved by admin."), - ActionsRequired: to.Ptr(""), - Status: to.Ptr("Approved"), - }, - }, - }, &armdatafactory.PrivateEndpointConnectionClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PrivateEndpointConnectionResource = armdatafactory.PrivateEndpointConnectionResource{ - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/privateEndpointConnections"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Properties: &armdatafactory.RemotePrivateEndpointConnection{ - // PrivateEndpoint: &armdatafactory.ArmIDWrapper{ - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/privateEndpoints/myPrivateEndpoint"), - // }, - // PrivateLinkServiceConnectionState: &armdatafactory.PrivateLinkConnectionState{ - // Description: to.Ptr("Approved by admin."), - // ActionsRequired: to.Ptr(""), - // Status: to.Ptr("Approved"), - // }, - // ProvisioningState: to.Ptr("Succeeded"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/GetPrivateEndpointConnection.json -func ExamplePrivateEndpointConnectionClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPrivateEndpointConnectionClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "connection", &armdatafactory.PrivateEndpointConnectionClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PrivateEndpointConnectionResource = armdatafactory.PrivateEndpointConnectionResource{ - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/privateEndpointConnections"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Properties: &armdatafactory.RemotePrivateEndpointConnection{ - // PrivateEndpoint: &armdatafactory.ArmIDWrapper{ - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/privateEndpoints/myPrivateEndpoint"), - // }, - // PrivateLinkServiceConnectionState: &armdatafactory.PrivateLinkConnectionState{ - // Description: to.Ptr("Approved by admin."), - // ActionsRequired: to.Ptr(""), - // Status: to.Ptr("Approved"), - // }, - // ProvisioningState: to.Ptr("Succeeded"), - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/DeletePrivateEndpointConnection.json -func ExamplePrivateEndpointConnectionClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewPrivateEndpointConnectionClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "connection", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/privateendpointconnections_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/privateendpointconnections_client_example_test.go deleted file mode 100644 index 1c322a2bd948..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/privateendpointconnections_client_example_test.go +++ /dev/null @@ -1,62 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/PrivateEndPointConnections_ListByFactory.json -func ExamplePrivateEndPointConnectionsClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewPrivateEndPointConnectionsClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.PrivateEndpointConnectionListResponse = armdatafactory.PrivateEndpointConnectionListResponse{ - // Value: []*armdatafactory.PrivateEndpointConnectionResource{ - // { - // Name: to.Ptr("factories"), - // Type: to.Ptr("Microsoft.DataFactory/factories/privateEndpointConnections"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/privateEndpoints/myPrivateEndpoint"), - // Properties: &armdatafactory.RemotePrivateEndpointConnection{ - // PrivateEndpoint: &armdatafactory.ArmIDWrapper{ - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/managedVirtualNetworks/myPrivateEndpoint"), - // }, - // PrivateLinkServiceConnectionState: &armdatafactory.PrivateLinkConnectionState{ - // Description: to.Ptr("Approved by admin."), - // ActionsRequired: to.Ptr("exampleActionsRequired"), - // Status: to.Ptr("Approved"), - // }, - // ProvisioningState: to.Ptr("Succeeded"), - // }, - // }}, - // } - } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/privatelinkresources_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/privatelinkresources_client_example_test.go deleted file mode 100644 index 2212eb05901e..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/privatelinkresources_client_example_test.go +++ /dev/null @@ -1,53 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/GetPrivateLinkResources.json -func ExamplePrivateLinkResourcesClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewPrivateLinkResourcesClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.PrivateLinkResourcesWrapper = armdatafactory.PrivateLinkResourcesWrapper{ - // Value: []*armdatafactory.PrivateLinkResource{ - // { - // Name: to.Ptr("exampleFactoryName"), - // Type: to.Ptr("Microsoft.DataFactory/factories/privateLinkResources"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName"), - // Properties: &armdatafactory.PrivateLinkResourceProperties{ - // GroupID: to.Ptr("dataFactory"), - // RequiredMembers: []*string{ - // to.Ptr("dataFactory")}, - // RequiredZoneNames: []*string{ - // to.Ptr("privatelink.datafactory.azure.net")}, - // }, - // }}, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/triggerruns_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/triggerruns_client_example_test.go deleted file mode 100644 index 8771f22c15b3..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/triggerruns_client_example_test.go +++ /dev/null @@ -1,103 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "time" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/TriggerRuns_Rerun.json -func ExampleTriggerRunsClient_Rerun() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewTriggerRunsClient().Rerun(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", "2f7fdb90-5df1-4b8e-ac2f-064cfa58202b", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/TriggerRuns_Cancel.json -func ExampleTriggerRunsClient_Cancel() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewTriggerRunsClient().Cancel(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", "2f7fdb90-5df1-4b8e-ac2f-064cfa58202b", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/TriggerRuns_QueryByFactory.json -func ExampleTriggerRunsClient_QueryByFactory() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewTriggerRunsClient().QueryByFactory(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.RunFilterParameters{ - Filters: []*armdatafactory.RunQueryFilter{ - { - Operand: to.Ptr(armdatafactory.RunQueryFilterOperandTriggerName), - Operator: to.Ptr(armdatafactory.RunQueryFilterOperatorEquals), - Values: []*string{ - to.Ptr("exampleTrigger")}, - }}, - LastUpdatedAfter: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:36:44.334Z"); return t }()), - LastUpdatedBefore: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:49:48.368Z"); return t }()), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerRunsQueryResponse = armdatafactory.TriggerRunsQueryResponse{ - // Value: []*armdatafactory.TriggerRun{ - // { - // Message: to.Ptr(""), - // Properties: map[string]*string{ - // "ScheduleTime": to.Ptr("6/16/2018 12:43:14 AM"), - // "TriggerTime": to.Ptr("6/16/2018 12:43:15 AM"), - // }, - // Status: to.Ptr(armdatafactory.TriggerRunStatusSucceeded), - // TriggerName: to.Ptr("exampleTrigger"), - // TriggerRunID: to.Ptr("08586724970898148904457116912CU27"), - // TriggerRunTimestamp: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:43:15.660Z"); return t}()), - // TriggerType: to.Ptr("ScheduleTrigger"), - // TriggeredPipelines: map[string]*string{ - // "examplePipeline": to.Ptr("9f3ce8b3-37d7-43eb-96ac-a656c0476283"), - // }, - // }}, - // } -} diff --git a/sdk/resourcemanager/datafactory/armdatafactory/triggers_client_example_test.go b/sdk/resourcemanager/datafactory/armdatafactory/triggers_client_example_test.go deleted file mode 100644 index 3d43c2bb7266..000000000000 --- a/sdk/resourcemanager/datafactory/armdatafactory/triggers_client_example_test.go +++ /dev/null @@ -1,465 +0,0 @@ -//go:build go1.18 -// +build go1.18 - -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. See License.txt in the project root for license information. -// Code generated by Microsoft (R) AutoRest Code Generator. -// Changes may cause incorrect behavior and will be lost if the code is regenerated. -// DO NOT EDIT. - -package armdatafactory_test - -import ( - "context" - "log" - - "time" - - "github.com/Azure/azure-sdk-for-go/sdk/azcore/to" - "github.com/Azure/azure-sdk-for-go/sdk/azidentity" - "github.com/Azure/azure-sdk-for-go/sdk/resourcemanager/datafactory/armdatafactory/v4" -) - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_ListByFactory.json -func ExampleTriggersClient_NewListByFactoryPager() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - pager := clientFactory.NewTriggersClient().NewListByFactoryPager("exampleResourceGroup", "exampleFactoryName", nil) - for pager.More() { - page, err := pager.NextPage(ctx) - if err != nil { - log.Fatalf("failed to advance page: %v", err) - } - for _, v := range page.Value { - // You could use page here. We use blank identifier for just demo purposes. - _ = v - } - // If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // page.TriggerListResponse = armdatafactory.TriggerListResponse{ - // Value: []*armdatafactory.TriggerResource{ - // { - // Name: to.Ptr("exampleTrigger"), - // Type: to.Ptr("Microsoft.DataFactory/factories/triggers"), - // Etag: to.Ptr("0a008ed4-0000-0000-0000-5b245c740000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/triggers/exampleTrigger"), - // Properties: &armdatafactory.ScheduleTrigger{ - // Type: to.Ptr("ScheduleTrigger"), - // Description: to.Ptr("Example description"), - // RuntimeState: to.Ptr(armdatafactory.TriggerRuntimeStateStarted), - // Pipelines: []*armdatafactory.TriggerPipelineReference{ - // { - // Parameters: map[string]any{ - // "OutputBlobNameList": []any{ - // "exampleoutput.csv", - // }, - // }, - // PipelineReference: &armdatafactory.PipelineReference{ - // Type: to.Ptr(armdatafactory.PipelineReferenceTypePipelineReference), - // ReferenceName: to.Ptr("examplePipeline"), - // }, - // }}, - // TypeProperties: &armdatafactory.ScheduleTriggerTypeProperties{ - // Recurrence: &armdatafactory.ScheduleTriggerRecurrence{ - // EndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:55:14.905Z"); return t}()), - // Frequency: to.Ptr(armdatafactory.RecurrenceFrequencyMinute), - // Interval: to.Ptr[int32](4), - // StartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:14.905Z"); return t}()), - // TimeZone: to.Ptr("UTC"), - // }, - // }, - // }, - // }}, - // } - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_QueryByFactory.json -func ExampleTriggersClient_QueryByFactory() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewTriggersClient().QueryByFactory(ctx, "exampleResourceGroup", "exampleFactoryName", armdatafactory.TriggerFilterParameters{ - ParentTriggerName: to.Ptr("exampleTrigger"), - }, nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerQueryResponse = armdatafactory.TriggerQueryResponse{ - // Value: []*armdatafactory.TriggerResource{ - // { - // Name: to.Ptr("exampleRerunTrigger"), - // Type: to.Ptr("Microsoft.DataFactory/factories/triggers"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/triggers/exampleRerunTrigger"), - // Properties: &armdatafactory.RerunTumblingWindowTrigger{ - // Type: to.Ptr("RerunTumblingWindowTrigger"), - // Description: to.Ptr("Example description"), - // TypeProperties: &armdatafactory.RerunTumblingWindowTriggerTypeProperties{ - // ParentTrigger: "exampleTrigger", - // RequestedEndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:55:14.905Z"); return t}()), - // RequestedStartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:14.905Z"); return t}()), - // RerunConcurrency: to.Ptr[int32](4), - // }, - // }, - // }}, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_Create.json -func ExampleTriggersClient_CreateOrUpdate_triggersCreate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewTriggersClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", armdatafactory.TriggerResource{ - Properties: &armdatafactory.ScheduleTrigger{ - Type: to.Ptr("ScheduleTrigger"), - Pipelines: []*armdatafactory.TriggerPipelineReference{ - { - Parameters: map[string]any{ - "OutputBlobNameList": []any{ - "exampleoutput.csv", - }, - }, - PipelineReference: &armdatafactory.PipelineReference{ - Type: to.Ptr(armdatafactory.PipelineReferenceTypePipelineReference), - ReferenceName: to.Ptr("examplePipeline"), - }, - }}, - TypeProperties: &armdatafactory.ScheduleTriggerTypeProperties{ - Recurrence: &armdatafactory.ScheduleTriggerRecurrence{ - EndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:55:13.844Z"); return t }()), - Frequency: to.Ptr(armdatafactory.RecurrenceFrequencyMinute), - Interval: to.Ptr[int32](4), - StartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:13.844Z"); return t }()), - TimeZone: to.Ptr("UTC"), - }, - }, - }, - }, &armdatafactory.TriggersClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerResource = armdatafactory.TriggerResource{ - // Name: to.Ptr("exampleTrigger"), - // Type: to.Ptr("Microsoft.DataFactory/factories/triggers"), - // Etag: to.Ptr("0a008ad4-0000-0000-0000-5b245c6e0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/triggers/exampleTrigger"), - // Properties: &armdatafactory.ScheduleTrigger{ - // Type: to.Ptr("ScheduleTrigger"), - // RuntimeState: to.Ptr(armdatafactory.TriggerRuntimeStateStopped), - // Pipelines: []*armdatafactory.TriggerPipelineReference{ - // { - // Parameters: map[string]any{ - // "OutputBlobNameList": []any{ - // "exampleoutput.csv", - // }, - // }, - // PipelineReference: &armdatafactory.PipelineReference{ - // Type: to.Ptr(armdatafactory.PipelineReferenceTypePipelineReference), - // ReferenceName: to.Ptr("examplePipeline"), - // }, - // }}, - // TypeProperties: &armdatafactory.ScheduleTriggerTypeProperties{ - // Recurrence: &armdatafactory.ScheduleTriggerRecurrence{ - // EndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:55:13.844Z"); return t}()), - // Frequency: to.Ptr(armdatafactory.RecurrenceFrequencyMinute), - // Interval: to.Ptr[int32](4), - // StartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:13.844Z"); return t}()), - // TimeZone: to.Ptr("UTC"), - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_Update.json -func ExampleTriggersClient_CreateOrUpdate_triggersUpdate() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewTriggersClient().CreateOrUpdate(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", armdatafactory.TriggerResource{ - Properties: &armdatafactory.ScheduleTrigger{ - Type: to.Ptr("ScheduleTrigger"), - Description: to.Ptr("Example description"), - Pipelines: []*armdatafactory.TriggerPipelineReference{ - { - Parameters: map[string]any{ - "OutputBlobNameList": []any{ - "exampleoutput.csv", - }, - }, - PipelineReference: &armdatafactory.PipelineReference{ - Type: to.Ptr(armdatafactory.PipelineReferenceTypePipelineReference), - ReferenceName: to.Ptr("examplePipeline"), - }, - }}, - TypeProperties: &armdatafactory.ScheduleTriggerTypeProperties{ - Recurrence: &armdatafactory.ScheduleTriggerRecurrence{ - EndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:55:14.905Z"); return t }()), - Frequency: to.Ptr(armdatafactory.RecurrenceFrequencyMinute), - Interval: to.Ptr[int32](4), - StartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:14.905Z"); return t }()), - TimeZone: to.Ptr("UTC"), - }, - }, - }, - }, &armdatafactory.TriggersClientCreateOrUpdateOptions{IfMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerResource = armdatafactory.TriggerResource{ - // Name: to.Ptr("exampleTrigger"), - // Type: to.Ptr("Microsoft.DataFactory/factories/triggers"), - // Etag: to.Ptr("0a008dd4-0000-0000-0000-5b245c6f0000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/triggers/exampleTrigger"), - // Properties: &armdatafactory.ScheduleTrigger{ - // Type: to.Ptr("ScheduleTrigger"), - // Description: to.Ptr("Example description"), - // RuntimeState: to.Ptr(armdatafactory.TriggerRuntimeStateStopped), - // Pipelines: []*armdatafactory.TriggerPipelineReference{ - // { - // Parameters: map[string]any{ - // "OutputBlobNameList": []any{ - // "exampleoutput.csv", - // }, - // }, - // PipelineReference: &armdatafactory.PipelineReference{ - // Type: to.Ptr(armdatafactory.PipelineReferenceTypePipelineReference), - // ReferenceName: to.Ptr("examplePipeline"), - // }, - // }}, - // TypeProperties: &armdatafactory.ScheduleTriggerTypeProperties{ - // Recurrence: &armdatafactory.ScheduleTriggerRecurrence{ - // EndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:55:14.905Z"); return t}()), - // Frequency: to.Ptr(armdatafactory.RecurrenceFrequencyMinute), - // Interval: to.Ptr[int32](4), - // StartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2018-06-16T00:39:14.905Z"); return t}()), - // TimeZone: to.Ptr("UTC"), - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_Get.json -func ExampleTriggersClient_Get() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewTriggersClient().Get(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", &armdatafactory.TriggersClientGetOptions{IfNoneMatch: nil}) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerResource = armdatafactory.TriggerResource{ - // Name: to.Ptr("exampleTrigger"), - // Type: to.Ptr("Microsoft.DataFactory/factories/triggers"), - // Etag: to.Ptr("1500544f-0000-0200-0000-5cbe09100000"), - // ID: to.Ptr("/subscriptions/12345678-1234-1234-1234-12345678abc/resourceGroups/exampleResourceGroup/providers/Microsoft.DataFactory/factories/exampleFactoryName/triggers/exampleTrigger"), - // Properties: &armdatafactory.ScheduleTrigger{ - // Type: to.Ptr("ScheduleTrigger"), - // RuntimeState: to.Ptr(armdatafactory.TriggerRuntimeStateStopped), - // Pipelines: []*armdatafactory.TriggerPipelineReference{ - // { - // Parameters: map[string]any{ - // "OutputBlobNameList": []any{ - // "exampleoutput.csv", - // }, - // }, - // PipelineReference: &armdatafactory.PipelineReference{ - // Type: to.Ptr(armdatafactory.PipelineReferenceTypePipelineReference), - // ReferenceName: to.Ptr("examplePipeline"), - // }, - // }}, - // TypeProperties: &armdatafactory.ScheduleTriggerTypeProperties{ - // Recurrence: &armdatafactory.ScheduleTriggerRecurrence{ - // EndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2019-04-22T18:48:52.528Z"); return t}()), - // Frequency: to.Ptr(armdatafactory.RecurrenceFrequencyMinute), - // Interval: to.Ptr[int32](4), - // StartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, "2019-04-22T18:32:52.527Z"); return t}()), - // TimeZone: to.Ptr("UTC"), - // }, - // }, - // }, - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_Delete.json -func ExampleTriggersClient_Delete() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - _, err = clientFactory.NewTriggersClient().Delete(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_SubscribeToEvents.json -func ExampleTriggersClient_BeginSubscribeToEvents() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewTriggersClient().BeginSubscribeToEvents(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - res, err := poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerSubscriptionOperationStatus = armdatafactory.TriggerSubscriptionOperationStatus{ - // Status: to.Ptr(armdatafactory.EventSubscriptionStatusEnabled), - // TriggerName: to.Ptr("exampleTrigger"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_GetEventSubscriptionStatus.json -func ExampleTriggersClient_GetEventSubscriptionStatus() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - res, err := clientFactory.NewTriggersClient().GetEventSubscriptionStatus(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerSubscriptionOperationStatus = armdatafactory.TriggerSubscriptionOperationStatus{ - // Status: to.Ptr(armdatafactory.EventSubscriptionStatusEnabled), - // TriggerName: to.Ptr("exampleTrigger"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_UnsubscribeFromEvents.json -func ExampleTriggersClient_BeginUnsubscribeFromEvents() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewTriggersClient().BeginUnsubscribeFromEvents(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - res, err := poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } - // You could use response here. We use blank identifier for just demo purposes. - _ = res - // If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes. - // res.TriggerSubscriptionOperationStatus = armdatafactory.TriggerSubscriptionOperationStatus{ - // Status: to.Ptr(armdatafactory.EventSubscriptionStatusDisabled), - // TriggerName: to.Ptr("exampleTrigger"), - // } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_Start.json -func ExampleTriggersClient_BeginStart() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewTriggersClient().BeginStart(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - _, err = poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } -} - -// Generated from example definition: https://github.com/Azure/azure-rest-api-specs/blob/69ece3818b8b0929b43a07c3fe25716427734882/specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Triggers_Stop.json -func ExampleTriggersClient_BeginStop() { - cred, err := azidentity.NewDefaultAzureCredential(nil) - if err != nil { - log.Fatalf("failed to obtain a credential: %v", err) - } - ctx := context.Background() - clientFactory, err := armdatafactory.NewClientFactory("", cred, nil) - if err != nil { - log.Fatalf("failed to create client: %v", err) - } - poller, err := clientFactory.NewTriggersClient().BeginStop(ctx, "exampleResourceGroup", "exampleFactoryName", "exampleTrigger", nil) - if err != nil { - log.Fatalf("failed to finish the request: %v", err) - } - _, err = poller.PollUntilDone(ctx, nil) - if err != nil { - log.Fatalf("failed to pull the result: %v", err) - } -}