Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

deps: Update Go SDK to 54.0.0 #11593

Merged
merged 5 commits into from
May 7, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -160,17 +160,17 @@ func dataSourceDataFactoryRead(d *schema.ResourceData, meta interface{}) error {
d.Set("vsts_configuration", []interface{}{})
d.Set("github_configuration", []interface{}{})
repoType, repo := flattenDataFactoryRepoConfiguration(&resp)
if repoType == datafactory.TypeFactoryVSTSConfiguration {
if repoType == datafactory.TypeBasicFactoryRepoConfigurationTypeFactoryVSTSConfiguration {
if err := d.Set("vsts_configuration", repo); err != nil {
return fmt.Errorf("Error setting `vsts_configuration`: %+v", err)
}
}
if repoType == datafactory.TypeFactoryGitHubConfiguration {
if repoType == datafactory.TypeBasicFactoryRepoConfigurationTypeFactoryGitHubConfiguration {
if err := d.Set("github_configuration", repo); err != nil {
return fmt.Errorf("Error setting `github_configuration`: %+v", err)
}
}
if repoType == datafactory.TypeFactoryRepoConfiguration {
if repoType == datafactory.TypeBasicFactoryRepoConfigurationTypeFactoryRepoConfiguration {
d.Set("vsts_configuration", repo)
d.Set("github_configuration", repo)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -216,7 +216,7 @@ func resourceDataFactoryDatasetAzureBlobCreateUpdate(d *schema.ResourceData, met
azureBlobTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeAzureBlob)
datasetType := string(datafactory.TypeBasicDatasetTypeAzureBlob)
dataset := datafactory.DatasetResource{
Properties: &azureBlobTableset,
Type: &datasetType,
Expand Down Expand Up @@ -269,7 +269,7 @@ func resourceDataFactoryDatasetAzureBlobRead(d *schema.ResourceData, meta interf

azureBlobTable, ok := resp.Properties.AsAzureBlobDataset()
if !ok {
return fmt.Errorf("Error classifying Data Factory Dataset Azure Blob %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifying Data Factory Dataset Azure Blob %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeAzureBlob, *resp.Type)
}

d.Set("additional_properties", azureBlobTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -210,7 +210,7 @@ func resourceDataFactoryDatasetCosmosDbSQLAPICreateUpdate(d *schema.ResourceData
cosmosDbTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypeRelationalTable)
dataset := datafactory.DatasetResource{
Properties: &cosmosDbTableset,
Type: &datasetType,
Expand Down Expand Up @@ -263,7 +263,7 @@ func resourceDataFactoryDatasetCosmosDbSQLAPIRead(d *schema.ResourceData, meta i

cosmosDbTable, ok := resp.Properties.AsCosmosDbSQLAPICollectionDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset CosmosDB SQL API%q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset CosmosDB SQL API%q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeRelationalTable, *resp.Type)
}

d.Set("additional_properties", cosmosDbTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -346,7 +346,7 @@ func resourceDataFactoryDatasetDelimitedTextCreateUpdate(d *schema.ResourceData,
delimited_textTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeDelimitedText)
datasetType := string(datafactory.TypeBasicDatasetTypeDelimitedText)
dataset := datafactory.DatasetResource{
Properties: &delimited_textTableset,
Type: &datasetType,
Expand Down Expand Up @@ -396,7 +396,7 @@ func resourceDataFactoryDatasetDelimitedTextRead(d *schema.ResourceData, meta in

delimited_textTable, ok := resp.Properties.AsDelimitedTextDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset DelimitedText %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset DelimitedText %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeBasicDatasetTypeDelimitedText, *resp.Type)
}

d.Set("additional_properties", delimited_textTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -226,7 +226,7 @@ func resourceDataFactoryDatasetHTTPCreateUpdate(d *schema.ResourceData, meta int
httpTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeHTTPFile)
datasetType := string(datafactory.TypeBasicDatasetTypeHTTPFile)
dataset := datafactory.DatasetResource{
Properties: &httpTableset,
Type: &datasetType,
Expand Down Expand Up @@ -279,7 +279,7 @@ func resourceDataFactoryDatasetHTTPRead(d *schema.ResourceData, meta interface{}

httpTable, ok := resp.Properties.AsHTTPDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset HTTP %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset HTTP %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeHTTPFile, *resp.Type)
}

d.Set("additional_properties", httpTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -272,7 +272,7 @@ func resourceDataFactoryDatasetJSONCreateUpdate(d *schema.ResourceData, meta int
jsonTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeJSON)
datasetType := string(datafactory.TypeBasicDatasetTypeJSON)
dataset := datafactory.DatasetResource{
Properties: &jsonTableset,
Type: &datasetType,
Expand Down Expand Up @@ -325,7 +325,7 @@ func resourceDataFactoryDatasetJSONRead(d *schema.ResourceData, meta interface{}

jsonTable, ok := resp.Properties.AsJSONDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset JSON %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset JSON %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeJSON, *resp.Type)
}

d.Set("additional_properties", jsonTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ func resourceDataFactoryDatasetMySQLCreateUpdate(d *schema.ResourceData, meta in
mysqlTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypeMySQLTable)
dataset := datafactory.DatasetResource{
Properties: &mysqlTableset,
Type: &datasetType,
Expand Down Expand Up @@ -261,7 +261,7 @@ func resourceDataFactoryDatasetMySQLRead(d *schema.ResourceData, meta interface{

mysqlTable, ok := resp.Properties.AsRelationalTableDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset MySQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset MySQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeMySQLTable, *resp.Type)
}

d.Set("additional_properties", mysqlTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -291,7 +291,7 @@ func resourceDataFactoryDatasetParquetCreateUpdate(d *schema.ResourceData, meta
parquetTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeParquet)
datasetType := string(datafactory.TypeBasicDatasetTypeParquet)
dataset := datafactory.DatasetResource{
Properties: &parquetTableset,
Type: &datasetType,
Expand Down Expand Up @@ -332,7 +332,7 @@ func resourceDataFactoryDatasetParquetRead(d *schema.ResourceData, meta interfac

parquetTable, ok := resp.Properties.AsParquetDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset Parquet %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset Parquet %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeBasicDatasetTypeParquet, *resp.Type)
}

d.Set("additional_properties", parquetTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ func resourceDataFactoryDatasetPostgreSQLCreateUpdate(d *schema.ResourceData, me
postgresqlTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypePostgreSQLTable)
dataset := datafactory.DatasetResource{
Properties: &postgresqlTableset,
Type: &datasetType,
Expand Down Expand Up @@ -261,7 +261,7 @@ func resourceDataFactoryDatasetPostgreSQLRead(d *schema.ResourceData, meta inter

postgresqlTable, ok := resp.Properties.AsRelationalTableDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset PostgreSQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset PostgreSQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypePostgreSQLTable, *resp.Type)
}

d.Set("additional_properties", postgresqlTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -215,7 +215,7 @@ func resourceDataFactoryDatasetSnowflakeCreateUpdate(d *schema.ResourceData, met
snowflakeTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypeSnowflakeTable)
dataset := datafactory.DatasetResource{
Properties: &snowflakeTableset,
Type: &datasetType,
Expand Down Expand Up @@ -268,7 +268,7 @@ func resourceDataFactoryDatasetSnowflakeRead(d *schema.ResourceData, meta interf

snowflakeTable, ok := resp.Properties.AsSnowflakeDataset()
if !ok {
return fmt.Errorf("Error classifying Data Factory Dataset Snowflake %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifying Data Factory Dataset Snowflake %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeSnowflakeTable, *resp.Type)
}

d.Set("additional_properties", snowflakeTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ func resourceDataFactoryDatasetSQLServerTableCreateUpdate(d *schema.ResourceData
sqlServerTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeSQLServerTable)
datasetType := string(datafactory.TypeBasicDatasetTypeSQLServerTable)
dataset := datafactory.DatasetResource{
Properties: &sqlServerTableset,
Type: &datasetType,
Expand Down Expand Up @@ -261,7 +261,7 @@ func resourceDataFactoryDatasetSQLServerTableRead(d *schema.ResourceData, meta i

sqlServerTable, ok := resp.Properties.AsSQLServerTableDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset SQL Server Table %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeSQLServerTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset SQL Server Table %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeSQLServerTable, *resp.Type)
}

d.Set("additional_properties", sqlServerTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -63,11 +63,11 @@ func resourceDataFactoryIntegrationRuntimeAzure() *schema.Resource {
"compute_type": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.General),
Default: string(datafactory.DataFlowComputeTypeGeneral),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.General),
string(datafactory.ComputeOptimized),
string(datafactory.MemoryOptimized),
string(datafactory.DataFlowComputeTypeGeneral),
string(datafactory.DataFlowComputeTypeComputeOptimized),
string(datafactory.DataFlowComputeTypeMemoryOptimized),
}, false),
},

Expand Down Expand Up @@ -116,7 +116,7 @@ func resourceDataFactoryIntegrationRuntimeAzureCreateUpdate(d *schema.ResourceDa

managedIntegrationRuntime := datafactory.ManagedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeManaged,
Type: datafactory.TypeBasicIntegrationRuntimeTypeManaged,
ManagedIntegrationRuntimeTypeProperties: &datafactory.ManagedIntegrationRuntimeTypeProperties{
ComputeProperties: expandDataFactoryIntegrationRuntimeAzureComputeProperties(d),
},
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -102,20 +102,20 @@ func resourceDataFactoryIntegrationRuntimeAzureSsis() *schema.Resource {
"edition": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.Standard),
Default: string(datafactory.IntegrationRuntimeEditionStandard),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.Standard),
string(datafactory.Enterprise),
string(datafactory.IntegrationRuntimeEditionStandard),
string(datafactory.IntegrationRuntimeEditionEnterprise),
}, false),
},

"license_type": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.LicenseIncluded),
Default: string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.LicenseIncluded),
string(datafactory.BasePrice),
string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
string(datafactory.IntegrationRuntimeLicenseTypeBasePrice),
}, false),
},

Expand Down Expand Up @@ -225,7 +225,7 @@ func resourceDataFactoryIntegrationRuntimeAzureSsisCreateUpdate(d *schema.Resour
description := d.Get("description").(string)
managedIntegrationRuntime := datafactory.ManagedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeManaged,
Type: datafactory.TypeBasicIntegrationRuntimeTypeManaged,
ManagedIntegrationRuntimeTypeProperties: &datafactory.ManagedIntegrationRuntimeTypeProperties{
ComputeProperties: expandDataFactoryIntegrationRuntimeAzureSsisComputeProperties(d),
SsisProperties: expandDataFactoryIntegrationRuntimeAzureSsisProperties(d),
Expand Down Expand Up @@ -395,7 +395,7 @@ func expandDataFactoryIntegrationRuntimeAzureSsisProperties(d *schema.ResourceDa
if adminPassword := catalogInfo["administrator_password"]; adminPassword.(string) != "" {
ssisProperties.CatalogInfo.CatalogAdminPassword = &datafactory.SecureString{
Value: utils.String(adminPassword.(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}
}
}
Expand All @@ -405,7 +405,7 @@ func expandDataFactoryIntegrationRuntimeAzureSsisProperties(d *schema.ResourceDa

sasToken := &datafactory.SecureString{
Value: utils.String(customSetupScript["sas_token"].(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}

ssisProperties.CustomSetupScriptProperties = &datafactory.IntegrationRuntimeCustomSetupScriptProperties{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -102,20 +102,20 @@ func resourceDataFactoryIntegrationRuntimeManaged() *schema.Resource {
"edition": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.Standard),
Default: string(datafactory.IntegrationRuntimeEditionStandard),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.Standard),
string(datafactory.Enterprise),
string(datafactory.IntegrationRuntimeEditionStandard),
string(datafactory.IntegrationRuntimeEditionEnterprise),
}, false),
},

"license_type": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.LicenseIncluded),
Default: string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.LicenseIncluded),
string(datafactory.BasePrice),
string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
string(datafactory.IntegrationRuntimeLicenseTypeBasePrice),
}, false),
},

Expand Down Expand Up @@ -225,7 +225,7 @@ func resourceDataFactoryIntegrationRuntimeManagedCreateUpdate(d *schema.Resource
description := d.Get("description").(string)
managedIntegrationRuntime := datafactory.ManagedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeManaged,
Type: datafactory.TypeBasicIntegrationRuntimeTypeManaged,
ManagedIntegrationRuntimeTypeProperties: &datafactory.ManagedIntegrationRuntimeTypeProperties{
ComputeProperties: expandDataFactoryIntegrationRuntimeManagedComputeProperties(d),
SsisProperties: expandDataFactoryIntegrationRuntimeManagedSsisProperties(d),
Expand Down Expand Up @@ -395,7 +395,7 @@ func expandDataFactoryIntegrationRuntimeManagedSsisProperties(d *schema.Resource
if adminPassword := catalogInfo["administrator_password"]; adminPassword.(string) != "" {
ssisProperties.CatalogInfo.CatalogAdminPassword = &datafactory.SecureString{
Value: utils.String(adminPassword.(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}
}
}
Expand All @@ -405,7 +405,7 @@ func expandDataFactoryIntegrationRuntimeManagedSsisProperties(d *schema.Resource

sasToken := &datafactory.SecureString{
Value: utils.String(customSetupScript["sas_token"].(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}

ssisProperties.CustomSetupScriptProperties = &datafactory.IntegrationRuntimeCustomSetupScriptProperties{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -117,7 +117,7 @@ func resourceDataFactoryIntegrationRuntimeSelfHostedCreateUpdate(d *schema.Resou

selfHostedIntegrationRuntime := datafactory.SelfHostedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeSelfHosted,
Type: datafactory.TypeBasicIntegrationRuntimeTypeSelfHosted,
}

properties := expandAzureRmDataFactoryIntegrationRuntimeSelfHostedTypeProperties(d)
Expand Down Expand Up @@ -246,7 +246,7 @@ func expandAzureRmDataFactoryIntegrationRuntimeSelfHostedTypeProperties(d *schem
linkedInfo := &datafactory.SelfHostedIntegrationRuntimeTypeProperties{
LinkedInfo: &datafactory.LinkedIntegrationRuntimeRbacAuthorization{
ResourceID: &rbac,
AuthorizationType: datafactory.AuthorizationTypeRBAC,
AuthorizationType: datafactory.AuthorizationTypeAuthorizationTypeRBAC,
},
}
return linkedInfo
Expand Down
Loading