Skip to content

Commit

Permalink
deps: Update Go SDK to 54.0.0 (#11593)
Browse files Browse the repository at this point in the history
Co-authored-by: Steve <[email protected]>
  • Loading branch information
favoretti and jackofallops authored May 7, 2021
1 parent e71b185 commit 922731e
Show file tree
Hide file tree
Showing 93 changed files with 12,899 additions and 10,722 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -160,17 +160,17 @@ func dataSourceDataFactoryRead(d *schema.ResourceData, meta interface{}) error {
d.Set("vsts_configuration", []interface{}{})
d.Set("github_configuration", []interface{}{})
repoType, repo := flattenDataFactoryRepoConfiguration(&resp)
if repoType == datafactory.TypeFactoryVSTSConfiguration {
if repoType == datafactory.TypeBasicFactoryRepoConfigurationTypeFactoryVSTSConfiguration {
if err := d.Set("vsts_configuration", repo); err != nil {
return fmt.Errorf("Error setting `vsts_configuration`: %+v", err)
}
}
if repoType == datafactory.TypeFactoryGitHubConfiguration {
if repoType == datafactory.TypeBasicFactoryRepoConfigurationTypeFactoryGitHubConfiguration {
if err := d.Set("github_configuration", repo); err != nil {
return fmt.Errorf("Error setting `github_configuration`: %+v", err)
}
}
if repoType == datafactory.TypeFactoryRepoConfiguration {
if repoType == datafactory.TypeBasicFactoryRepoConfigurationTypeFactoryRepoConfiguration {
d.Set("vsts_configuration", repo)
d.Set("github_configuration", repo)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -216,7 +216,7 @@ func resourceDataFactoryDatasetAzureBlobCreateUpdate(d *schema.ResourceData, met
azureBlobTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeAzureBlob)
datasetType := string(datafactory.TypeBasicDatasetTypeAzureBlob)
dataset := datafactory.DatasetResource{
Properties: &azureBlobTableset,
Type: &datasetType,
Expand Down Expand Up @@ -269,7 +269,7 @@ func resourceDataFactoryDatasetAzureBlobRead(d *schema.ResourceData, meta interf

azureBlobTable, ok := resp.Properties.AsAzureBlobDataset()
if !ok {
return fmt.Errorf("Error classifying Data Factory Dataset Azure Blob %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifying Data Factory Dataset Azure Blob %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeAzureBlob, *resp.Type)
}

d.Set("additional_properties", azureBlobTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -210,7 +210,7 @@ func resourceDataFactoryDatasetCosmosDbSQLAPICreateUpdate(d *schema.ResourceData
cosmosDbTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypeRelationalTable)
dataset := datafactory.DatasetResource{
Properties: &cosmosDbTableset,
Type: &datasetType,
Expand Down Expand Up @@ -263,7 +263,7 @@ func resourceDataFactoryDatasetCosmosDbSQLAPIRead(d *schema.ResourceData, meta i

cosmosDbTable, ok := resp.Properties.AsCosmosDbSQLAPICollectionDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset CosmosDB SQL API%q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset CosmosDB SQL API%q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeRelationalTable, *resp.Type)
}

d.Set("additional_properties", cosmosDbTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -346,7 +346,7 @@ func resourceDataFactoryDatasetDelimitedTextCreateUpdate(d *schema.ResourceData,
delimited_textTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeDelimitedText)
datasetType := string(datafactory.TypeBasicDatasetTypeDelimitedText)
dataset := datafactory.DatasetResource{
Properties: &delimited_textTableset,
Type: &datasetType,
Expand Down Expand Up @@ -396,7 +396,7 @@ func resourceDataFactoryDatasetDelimitedTextRead(d *schema.ResourceData, meta in

delimited_textTable, ok := resp.Properties.AsDelimitedTextDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset DelimitedText %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset DelimitedText %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeBasicDatasetTypeDelimitedText, *resp.Type)
}

d.Set("additional_properties", delimited_textTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -226,7 +226,7 @@ func resourceDataFactoryDatasetHTTPCreateUpdate(d *schema.ResourceData, meta int
httpTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeHTTPFile)
datasetType := string(datafactory.TypeBasicDatasetTypeHTTPFile)
dataset := datafactory.DatasetResource{
Properties: &httpTableset,
Type: &datasetType,
Expand Down Expand Up @@ -279,7 +279,7 @@ func resourceDataFactoryDatasetHTTPRead(d *schema.ResourceData, meta interface{}

httpTable, ok := resp.Properties.AsHTTPDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset HTTP %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset HTTP %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeHTTPFile, *resp.Type)
}

d.Set("additional_properties", httpTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -272,7 +272,7 @@ func resourceDataFactoryDatasetJSONCreateUpdate(d *schema.ResourceData, meta int
jsonTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeJSON)
datasetType := string(datafactory.TypeBasicDatasetTypeJSON)
dataset := datafactory.DatasetResource{
Properties: &jsonTableset,
Type: &datasetType,
Expand Down Expand Up @@ -325,7 +325,7 @@ func resourceDataFactoryDatasetJSONRead(d *schema.ResourceData, meta interface{}

jsonTable, ok := resp.Properties.AsJSONDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset JSON %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset JSON %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeJSON, *resp.Type)
}

d.Set("additional_properties", jsonTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ func resourceDataFactoryDatasetMySQLCreateUpdate(d *schema.ResourceData, meta in
mysqlTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypeMySQLTable)
dataset := datafactory.DatasetResource{
Properties: &mysqlTableset,
Type: &datasetType,
Expand Down Expand Up @@ -261,7 +261,7 @@ func resourceDataFactoryDatasetMySQLRead(d *schema.ResourceData, meta interface{

mysqlTable, ok := resp.Properties.AsRelationalTableDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset MySQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset MySQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeMySQLTable, *resp.Type)
}

d.Set("additional_properties", mysqlTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -291,7 +291,7 @@ func resourceDataFactoryDatasetParquetCreateUpdate(d *schema.ResourceData, meta
parquetTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeParquet)
datasetType := string(datafactory.TypeBasicDatasetTypeParquet)
dataset := datafactory.DatasetResource{
Properties: &parquetTableset,
Type: &datasetType,
Expand Down Expand Up @@ -332,7 +332,7 @@ func resourceDataFactoryDatasetParquetRead(d *schema.ResourceData, meta interfac

parquetTable, ok := resp.Properties.AsParquetDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset Parquet %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset Parquet %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", id.Name, id.FactoryName, id.ResourceGroup, datafactory.TypeBasicDatasetTypeParquet, *resp.Type)
}

d.Set("additional_properties", parquetTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ func resourceDataFactoryDatasetPostgreSQLCreateUpdate(d *schema.ResourceData, me
postgresqlTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypePostgreSQLTable)
dataset := datafactory.DatasetResource{
Properties: &postgresqlTableset,
Type: &datasetType,
Expand Down Expand Up @@ -261,7 +261,7 @@ func resourceDataFactoryDatasetPostgreSQLRead(d *schema.ResourceData, meta inter

postgresqlTable, ok := resp.Properties.AsRelationalTableDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset PostgreSQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset PostgreSQL %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypePostgreSQLTable, *resp.Type)
}

d.Set("additional_properties", postgresqlTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -215,7 +215,7 @@ func resourceDataFactoryDatasetSnowflakeCreateUpdate(d *schema.ResourceData, met
snowflakeTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeRelationalTable)
datasetType := string(datafactory.TypeBasicDatasetTypeSnowflakeTable)
dataset := datafactory.DatasetResource{
Properties: &snowflakeTableset,
Type: &datasetType,
Expand Down Expand Up @@ -268,7 +268,7 @@ func resourceDataFactoryDatasetSnowflakeRead(d *schema.ResourceData, meta interf

snowflakeTable, ok := resp.Properties.AsSnowflakeDataset()
if !ok {
return fmt.Errorf("Error classifying Data Factory Dataset Snowflake %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeRelationalTable, *resp.Type)
return fmt.Errorf("Error classifying Data Factory Dataset Snowflake %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeSnowflakeTable, *resp.Type)
}

d.Set("additional_properties", snowflakeTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -208,7 +208,7 @@ func resourceDataFactoryDatasetSQLServerTableCreateUpdate(d *schema.ResourceData
sqlServerTableset.Structure = expandDataFactoryDatasetStructure(v.([]interface{}))
}

datasetType := string(datafactory.TypeSQLServerTable)
datasetType := string(datafactory.TypeBasicDatasetTypeSQLServerTable)
dataset := datafactory.DatasetResource{
Properties: &sqlServerTableset,
Type: &datasetType,
Expand Down Expand Up @@ -261,7 +261,7 @@ func resourceDataFactoryDatasetSQLServerTableRead(d *schema.ResourceData, meta i

sqlServerTable, ok := resp.Properties.AsSQLServerTableDataset()
if !ok {
return fmt.Errorf("Error classifiying Data Factory Dataset SQL Server Table %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeSQLServerTable, *resp.Type)
return fmt.Errorf("Error classifiying Data Factory Dataset SQL Server Table %q (Data Factory %q / Resource Group %q): Expected: %q Received: %q", name, dataFactoryName, resourceGroup, datafactory.TypeBasicDatasetTypeSQLServerTable, *resp.Type)
}

d.Set("additional_properties", sqlServerTable.AdditionalProperties)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -63,11 +63,11 @@ func resourceDataFactoryIntegrationRuntimeAzure() *schema.Resource {
"compute_type": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.General),
Default: string(datafactory.DataFlowComputeTypeGeneral),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.General),
string(datafactory.ComputeOptimized),
string(datafactory.MemoryOptimized),
string(datafactory.DataFlowComputeTypeGeneral),
string(datafactory.DataFlowComputeTypeComputeOptimized),
string(datafactory.DataFlowComputeTypeMemoryOptimized),
}, false),
},

Expand Down Expand Up @@ -116,7 +116,7 @@ func resourceDataFactoryIntegrationRuntimeAzureCreateUpdate(d *schema.ResourceDa

managedIntegrationRuntime := datafactory.ManagedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeManaged,
Type: datafactory.TypeBasicIntegrationRuntimeTypeManaged,
ManagedIntegrationRuntimeTypeProperties: &datafactory.ManagedIntegrationRuntimeTypeProperties{
ComputeProperties: expandDataFactoryIntegrationRuntimeAzureComputeProperties(d),
},
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -102,20 +102,20 @@ func resourceDataFactoryIntegrationRuntimeAzureSsis() *schema.Resource {
"edition": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.Standard),
Default: string(datafactory.IntegrationRuntimeEditionStandard),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.Standard),
string(datafactory.Enterprise),
string(datafactory.IntegrationRuntimeEditionStandard),
string(datafactory.IntegrationRuntimeEditionEnterprise),
}, false),
},

"license_type": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.LicenseIncluded),
Default: string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.LicenseIncluded),
string(datafactory.BasePrice),
string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
string(datafactory.IntegrationRuntimeLicenseTypeBasePrice),
}, false),
},

Expand Down Expand Up @@ -225,7 +225,7 @@ func resourceDataFactoryIntegrationRuntimeAzureSsisCreateUpdate(d *schema.Resour
description := d.Get("description").(string)
managedIntegrationRuntime := datafactory.ManagedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeManaged,
Type: datafactory.TypeBasicIntegrationRuntimeTypeManaged,
ManagedIntegrationRuntimeTypeProperties: &datafactory.ManagedIntegrationRuntimeTypeProperties{
ComputeProperties: expandDataFactoryIntegrationRuntimeAzureSsisComputeProperties(d),
SsisProperties: expandDataFactoryIntegrationRuntimeAzureSsisProperties(d),
Expand Down Expand Up @@ -395,7 +395,7 @@ func expandDataFactoryIntegrationRuntimeAzureSsisProperties(d *schema.ResourceDa
if adminPassword := catalogInfo["administrator_password"]; adminPassword.(string) != "" {
ssisProperties.CatalogInfo.CatalogAdminPassword = &datafactory.SecureString{
Value: utils.String(adminPassword.(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}
}
}
Expand All @@ -405,7 +405,7 @@ func expandDataFactoryIntegrationRuntimeAzureSsisProperties(d *schema.ResourceDa

sasToken := &datafactory.SecureString{
Value: utils.String(customSetupScript["sas_token"].(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}

ssisProperties.CustomSetupScriptProperties = &datafactory.IntegrationRuntimeCustomSetupScriptProperties{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -102,20 +102,20 @@ func resourceDataFactoryIntegrationRuntimeManaged() *schema.Resource {
"edition": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.Standard),
Default: string(datafactory.IntegrationRuntimeEditionStandard),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.Standard),
string(datafactory.Enterprise),
string(datafactory.IntegrationRuntimeEditionStandard),
string(datafactory.IntegrationRuntimeEditionEnterprise),
}, false),
},

"license_type": {
Type: schema.TypeString,
Optional: true,
Default: string(datafactory.LicenseIncluded),
Default: string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
ValidateFunc: validation.StringInSlice([]string{
string(datafactory.LicenseIncluded),
string(datafactory.BasePrice),
string(datafactory.IntegrationRuntimeLicenseTypeLicenseIncluded),
string(datafactory.IntegrationRuntimeLicenseTypeBasePrice),
}, false),
},

Expand Down Expand Up @@ -225,7 +225,7 @@ func resourceDataFactoryIntegrationRuntimeManagedCreateUpdate(d *schema.Resource
description := d.Get("description").(string)
managedIntegrationRuntime := datafactory.ManagedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeManaged,
Type: datafactory.TypeBasicIntegrationRuntimeTypeManaged,
ManagedIntegrationRuntimeTypeProperties: &datafactory.ManagedIntegrationRuntimeTypeProperties{
ComputeProperties: expandDataFactoryIntegrationRuntimeManagedComputeProperties(d),
SsisProperties: expandDataFactoryIntegrationRuntimeManagedSsisProperties(d),
Expand Down Expand Up @@ -395,7 +395,7 @@ func expandDataFactoryIntegrationRuntimeManagedSsisProperties(d *schema.Resource
if adminPassword := catalogInfo["administrator_password"]; adminPassword.(string) != "" {
ssisProperties.CatalogInfo.CatalogAdminPassword = &datafactory.SecureString{
Value: utils.String(adminPassword.(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}
}
}
Expand All @@ -405,7 +405,7 @@ func expandDataFactoryIntegrationRuntimeManagedSsisProperties(d *schema.Resource

sasToken := &datafactory.SecureString{
Value: utils.String(customSetupScript["sas_token"].(string)),
Type: datafactory.TypeSecureString,
Type: datafactory.TypeTypeSecureString,
}

ssisProperties.CustomSetupScriptProperties = &datafactory.IntegrationRuntimeCustomSetupScriptProperties{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -117,7 +117,7 @@ func resourceDataFactoryIntegrationRuntimeSelfHostedCreateUpdate(d *schema.Resou

selfHostedIntegrationRuntime := datafactory.SelfHostedIntegrationRuntime{
Description: &description,
Type: datafactory.TypeSelfHosted,
Type: datafactory.TypeBasicIntegrationRuntimeTypeSelfHosted,
}

properties := expandAzureRmDataFactoryIntegrationRuntimeSelfHostedTypeProperties(d)
Expand Down Expand Up @@ -246,7 +246,7 @@ func expandAzureRmDataFactoryIntegrationRuntimeSelfHostedTypeProperties(d *schem
linkedInfo := &datafactory.SelfHostedIntegrationRuntimeTypeProperties{
LinkedInfo: &datafactory.LinkedIntegrationRuntimeRbacAuthorization{
ResourceID: &rbac,
AuthorizationType: datafactory.AuthorizationTypeRBAC,
AuthorizationType: datafactory.AuthorizationTypeAuthorizationTypeRBAC,
},
}
return linkedInfo
Expand Down
Loading

0 comments on commit 922731e

Please sign in to comment.