diff --git a/.openapi/kas-fleet-manager.yaml b/.openapi/kas-fleet-manager.yaml index 3dcfcbd8..d46f6d0f 100644 --- a/.openapi/kas-fleet-manager.yaml +++ b/.openapi/kas-fleet-manager.yaml @@ -967,7 +967,7 @@ paths: content: application/json: schema: - $ref: '#/components/schemas/EnterpriseClusterRegistrationResponse' + $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters' description: Enterprise cluster registered "400": content: @@ -1100,6 +1100,128 @@ paths: description: Unexpected error occurred security: - Bearer: [ ] + get: + tags: + - enterprise-dataplane-clusters + operationId: getEnterpriseClusterById + description: Returns enterprise data plane cluster by ID + parameters: + - in: path + name: id + description: ID of the enterprise data plane cluster + schema: + type: string + required: true + responses: + "200": + content: + application/json: + schema: + $ref: '#/components/schemas/EnterpriseCluster' + examples: + EnterpriseClusterExample: + $ref: '#/components/examples/EnterpriseClusterExample' + description: Found enterprise data plane cluster with provided ID + "401": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 401Example: + $ref: '#/components/examples/401Example' + description: Auth token is invalid + "403": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 403Example: + $ref: '#/components/examples/403Example' + description: User not authorized to access the service + "404": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 404Example: + $ref: '#/components/examples/404Example' + description: No Enterprise data plane cluster with specified ID exists + "500": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 500Example: + $ref: '#/components/examples/500Example' + description: Unexpected error occurred + security: + - Bearer: [ ] + + /api/kafkas_mgmt/v1/clusters/{id}/addon_parameters: + get: + tags: + - enterprise-dataplane-clusters + operationId: getEnterpriseClusterWithAddonParameters + description: Returns enterprise data plane cluster by ID along with its addon parameters + parameters: + - in: path + name: id + description: ID of the enterprise data plane cluster + schema: + type: string + required: true + responses: + "200": + content: + application/json: + schema: + $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters' + examples: + EnterpriseClusterWithAddonsExample: + $ref: '#/components/examples/EnterpriseClusterWithAddonParametersExample' + description: Returns the enterprise data plane cluster with addon parameters for an existing enterprise data plane cluster with provided ID + "401": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 401Example: + $ref: '#/components/examples/401Example' + description: Auth token is invalid + "403": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 403Example: + $ref: '#/components/examples/403Example' + description: User not authorized to access the service + "404": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 404Example: + $ref: '#/components/examples/404Example' + description: No Enterprise data plane cluster with specified ID exists + "500": + content: + application/json: + schema: + $ref: '#/components/schemas/Error' + examples: + 500Example: + $ref: '#/components/examples/500Example' + description: Unexpected error occurred + security: + - Bearer: [ ] components: schemas: @@ -1818,8 +1940,8 @@ components: The node count value has to be a multiple of 3 with a minimum of 3 nodes. type: integer format: int32 - EnterpriseClusterRegistrationResponse: - description: Enterprise cluster registration endpoint response + EnterpriseClusterWithAddonParameters: + description: Enterprise cluster with addon parameters allOf: - $ref: "#/components/schemas/ObjectReference" - type: object @@ -2028,6 +2150,13 @@ components: name: "test_kafka" plan: "standard.x1" EnterpriseClusterExample: + value: + id: "abcd1234ascd3456fdks9485lskd030h" + kind: "cluster" + href: "/api/kafkas_mgmr/v1/clusters/abcd1234ascd3456fdks9485lskd030h" + status: "ready" + cluster_id: "abcd1234ascd3456fdks9485lskd030h" + EnterpriseClusterWithAddonParametersExample: value: id: "abcd1234ascd3456fdks9485lskd030h" kind: "cluster" @@ -2035,6 +2164,32 @@ components: status: "ready" cluster_id: "abcd1234ascd3456fdks9485lskd030h" access_kafkas_via_private_network: true + fleetshard_parameters: [ + { + id: "sso-auth-server-url", + value: "http://someurl/auth/realms/rhoas" + }, + { + id: "sso-client-id", + value: "kas-fleetshard-agent-1234abcd1234abcd1234abcd1234abcd" + }, + { + id: "sso-secret", + value: "some-secret" + }, + { + id: "cluster-id", + value: "1234abcd1234abcd1234abcd1234abcd" + }, + { + id: "poll-interval", + value: "15s" + }, + { + id: "resync-interval", + value: "60s" + } + ] KafkaRequestExample: value: id: "1iSY6RQ3JKI8Q0OTmjQFd3ocFRg" diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/.openapi-generator/FILES b/app-services-sdk-go/kafkamgmt/apiv1/client/.openapi-generator/FILES index fefcd86d..5230a268 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/.openapi-generator/FILES +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/.openapi-generator/FILES @@ -16,8 +16,8 @@ docs/EnterpriseCluster.md docs/EnterpriseClusterAllOf.md docs/EnterpriseClusterList.md docs/EnterpriseClusterListAllOf.md -docs/EnterpriseClusterRegistrationResponse.md -docs/EnterpriseClusterRegistrationResponseAllOf.md +docs/EnterpriseClusterWithAddonParameters.md +docs/EnterpriseClusterWithAddonParametersAllOf.md docs/EnterpriseDataplaneClustersApi.md docs/EnterpriseOsdClusterPayload.md docs/Error.md @@ -72,8 +72,8 @@ model_enterprise_cluster.go model_enterprise_cluster_all_of.go model_enterprise_cluster_list.go model_enterprise_cluster_list_all_of.go -model_enterprise_cluster_registration_response.go -model_enterprise_cluster_registration_response_all_of.go +model_enterprise_cluster_with_addon_parameters.go +model_enterprise_cluster_with_addon_parameters_all_of.go model_enterprise_osd_cluster_payload.go model_error.go model_error_list.go diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/api/openapi.yaml b/app-services-sdk-go/kafkamgmt/apiv1/client/api/openapi.yaml index 87c06783..a510f4e9 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/api/openapi.yaml +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/api/openapi.yaml @@ -1248,7 +1248,7 @@ paths: content: application/json: schema: - $ref: '#/components/schemas/EnterpriseClusterRegistrationResponse' + $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters' description: Enterprise cluster registered "400": content: @@ -1387,6 +1387,133 @@ paths: - Bearer: [] tags: - enterprise-dataplane-clusters + get: + description: Returns enterprise data plane cluster by ID + operationId: getEnterpriseClusterById + parameters: + - description: ID of the enterprise data plane cluster + explode: false + in: path + name: id + required: true + schema: + type: string + style: simple + responses: + "200": + content: + application/json: + examples: + EnterpriseClusterExample: + $ref: '#/components/examples/EnterpriseClusterExample' + schema: + $ref: '#/components/schemas/EnterpriseCluster' + description: Found enterprise data plane cluster with provided ID + "401": + content: + application/json: + examples: + "401Example": + $ref: '#/components/examples/401Example' + schema: + $ref: '#/components/schemas/Error' + description: Auth token is invalid + "403": + content: + application/json: + examples: + "403Example": + $ref: '#/components/examples/403Example' + schema: + $ref: '#/components/schemas/Error' + description: User not authorized to access the service + "404": + content: + application/json: + examples: + "404Example": + $ref: '#/components/examples/404Example' + schema: + $ref: '#/components/schemas/Error' + description: No Enterprise data plane cluster with specified ID exists + "500": + content: + application/json: + examples: + "500Example": + $ref: '#/components/examples/500Example' + schema: + $ref: '#/components/schemas/Error' + description: Unexpected error occurred + security: + - Bearer: [] + tags: + - enterprise-dataplane-clusters + /api/kafkas_mgmt/v1/clusters/{id}/addon_parameters: + get: + description: Returns enterprise data plane cluster by ID along with its addon + parameters + operationId: getEnterpriseClusterWithAddonParameters + parameters: + - description: ID of the enterprise data plane cluster + explode: false + in: path + name: id + required: true + schema: + type: string + style: simple + responses: + "200": + content: + application/json: + examples: + EnterpriseClusterWithAddonsExample: + $ref: '#/components/examples/EnterpriseClusterWithAddonParametersExample' + schema: + $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters' + description: Returns the enterprise data plane cluster with addon parameters + for an existing enterprise data plane cluster with provided ID + "401": + content: + application/json: + examples: + "401Example": + $ref: '#/components/examples/401Example' + schema: + $ref: '#/components/schemas/Error' + description: Auth token is invalid + "403": + content: + application/json: + examples: + "403Example": + $ref: '#/components/examples/403Example' + schema: + $ref: '#/components/schemas/Error' + description: User not authorized to access the service + "404": + content: + application/json: + examples: + "404Example": + $ref: '#/components/examples/404Example' + schema: + $ref: '#/components/schemas/Error' + description: No Enterprise data plane cluster with specified ID exists + "500": + content: + application/json: + examples: + "500Example": + $ref: '#/components/examples/500Example' + schema: + $ref: '#/components/schemas/Error' + description: Unexpected error occurred + security: + - Bearer: [] + tags: + - enterprise-dataplane-clusters components: examples: USRegionExample: @@ -1406,6 +1533,13 @@ components: name: test_kafka plan: standard.x1 EnterpriseClusterExample: + value: + id: abcd1234ascd3456fdks9485lskd030h + kind: cluster + href: /api/kafkas_mgmr/v1/clusters/abcd1234ascd3456fdks9485lskd030h + status: ready + cluster_id: abcd1234ascd3456fdks9485lskd030h + EnterpriseClusterWithAddonParametersExample: value: id: abcd1234ascd3456fdks9485lskd030h kind: cluster @@ -1413,6 +1547,19 @@ components: status: ready cluster_id: abcd1234ascd3456fdks9485lskd030h access_kafkas_via_private_network: true + fleetshard_parameters: + - id: sso-auth-server-url + value: http://someurl/auth/realms/rhoas + - id: sso-client-id + value: kas-fleetshard-agent-1234abcd1234abcd1234abcd1234abcd + - id: sso-secret + value: some-secret + - id: cluster-id + value: 1234abcd1234abcd1234abcd1234abcd + - id: poll-interval + value: 15s + - id: resync-interval + value: 60s KafkaRequestExample: value: id: 1iSY6RQ3JKI8Q0OTmjQFd3ocFRg @@ -2382,11 +2529,11 @@ components: - cluster_ingress_dns_name - kafka_machine_pool_node_count type: object - EnterpriseClusterRegistrationResponse: + EnterpriseClusterWithAddonParameters: allOf: - $ref: '#/components/schemas/ObjectReference' - - $ref: '#/components/schemas/EnterpriseClusterRegistrationResponse_allOf' - description: Enterprise cluster registration endpoint response + - $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters_allOf' + description: Enterprise cluster with addon parameters FleetshardParameter: description: Fleetshard parameter consumed by enterprise cluster properties: @@ -2700,7 +2847,7 @@ components: - $ref: '#/components/schemas/InstantQuery' type: array type: object - EnterpriseClusterRegistrationResponse_allOf: + EnterpriseClusterWithAddonParameters_allOf: properties: access_kafkas_via_private_network: description: Indicates whether Kafkas created on this data plane cluster diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/api_enterprise_dataplane_clusters.go b/app-services-sdk-go/kafkamgmt/apiv1/client/api_enterprise_dataplane_clusters.go index 4c5319b7..269ac1ff 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/api_enterprise_dataplane_clusters.go +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/api_enterprise_dataplane_clusters.go @@ -36,6 +36,36 @@ type EnterpriseDataplaneClustersApi interface { // @return Error DeleteEnterpriseClusterByIdExecute(r ApiDeleteEnterpriseClusterByIdRequest) (*Error, *http.Response, error) + /* + GetEnterpriseClusterById Method for GetEnterpriseClusterById + + Returns enterprise data plane cluster by ID + + @param ctx context.Context - for authentication, logging, cancellation, deadlines, tracing, etc. Passed from http.Request or context.Background(). + @param id ID of the enterprise data plane cluster + @return ApiGetEnterpriseClusterByIdRequest + */ + GetEnterpriseClusterById(ctx context.Context, id string) ApiGetEnterpriseClusterByIdRequest + + // GetEnterpriseClusterByIdExecute executes the request + // @return EnterpriseCluster + GetEnterpriseClusterByIdExecute(r ApiGetEnterpriseClusterByIdRequest) (*EnterpriseCluster, *http.Response, error) + + /* + GetEnterpriseClusterWithAddonParameters Method for GetEnterpriseClusterWithAddonParameters + + Returns enterprise data plane cluster by ID along with its addon parameters + + @param ctx context.Context - for authentication, logging, cancellation, deadlines, tracing, etc. Passed from http.Request or context.Background(). + @param id ID of the enterprise data plane cluster + @return ApiGetEnterpriseClusterWithAddonParametersRequest + */ + GetEnterpriseClusterWithAddonParameters(ctx context.Context, id string) ApiGetEnterpriseClusterWithAddonParametersRequest + + // GetEnterpriseClusterWithAddonParametersExecute executes the request + // @return EnterpriseClusterWithAddonParameters + GetEnterpriseClusterWithAddonParametersExecute(r ApiGetEnterpriseClusterWithAddonParametersRequest) (*EnterpriseClusterWithAddonParameters, *http.Response, error) + /* GetEnterpriseOsdClusters Method for GetEnterpriseOsdClusters @@ -61,8 +91,8 @@ type EnterpriseDataplaneClustersApi interface { RegisterEnterpriseOsdCluster(ctx context.Context) ApiRegisterEnterpriseOsdClusterRequest // RegisterEnterpriseOsdClusterExecute executes the request - // @return EnterpriseClusterRegistrationResponse - RegisterEnterpriseOsdClusterExecute(r ApiRegisterEnterpriseOsdClusterRequest) (*EnterpriseClusterRegistrationResponse, *http.Response, error) + // @return EnterpriseClusterWithAddonParameters + RegisterEnterpriseOsdClusterExecute(r ApiRegisterEnterpriseOsdClusterRequest) (*EnterpriseClusterWithAddonParameters, *http.Response, error) } // EnterpriseDataplaneClustersApiService EnterpriseDataplaneClustersApi service @@ -244,6 +274,298 @@ func (a *EnterpriseDataplaneClustersApiService) DeleteEnterpriseClusterByIdExecu return localVarReturnValue, localVarHTTPResponse, nil } +type ApiGetEnterpriseClusterByIdRequest struct { + ctx context.Context + ApiService EnterpriseDataplaneClustersApi + id string +} + +func (r ApiGetEnterpriseClusterByIdRequest) Execute() (*EnterpriseCluster, *http.Response, error) { + return r.ApiService.GetEnterpriseClusterByIdExecute(r) +} + +/* +GetEnterpriseClusterById Method for GetEnterpriseClusterById + +Returns enterprise data plane cluster by ID + + @param ctx context.Context - for authentication, logging, cancellation, deadlines, tracing, etc. Passed from http.Request or context.Background(). + @param id ID of the enterprise data plane cluster + @return ApiGetEnterpriseClusterByIdRequest +*/ +func (a *EnterpriseDataplaneClustersApiService) GetEnterpriseClusterById(ctx context.Context, id string) ApiGetEnterpriseClusterByIdRequest { + return ApiGetEnterpriseClusterByIdRequest{ + ApiService: a, + ctx: ctx, + id: id, + } +} + +// Execute executes the request +// @return EnterpriseCluster +func (a *EnterpriseDataplaneClustersApiService) GetEnterpriseClusterByIdExecute(r ApiGetEnterpriseClusterByIdRequest) (*EnterpriseCluster, *http.Response, error) { + var ( + localVarHTTPMethod = http.MethodGet + localVarPostBody interface{} + formFiles []formFile + localVarReturnValue *EnterpriseCluster + ) + + localBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, "EnterpriseDataplaneClustersApiService.GetEnterpriseClusterById") + if err != nil { + return localVarReturnValue, nil, &GenericOpenAPIError{error: err.Error()} + } + + localVarPath := localBasePath + "/api/kafkas_mgmt/v1/clusters/{id}" + localVarPath = strings.Replace(localVarPath, "{"+"id"+"}", url.PathEscape(parameterToString(r.id, "")), -1) + + localVarHeaderParams := make(map[string]string) + localVarQueryParams := url.Values{} + localVarFormParams := url.Values{} + + // to determine the Content-Type header + localVarHTTPContentTypes := []string{} + + // set Content-Type header + localVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes) + if localVarHTTPContentType != "" { + localVarHeaderParams["Content-Type"] = localVarHTTPContentType + } + + // to determine the Accept header + localVarHTTPHeaderAccepts := []string{"application/json"} + + // set Accept header + localVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts) + if localVarHTTPHeaderAccept != "" { + localVarHeaderParams["Accept"] = localVarHTTPHeaderAccept + } + req, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, formFiles) + if err != nil { + return localVarReturnValue, nil, err + } + + localVarHTTPResponse, err := a.client.callAPI(req) + if err != nil || localVarHTTPResponse == nil { + return localVarReturnValue, localVarHTTPResponse, err + } + + localVarBody, err := ioutil.ReadAll(localVarHTTPResponse.Body) + localVarHTTPResponse.Body.Close() + localVarHTTPResponse.Body = ioutil.NopCloser(bytes.NewBuffer(localVarBody)) + if err != nil { + return localVarReturnValue, localVarHTTPResponse, err + } + + if localVarHTTPResponse.StatusCode >= 300 { + newErr := &GenericOpenAPIError{ + body: localVarBody, + error: localVarHTTPResponse.Status, + } + if localVarHTTPResponse.StatusCode == 401 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + return localVarReturnValue, localVarHTTPResponse, newErr + } + if localVarHTTPResponse.StatusCode == 403 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + return localVarReturnValue, localVarHTTPResponse, newErr + } + if localVarHTTPResponse.StatusCode == 404 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + return localVarReturnValue, localVarHTTPResponse, newErr + } + if localVarHTTPResponse.StatusCode == 500 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + } + return localVarReturnValue, localVarHTTPResponse, newErr + } + + err = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr := &GenericOpenAPIError{ + body: localVarBody, + error: err.Error(), + } + return localVarReturnValue, localVarHTTPResponse, newErr + } + + return localVarReturnValue, localVarHTTPResponse, nil +} + +type ApiGetEnterpriseClusterWithAddonParametersRequest struct { + ctx context.Context + ApiService EnterpriseDataplaneClustersApi + id string +} + +func (r ApiGetEnterpriseClusterWithAddonParametersRequest) Execute() (*EnterpriseClusterWithAddonParameters, *http.Response, error) { + return r.ApiService.GetEnterpriseClusterWithAddonParametersExecute(r) +} + +/* +GetEnterpriseClusterWithAddonParameters Method for GetEnterpriseClusterWithAddonParameters + +Returns enterprise data plane cluster by ID along with its addon parameters + + @param ctx context.Context - for authentication, logging, cancellation, deadlines, tracing, etc. Passed from http.Request or context.Background(). + @param id ID of the enterprise data plane cluster + @return ApiGetEnterpriseClusterWithAddonParametersRequest +*/ +func (a *EnterpriseDataplaneClustersApiService) GetEnterpriseClusterWithAddonParameters(ctx context.Context, id string) ApiGetEnterpriseClusterWithAddonParametersRequest { + return ApiGetEnterpriseClusterWithAddonParametersRequest{ + ApiService: a, + ctx: ctx, + id: id, + } +} + +// Execute executes the request +// @return EnterpriseClusterWithAddonParameters +func (a *EnterpriseDataplaneClustersApiService) GetEnterpriseClusterWithAddonParametersExecute(r ApiGetEnterpriseClusterWithAddonParametersRequest) (*EnterpriseClusterWithAddonParameters, *http.Response, error) { + var ( + localVarHTTPMethod = http.MethodGet + localVarPostBody interface{} + formFiles []formFile + localVarReturnValue *EnterpriseClusterWithAddonParameters + ) + + localBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, "EnterpriseDataplaneClustersApiService.GetEnterpriseClusterWithAddonParameters") + if err != nil { + return localVarReturnValue, nil, &GenericOpenAPIError{error: err.Error()} + } + + localVarPath := localBasePath + "/api/kafkas_mgmt/v1/clusters/{id}/addon_parameters" + localVarPath = strings.Replace(localVarPath, "{"+"id"+"}", url.PathEscape(parameterToString(r.id, "")), -1) + + localVarHeaderParams := make(map[string]string) + localVarQueryParams := url.Values{} + localVarFormParams := url.Values{} + + // to determine the Content-Type header + localVarHTTPContentTypes := []string{} + + // set Content-Type header + localVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes) + if localVarHTTPContentType != "" { + localVarHeaderParams["Content-Type"] = localVarHTTPContentType + } + + // to determine the Accept header + localVarHTTPHeaderAccepts := []string{"application/json"} + + // set Accept header + localVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts) + if localVarHTTPHeaderAccept != "" { + localVarHeaderParams["Accept"] = localVarHTTPHeaderAccept + } + req, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, formFiles) + if err != nil { + return localVarReturnValue, nil, err + } + + localVarHTTPResponse, err := a.client.callAPI(req) + if err != nil || localVarHTTPResponse == nil { + return localVarReturnValue, localVarHTTPResponse, err + } + + localVarBody, err := ioutil.ReadAll(localVarHTTPResponse.Body) + localVarHTTPResponse.Body.Close() + localVarHTTPResponse.Body = ioutil.NopCloser(bytes.NewBuffer(localVarBody)) + if err != nil { + return localVarReturnValue, localVarHTTPResponse, err + } + + if localVarHTTPResponse.StatusCode >= 300 { + newErr := &GenericOpenAPIError{ + body: localVarBody, + error: localVarHTTPResponse.Status, + } + if localVarHTTPResponse.StatusCode == 401 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + return localVarReturnValue, localVarHTTPResponse, newErr + } + if localVarHTTPResponse.StatusCode == 403 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + return localVarReturnValue, localVarHTTPResponse, newErr + } + if localVarHTTPResponse.StatusCode == 404 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + return localVarReturnValue, localVarHTTPResponse, newErr + } + if localVarHTTPResponse.StatusCode == 500 { + var v Error + err = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr.error = err.Error() + return localVarReturnValue, localVarHTTPResponse, newErr + } + newErr.error = formatErrorMessage(localVarHTTPResponse.Status, &v) + newErr.model = v + } + return localVarReturnValue, localVarHTTPResponse, newErr + } + + err = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get("Content-Type")) + if err != nil { + newErr := &GenericOpenAPIError{ + body: localVarBody, + error: err.Error(), + } + return localVarReturnValue, localVarHTTPResponse, newErr + } + + return localVarReturnValue, localVarHTTPResponse, nil +} + type ApiGetEnterpriseOsdClustersRequest struct { ctx context.Context ApiService EnterpriseDataplaneClustersApi @@ -398,7 +720,7 @@ func (r ApiRegisterEnterpriseOsdClusterRequest) EnterpriseOsdClusterPayload(ente return r } -func (r ApiRegisterEnterpriseOsdClusterRequest) Execute() (*EnterpriseClusterRegistrationResponse, *http.Response, error) { +func (r ApiRegisterEnterpriseOsdClusterRequest) Execute() (*EnterpriseClusterWithAddonParameters, *http.Response, error) { return r.ApiService.RegisterEnterpriseOsdClusterExecute(r) } @@ -418,13 +740,13 @@ func (a *EnterpriseDataplaneClustersApiService) RegisterEnterpriseOsdCluster(ctx } // Execute executes the request -// @return EnterpriseClusterRegistrationResponse -func (a *EnterpriseDataplaneClustersApiService) RegisterEnterpriseOsdClusterExecute(r ApiRegisterEnterpriseOsdClusterRequest) (*EnterpriseClusterRegistrationResponse, *http.Response, error) { +// @return EnterpriseClusterWithAddonParameters +func (a *EnterpriseDataplaneClustersApiService) RegisterEnterpriseOsdClusterExecute(r ApiRegisterEnterpriseOsdClusterRequest) (*EnterpriseClusterWithAddonParameters, *http.Response, error) { var ( localVarHTTPMethod = http.MethodPost localVarPostBody interface{} formFiles []formFile - localVarReturnValue *EnterpriseClusterRegistrationResponse + localVarReturnValue *EnterpriseClusterWithAddonParameters ) localBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, "EnterpriseDataplaneClustersApiService.RegisterEnterpriseOsdCluster") diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterRegistrationResponse.md b/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterWithAddonParameters.md similarity index 60% rename from app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterRegistrationResponse.md rename to app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterWithAddonParameters.md index 6f5f3427..9f561636 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterRegistrationResponse.md +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterWithAddonParameters.md @@ -1,4 +1,4 @@ -# EnterpriseClusterRegistrationResponse +# EnterpriseClusterWithAddonParameters ## Properties @@ -14,175 +14,175 @@ Name | Type | Description | Notes ## Methods -### NewEnterpriseClusterRegistrationResponse +### NewEnterpriseClusterWithAddonParameters -`func NewEnterpriseClusterRegistrationResponse(id string, kind string, href string, accessKafkasViaPrivateNetwork bool, ) *EnterpriseClusterRegistrationResponse` +`func NewEnterpriseClusterWithAddonParameters(id string, kind string, href string, accessKafkasViaPrivateNetwork bool, ) *EnterpriseClusterWithAddonParameters` -NewEnterpriseClusterRegistrationResponse instantiates a new EnterpriseClusterRegistrationResponse object +NewEnterpriseClusterWithAddonParameters instantiates a new EnterpriseClusterWithAddonParameters object This constructor will assign default values to properties that have it defined, and makes sure properties required by API are set, but the set of arguments will change when the set of required properties is changed -### NewEnterpriseClusterRegistrationResponseWithDefaults +### NewEnterpriseClusterWithAddonParametersWithDefaults -`func NewEnterpriseClusterRegistrationResponseWithDefaults() *EnterpriseClusterRegistrationResponse` +`func NewEnterpriseClusterWithAddonParametersWithDefaults() *EnterpriseClusterWithAddonParameters` -NewEnterpriseClusterRegistrationResponseWithDefaults instantiates a new EnterpriseClusterRegistrationResponse object +NewEnterpriseClusterWithAddonParametersWithDefaults instantiates a new EnterpriseClusterWithAddonParameters object This constructor will only assign default values to properties that have it defined, but it doesn't guarantee that properties required by API are set ### GetId -`func (o *EnterpriseClusterRegistrationResponse) GetId() string` +`func (o *EnterpriseClusterWithAddonParameters) GetId() string` GetId returns the Id field if non-nil, zero value otherwise. ### GetIdOk -`func (o *EnterpriseClusterRegistrationResponse) GetIdOk() (*string, bool)` +`func (o *EnterpriseClusterWithAddonParameters) GetIdOk() (*string, bool)` GetIdOk returns a tuple with the Id field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetId -`func (o *EnterpriseClusterRegistrationResponse) SetId(v string)` +`func (o *EnterpriseClusterWithAddonParameters) SetId(v string)` SetId sets Id field to given value. ### GetKind -`func (o *EnterpriseClusterRegistrationResponse) GetKind() string` +`func (o *EnterpriseClusterWithAddonParameters) GetKind() string` GetKind returns the Kind field if non-nil, zero value otherwise. ### GetKindOk -`func (o *EnterpriseClusterRegistrationResponse) GetKindOk() (*string, bool)` +`func (o *EnterpriseClusterWithAddonParameters) GetKindOk() (*string, bool)` GetKindOk returns a tuple with the Kind field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetKind -`func (o *EnterpriseClusterRegistrationResponse) SetKind(v string)` +`func (o *EnterpriseClusterWithAddonParameters) SetKind(v string)` SetKind sets Kind field to given value. ### GetHref -`func (o *EnterpriseClusterRegistrationResponse) GetHref() string` +`func (o *EnterpriseClusterWithAddonParameters) GetHref() string` GetHref returns the Href field if non-nil, zero value otherwise. ### GetHrefOk -`func (o *EnterpriseClusterRegistrationResponse) GetHrefOk() (*string, bool)` +`func (o *EnterpriseClusterWithAddonParameters) GetHrefOk() (*string, bool)` GetHrefOk returns a tuple with the Href field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetHref -`func (o *EnterpriseClusterRegistrationResponse) SetHref(v string)` +`func (o *EnterpriseClusterWithAddonParameters) SetHref(v string)` SetHref sets Href field to given value. ### GetAccessKafkasViaPrivateNetwork -`func (o *EnterpriseClusterRegistrationResponse) GetAccessKafkasViaPrivateNetwork() bool` +`func (o *EnterpriseClusterWithAddonParameters) GetAccessKafkasViaPrivateNetwork() bool` GetAccessKafkasViaPrivateNetwork returns the AccessKafkasViaPrivateNetwork field if non-nil, zero value otherwise. ### GetAccessKafkasViaPrivateNetworkOk -`func (o *EnterpriseClusterRegistrationResponse) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool)` +`func (o *EnterpriseClusterWithAddonParameters) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool)` GetAccessKafkasViaPrivateNetworkOk returns a tuple with the AccessKafkasViaPrivateNetwork field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetAccessKafkasViaPrivateNetwork -`func (o *EnterpriseClusterRegistrationResponse) SetAccessKafkasViaPrivateNetwork(v bool)` +`func (o *EnterpriseClusterWithAddonParameters) SetAccessKafkasViaPrivateNetwork(v bool)` SetAccessKafkasViaPrivateNetwork sets AccessKafkasViaPrivateNetwork field to given value. ### GetClusterId -`func (o *EnterpriseClusterRegistrationResponse) GetClusterId() string` +`func (o *EnterpriseClusterWithAddonParameters) GetClusterId() string` GetClusterId returns the ClusterId field if non-nil, zero value otherwise. ### GetClusterIdOk -`func (o *EnterpriseClusterRegistrationResponse) GetClusterIdOk() (*string, bool)` +`func (o *EnterpriseClusterWithAddonParameters) GetClusterIdOk() (*string, bool)` GetClusterIdOk returns a tuple with the ClusterId field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetClusterId -`func (o *EnterpriseClusterRegistrationResponse) SetClusterId(v string)` +`func (o *EnterpriseClusterWithAddonParameters) SetClusterId(v string)` SetClusterId sets ClusterId field to given value. ### HasClusterId -`func (o *EnterpriseClusterRegistrationResponse) HasClusterId() bool` +`func (o *EnterpriseClusterWithAddonParameters) HasClusterId() bool` HasClusterId returns a boolean if a field has been set. ### GetStatus -`func (o *EnterpriseClusterRegistrationResponse) GetStatus() string` +`func (o *EnterpriseClusterWithAddonParameters) GetStatus() string` GetStatus returns the Status field if non-nil, zero value otherwise. ### GetStatusOk -`func (o *EnterpriseClusterRegistrationResponse) GetStatusOk() (*string, bool)` +`func (o *EnterpriseClusterWithAddonParameters) GetStatusOk() (*string, bool)` GetStatusOk returns a tuple with the Status field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetStatus -`func (o *EnterpriseClusterRegistrationResponse) SetStatus(v string)` +`func (o *EnterpriseClusterWithAddonParameters) SetStatus(v string)` SetStatus sets Status field to given value. ### HasStatus -`func (o *EnterpriseClusterRegistrationResponse) HasStatus() bool` +`func (o *EnterpriseClusterWithAddonParameters) HasStatus() bool` HasStatus returns a boolean if a field has been set. ### GetFleetshardParameters -`func (o *EnterpriseClusterRegistrationResponse) GetFleetshardParameters() []FleetshardParameter` +`func (o *EnterpriseClusterWithAddonParameters) GetFleetshardParameters() []FleetshardParameter` GetFleetshardParameters returns the FleetshardParameters field if non-nil, zero value otherwise. ### GetFleetshardParametersOk -`func (o *EnterpriseClusterRegistrationResponse) GetFleetshardParametersOk() (*[]FleetshardParameter, bool)` +`func (o *EnterpriseClusterWithAddonParameters) GetFleetshardParametersOk() (*[]FleetshardParameter, bool)` GetFleetshardParametersOk returns a tuple with the FleetshardParameters field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetFleetshardParameters -`func (o *EnterpriseClusterRegistrationResponse) SetFleetshardParameters(v []FleetshardParameter)` +`func (o *EnterpriseClusterWithAddonParameters) SetFleetshardParameters(v []FleetshardParameter)` SetFleetshardParameters sets FleetshardParameters field to given value. ### HasFleetshardParameters -`func (o *EnterpriseClusterRegistrationResponse) HasFleetshardParameters() bool` +`func (o *EnterpriseClusterWithAddonParameters) HasFleetshardParameters() bool` HasFleetshardParameters returns a boolean if a field has been set. diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterRegistrationResponseAllOf.md b/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterWithAddonParametersAllOf.md similarity index 59% rename from app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterRegistrationResponseAllOf.md rename to app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterWithAddonParametersAllOf.md index 414b556a..0bba2887 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterRegistrationResponseAllOf.md +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseClusterWithAddonParametersAllOf.md @@ -1,4 +1,4 @@ -# EnterpriseClusterRegistrationResponseAllOf +# EnterpriseClusterWithAddonParametersAllOf ## Properties @@ -11,115 +11,115 @@ Name | Type | Description | Notes ## Methods -### NewEnterpriseClusterRegistrationResponseAllOf +### NewEnterpriseClusterWithAddonParametersAllOf -`func NewEnterpriseClusterRegistrationResponseAllOf(accessKafkasViaPrivateNetwork bool, ) *EnterpriseClusterRegistrationResponseAllOf` +`func NewEnterpriseClusterWithAddonParametersAllOf(accessKafkasViaPrivateNetwork bool, ) *EnterpriseClusterWithAddonParametersAllOf` -NewEnterpriseClusterRegistrationResponseAllOf instantiates a new EnterpriseClusterRegistrationResponseAllOf object +NewEnterpriseClusterWithAddonParametersAllOf instantiates a new EnterpriseClusterWithAddonParametersAllOf object This constructor will assign default values to properties that have it defined, and makes sure properties required by API are set, but the set of arguments will change when the set of required properties is changed -### NewEnterpriseClusterRegistrationResponseAllOfWithDefaults +### NewEnterpriseClusterWithAddonParametersAllOfWithDefaults -`func NewEnterpriseClusterRegistrationResponseAllOfWithDefaults() *EnterpriseClusterRegistrationResponseAllOf` +`func NewEnterpriseClusterWithAddonParametersAllOfWithDefaults() *EnterpriseClusterWithAddonParametersAllOf` -NewEnterpriseClusterRegistrationResponseAllOfWithDefaults instantiates a new EnterpriseClusterRegistrationResponseAllOf object +NewEnterpriseClusterWithAddonParametersAllOfWithDefaults instantiates a new EnterpriseClusterWithAddonParametersAllOf object This constructor will only assign default values to properties that have it defined, but it doesn't guarantee that properties required by API are set ### GetAccessKafkasViaPrivateNetwork -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetAccessKafkasViaPrivateNetwork() bool` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetAccessKafkasViaPrivateNetwork() bool` GetAccessKafkasViaPrivateNetwork returns the AccessKafkasViaPrivateNetwork field if non-nil, zero value otherwise. ### GetAccessKafkasViaPrivateNetworkOk -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool)` GetAccessKafkasViaPrivateNetworkOk returns a tuple with the AccessKafkasViaPrivateNetwork field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetAccessKafkasViaPrivateNetwork -`func (o *EnterpriseClusterRegistrationResponseAllOf) SetAccessKafkasViaPrivateNetwork(v bool)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) SetAccessKafkasViaPrivateNetwork(v bool)` SetAccessKafkasViaPrivateNetwork sets AccessKafkasViaPrivateNetwork field to given value. ### GetClusterId -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetClusterId() string` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetClusterId() string` GetClusterId returns the ClusterId field if non-nil, zero value otherwise. ### GetClusterIdOk -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetClusterIdOk() (*string, bool)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetClusterIdOk() (*string, bool)` GetClusterIdOk returns a tuple with the ClusterId field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetClusterId -`func (o *EnterpriseClusterRegistrationResponseAllOf) SetClusterId(v string)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) SetClusterId(v string)` SetClusterId sets ClusterId field to given value. ### HasClusterId -`func (o *EnterpriseClusterRegistrationResponseAllOf) HasClusterId() bool` +`func (o *EnterpriseClusterWithAddonParametersAllOf) HasClusterId() bool` HasClusterId returns a boolean if a field has been set. ### GetStatus -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetStatus() string` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetStatus() string` GetStatus returns the Status field if non-nil, zero value otherwise. ### GetStatusOk -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetStatusOk() (*string, bool)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetStatusOk() (*string, bool)` GetStatusOk returns a tuple with the Status field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetStatus -`func (o *EnterpriseClusterRegistrationResponseAllOf) SetStatus(v string)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) SetStatus(v string)` SetStatus sets Status field to given value. ### HasStatus -`func (o *EnterpriseClusterRegistrationResponseAllOf) HasStatus() bool` +`func (o *EnterpriseClusterWithAddonParametersAllOf) HasStatus() bool` HasStatus returns a boolean if a field has been set. ### GetFleetshardParameters -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetFleetshardParameters() []FleetshardParameter` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetFleetshardParameters() []FleetshardParameter` GetFleetshardParameters returns the FleetshardParameters field if non-nil, zero value otherwise. ### GetFleetshardParametersOk -`func (o *EnterpriseClusterRegistrationResponseAllOf) GetFleetshardParametersOk() (*[]FleetshardParameter, bool)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) GetFleetshardParametersOk() (*[]FleetshardParameter, bool)` GetFleetshardParametersOk returns a tuple with the FleetshardParameters field if it's non-nil, zero value otherwise and a boolean to check if the value has been set. ### SetFleetshardParameters -`func (o *EnterpriseClusterRegistrationResponseAllOf) SetFleetshardParameters(v []FleetshardParameter)` +`func (o *EnterpriseClusterWithAddonParametersAllOf) SetFleetshardParameters(v []FleetshardParameter)` SetFleetshardParameters sets FleetshardParameters field to given value. ### HasFleetshardParameters -`func (o *EnterpriseClusterRegistrationResponseAllOf) HasFleetshardParameters() bool` +`func (o *EnterpriseClusterWithAddonParametersAllOf) HasFleetshardParameters() bool` HasFleetshardParameters returns a boolean if a field has been set. diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseDataplaneClustersApi.md b/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseDataplaneClustersApi.md index 8f4fcfb2..ded34230 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseDataplaneClustersApi.md +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/docs/EnterpriseDataplaneClustersApi.md @@ -5,6 +5,8 @@ All URIs are relative to *https://api.openshift.com* Method | HTTP request | Description ------------- | ------------- | ------------- [**DeleteEnterpriseClusterById**](EnterpriseDataplaneClustersApi.md#DeleteEnterpriseClusterById) | **Delete** /api/kafkas_mgmt/v1/clusters/{id} | +[**GetEnterpriseClusterById**](EnterpriseDataplaneClustersApi.md#GetEnterpriseClusterById) | **Get** /api/kafkas_mgmt/v1/clusters/{id} | +[**GetEnterpriseClusterWithAddonParameters**](EnterpriseDataplaneClustersApi.md#GetEnterpriseClusterWithAddonParameters) | **Get** /api/kafkas_mgmt/v1/clusters/{id}/addon_parameters | [**GetEnterpriseOsdClusters**](EnterpriseDataplaneClustersApi.md#GetEnterpriseOsdClusters) | **Get** /api/kafkas_mgmt/v1/clusters | [**RegisterEnterpriseOsdCluster**](EnterpriseDataplaneClustersApi.md#RegisterEnterpriseOsdCluster) | **Post** /api/kafkas_mgmt/v1/clusters | @@ -82,6 +84,146 @@ Name | Type | Description | Notes [[Back to README]](../README.md) +## GetEnterpriseClusterById + +> EnterpriseCluster GetEnterpriseClusterById(ctx, id).Execute() + + + + + +### Example + +```go +package main + +import ( + "context" + "fmt" + "os" + openapiclient "./openapi" +) + +func main() { + id := "id_example" // string | ID of the enterprise data plane cluster + + configuration := openapiclient.NewConfiguration() + apiClient := openapiclient.NewAPIClient(configuration) + resp, r, err := apiClient.EnterpriseDataplaneClustersApi.GetEnterpriseClusterById(context.Background(), id).Execute() + if err != nil { + fmt.Fprintf(os.Stderr, "Error when calling `EnterpriseDataplaneClustersApi.GetEnterpriseClusterById``: %v\n", err) + fmt.Fprintf(os.Stderr, "Full HTTP response: %v\n", r) + } + // response from `GetEnterpriseClusterById`: EnterpriseCluster + fmt.Fprintf(os.Stdout, "Response from `EnterpriseDataplaneClustersApi.GetEnterpriseClusterById`: %v\n", resp) +} +``` + +### Path Parameters + + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- +**ctx** | **context.Context** | context for authentication, logging, cancellation, deadlines, tracing, etc. +**id** | **string** | ID of the enterprise data plane cluster | + +### Other Parameters + +Other parameters are passed through a pointer to a apiGetEnterpriseClusterByIdRequest struct via the builder pattern + + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- + + +### Return type + +[**EnterpriseCluster**](EnterpriseCluster.md) + +### Authorization + +[Bearer](../README.md#Bearer) + +### HTTP request headers + +- **Content-Type**: Not defined +- **Accept**: application/json + +[[Back to top]](#) [[Back to API list]](../README.md#documentation-for-api-endpoints) +[[Back to Model list]](../README.md#documentation-for-models) +[[Back to README]](../README.md) + + +## GetEnterpriseClusterWithAddonParameters + +> EnterpriseClusterWithAddonParameters GetEnterpriseClusterWithAddonParameters(ctx, id).Execute() + + + + + +### Example + +```go +package main + +import ( + "context" + "fmt" + "os" + openapiclient "./openapi" +) + +func main() { + id := "id_example" // string | ID of the enterprise data plane cluster + + configuration := openapiclient.NewConfiguration() + apiClient := openapiclient.NewAPIClient(configuration) + resp, r, err := apiClient.EnterpriseDataplaneClustersApi.GetEnterpriseClusterWithAddonParameters(context.Background(), id).Execute() + if err != nil { + fmt.Fprintf(os.Stderr, "Error when calling `EnterpriseDataplaneClustersApi.GetEnterpriseClusterWithAddonParameters``: %v\n", err) + fmt.Fprintf(os.Stderr, "Full HTTP response: %v\n", r) + } + // response from `GetEnterpriseClusterWithAddonParameters`: EnterpriseClusterWithAddonParameters + fmt.Fprintf(os.Stdout, "Response from `EnterpriseDataplaneClustersApi.GetEnterpriseClusterWithAddonParameters`: %v\n", resp) +} +``` + +### Path Parameters + + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- +**ctx** | **context.Context** | context for authentication, logging, cancellation, deadlines, tracing, etc. +**id** | **string** | ID of the enterprise data plane cluster | + +### Other Parameters + +Other parameters are passed through a pointer to a apiGetEnterpriseClusterWithAddonParametersRequest struct via the builder pattern + + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- + + +### Return type + +[**EnterpriseClusterWithAddonParameters**](EnterpriseClusterWithAddonParameters.md) + +### Authorization + +[Bearer](../README.md#Bearer) + +### HTTP request headers + +- **Content-Type**: Not defined +- **Accept**: application/json + +[[Back to top]](#) [[Back to API list]](../README.md#documentation-for-api-endpoints) +[[Back to Model list]](../README.md#documentation-for-models) +[[Back to README]](../README.md) + + ## GetEnterpriseOsdClusters > EnterpriseClusterList GetEnterpriseOsdClusters(ctx).Execute() @@ -145,7 +287,7 @@ Other parameters are passed through a pointer to a apiGetEnterpriseOsdClustersRe ## RegisterEnterpriseOsdCluster -> EnterpriseClusterRegistrationResponse RegisterEnterpriseOsdCluster(ctx).EnterpriseOsdClusterPayload(enterpriseOsdClusterPayload).Execute() +> EnterpriseClusterWithAddonParameters RegisterEnterpriseOsdCluster(ctx).EnterpriseOsdClusterPayload(enterpriseOsdClusterPayload).Execute() @@ -173,7 +315,7 @@ func main() { fmt.Fprintf(os.Stderr, "Error when calling `EnterpriseDataplaneClustersApi.RegisterEnterpriseOsdCluster``: %v\n", err) fmt.Fprintf(os.Stderr, "Full HTTP response: %v\n", r) } - // response from `RegisterEnterpriseOsdCluster`: EnterpriseClusterRegistrationResponse + // response from `RegisterEnterpriseOsdCluster`: EnterpriseClusterWithAddonParameters fmt.Fprintf(os.Stdout, "Response from `EnterpriseDataplaneClustersApi.RegisterEnterpriseOsdCluster`: %v\n", resp) } ``` @@ -193,7 +335,7 @@ Name | Type | Description | Notes ### Return type -[**EnterpriseClusterRegistrationResponse**](EnterpriseClusterRegistrationResponse.md) +[**EnterpriseClusterWithAddonParameters**](EnterpriseClusterWithAddonParameters.md) ### Authorization diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_registration_response.go b/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_with_addon_parameters.go similarity index 61% rename from app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_registration_response.go rename to app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_with_addon_parameters.go index ebf1be13..ff7203a4 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_registration_response.go +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_with_addon_parameters.go @@ -15,8 +15,8 @@ import ( "encoding/json" ) -// EnterpriseClusterRegistrationResponse Enterprise cluster registration endpoint response -type EnterpriseClusterRegistrationResponse struct { +// EnterpriseClusterWithAddonParameters Enterprise cluster with addon parameters +type EnterpriseClusterWithAddonParameters struct { Id string `json:"id"` Kind string `json:"kind"` Href string `json:"href"` @@ -29,12 +29,12 @@ type EnterpriseClusterRegistrationResponse struct { FleetshardParameters []FleetshardParameter `json:"fleetshard_parameters,omitempty"` } -// NewEnterpriseClusterRegistrationResponse instantiates a new EnterpriseClusterRegistrationResponse object +// NewEnterpriseClusterWithAddonParameters instantiates a new EnterpriseClusterWithAddonParameters object // This constructor will assign default values to properties that have it defined, // and makes sure properties required by API are set, but the set of arguments // will change when the set of required properties is changed -func NewEnterpriseClusterRegistrationResponse(id string, kind string, href string, accessKafkasViaPrivateNetwork bool) *EnterpriseClusterRegistrationResponse { - this := EnterpriseClusterRegistrationResponse{} +func NewEnterpriseClusterWithAddonParameters(id string, kind string, href string, accessKafkasViaPrivateNetwork bool) *EnterpriseClusterWithAddonParameters { + this := EnterpriseClusterWithAddonParameters{} this.Id = id this.Kind = kind this.Href = href @@ -42,16 +42,16 @@ func NewEnterpriseClusterRegistrationResponse(id string, kind string, href strin return &this } -// NewEnterpriseClusterRegistrationResponseWithDefaults instantiates a new EnterpriseClusterRegistrationResponse object +// NewEnterpriseClusterWithAddonParametersWithDefaults instantiates a new EnterpriseClusterWithAddonParameters object // This constructor will only assign default values to properties that have it defined, // but it doesn't guarantee that properties required by API are set -func NewEnterpriseClusterRegistrationResponseWithDefaults() *EnterpriseClusterRegistrationResponse { - this := EnterpriseClusterRegistrationResponse{} +func NewEnterpriseClusterWithAddonParametersWithDefaults() *EnterpriseClusterWithAddonParameters { + this := EnterpriseClusterWithAddonParameters{} return &this } // GetId returns the Id field value -func (o *EnterpriseClusterRegistrationResponse) GetId() string { +func (o *EnterpriseClusterWithAddonParameters) GetId() string { if o == nil { var ret string return ret @@ -62,7 +62,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetId() string { // GetIdOk returns a tuple with the Id field value // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponse) GetIdOk() (*string, bool) { +func (o *EnterpriseClusterWithAddonParameters) GetIdOk() (*string, bool) { if o == nil { return nil, false } @@ -70,12 +70,12 @@ func (o *EnterpriseClusterRegistrationResponse) GetIdOk() (*string, bool) { } // SetId sets field value -func (o *EnterpriseClusterRegistrationResponse) SetId(v string) { +func (o *EnterpriseClusterWithAddonParameters) SetId(v string) { o.Id = v } // GetKind returns the Kind field value -func (o *EnterpriseClusterRegistrationResponse) GetKind() string { +func (o *EnterpriseClusterWithAddonParameters) GetKind() string { if o == nil { var ret string return ret @@ -86,7 +86,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetKind() string { // GetKindOk returns a tuple with the Kind field value // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponse) GetKindOk() (*string, bool) { +func (o *EnterpriseClusterWithAddonParameters) GetKindOk() (*string, bool) { if o == nil { return nil, false } @@ -94,12 +94,12 @@ func (o *EnterpriseClusterRegistrationResponse) GetKindOk() (*string, bool) { } // SetKind sets field value -func (o *EnterpriseClusterRegistrationResponse) SetKind(v string) { +func (o *EnterpriseClusterWithAddonParameters) SetKind(v string) { o.Kind = v } // GetHref returns the Href field value -func (o *EnterpriseClusterRegistrationResponse) GetHref() string { +func (o *EnterpriseClusterWithAddonParameters) GetHref() string { if o == nil { var ret string return ret @@ -110,7 +110,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetHref() string { // GetHrefOk returns a tuple with the Href field value // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponse) GetHrefOk() (*string, bool) { +func (o *EnterpriseClusterWithAddonParameters) GetHrefOk() (*string, bool) { if o == nil { return nil, false } @@ -118,12 +118,12 @@ func (o *EnterpriseClusterRegistrationResponse) GetHrefOk() (*string, bool) { } // SetHref sets field value -func (o *EnterpriseClusterRegistrationResponse) SetHref(v string) { +func (o *EnterpriseClusterWithAddonParameters) SetHref(v string) { o.Href = v } // GetAccessKafkasViaPrivateNetwork returns the AccessKafkasViaPrivateNetwork field value -func (o *EnterpriseClusterRegistrationResponse) GetAccessKafkasViaPrivateNetwork() bool { +func (o *EnterpriseClusterWithAddonParameters) GetAccessKafkasViaPrivateNetwork() bool { if o == nil { var ret bool return ret @@ -134,7 +134,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetAccessKafkasViaPrivateNetwork // GetAccessKafkasViaPrivateNetworkOk returns a tuple with the AccessKafkasViaPrivateNetwork field value // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponse) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool) { +func (o *EnterpriseClusterWithAddonParameters) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool) { if o == nil { return nil, false } @@ -142,12 +142,12 @@ func (o *EnterpriseClusterRegistrationResponse) GetAccessKafkasViaPrivateNetwork } // SetAccessKafkasViaPrivateNetwork sets field value -func (o *EnterpriseClusterRegistrationResponse) SetAccessKafkasViaPrivateNetwork(v bool) { +func (o *EnterpriseClusterWithAddonParameters) SetAccessKafkasViaPrivateNetwork(v bool) { o.AccessKafkasViaPrivateNetwork = v } // GetClusterId returns the ClusterId field value if set, zero value otherwise. -func (o *EnterpriseClusterRegistrationResponse) GetClusterId() string { +func (o *EnterpriseClusterWithAddonParameters) GetClusterId() string { if o == nil || isNil(o.ClusterId) { var ret string return ret @@ -157,7 +157,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetClusterId() string { // GetClusterIdOk returns a tuple with the ClusterId field value if set, nil otherwise // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponse) GetClusterIdOk() (*string, bool) { +func (o *EnterpriseClusterWithAddonParameters) GetClusterIdOk() (*string, bool) { if o == nil || isNil(o.ClusterId) { return nil, false } @@ -165,7 +165,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetClusterIdOk() (*string, bool) } // HasClusterId returns a boolean if a field has been set. -func (o *EnterpriseClusterRegistrationResponse) HasClusterId() bool { +func (o *EnterpriseClusterWithAddonParameters) HasClusterId() bool { if o != nil && !isNil(o.ClusterId) { return true } @@ -174,12 +174,12 @@ func (o *EnterpriseClusterRegistrationResponse) HasClusterId() bool { } // SetClusterId gets a reference to the given string and assigns it to the ClusterId field. -func (o *EnterpriseClusterRegistrationResponse) SetClusterId(v string) { +func (o *EnterpriseClusterWithAddonParameters) SetClusterId(v string) { o.ClusterId = &v } // GetStatus returns the Status field value if set, zero value otherwise. -func (o *EnterpriseClusterRegistrationResponse) GetStatus() string { +func (o *EnterpriseClusterWithAddonParameters) GetStatus() string { if o == nil || isNil(o.Status) { var ret string return ret @@ -189,7 +189,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetStatus() string { // GetStatusOk returns a tuple with the Status field value if set, nil otherwise // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponse) GetStatusOk() (*string, bool) { +func (o *EnterpriseClusterWithAddonParameters) GetStatusOk() (*string, bool) { if o == nil || isNil(o.Status) { return nil, false } @@ -197,7 +197,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetStatusOk() (*string, bool) { } // HasStatus returns a boolean if a field has been set. -func (o *EnterpriseClusterRegistrationResponse) HasStatus() bool { +func (o *EnterpriseClusterWithAddonParameters) HasStatus() bool { if o != nil && !isNil(o.Status) { return true } @@ -206,12 +206,12 @@ func (o *EnterpriseClusterRegistrationResponse) HasStatus() bool { } // SetStatus gets a reference to the given string and assigns it to the Status field. -func (o *EnterpriseClusterRegistrationResponse) SetStatus(v string) { +func (o *EnterpriseClusterWithAddonParameters) SetStatus(v string) { o.Status = &v } // GetFleetshardParameters returns the FleetshardParameters field value if set, zero value otherwise. -func (o *EnterpriseClusterRegistrationResponse) GetFleetshardParameters() []FleetshardParameter { +func (o *EnterpriseClusterWithAddonParameters) GetFleetshardParameters() []FleetshardParameter { if o == nil || isNil(o.FleetshardParameters) { var ret []FleetshardParameter return ret @@ -221,7 +221,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetFleetshardParameters() []Flee // GetFleetshardParametersOk returns a tuple with the FleetshardParameters field value if set, nil otherwise // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponse) GetFleetshardParametersOk() ([]FleetshardParameter, bool) { +func (o *EnterpriseClusterWithAddonParameters) GetFleetshardParametersOk() ([]FleetshardParameter, bool) { if o == nil || isNil(o.FleetshardParameters) { return nil, false } @@ -229,7 +229,7 @@ func (o *EnterpriseClusterRegistrationResponse) GetFleetshardParametersOk() ([]F } // HasFleetshardParameters returns a boolean if a field has been set. -func (o *EnterpriseClusterRegistrationResponse) HasFleetshardParameters() bool { +func (o *EnterpriseClusterWithAddonParameters) HasFleetshardParameters() bool { if o != nil && !isNil(o.FleetshardParameters) { return true } @@ -238,11 +238,11 @@ func (o *EnterpriseClusterRegistrationResponse) HasFleetshardParameters() bool { } // SetFleetshardParameters gets a reference to the given []FleetshardParameter and assigns it to the FleetshardParameters field. -func (o *EnterpriseClusterRegistrationResponse) SetFleetshardParameters(v []FleetshardParameter) { +func (o *EnterpriseClusterWithAddonParameters) SetFleetshardParameters(v []FleetshardParameter) { o.FleetshardParameters = v } -func (o EnterpriseClusterRegistrationResponse) MarshalJSON() ([]byte, error) { +func (o EnterpriseClusterWithAddonParameters) MarshalJSON() ([]byte, error) { toSerialize := map[string]interface{}{} if true { toSerialize["id"] = o.Id @@ -268,38 +268,38 @@ func (o EnterpriseClusterRegistrationResponse) MarshalJSON() ([]byte, error) { return json.Marshal(toSerialize) } -type NullableEnterpriseClusterRegistrationResponse struct { - value *EnterpriseClusterRegistrationResponse +type NullableEnterpriseClusterWithAddonParameters struct { + value *EnterpriseClusterWithAddonParameters isSet bool } -func (v NullableEnterpriseClusterRegistrationResponse) Get() *EnterpriseClusterRegistrationResponse { +func (v NullableEnterpriseClusterWithAddonParameters) Get() *EnterpriseClusterWithAddonParameters { return v.value } -func (v *NullableEnterpriseClusterRegistrationResponse) Set(val *EnterpriseClusterRegistrationResponse) { +func (v *NullableEnterpriseClusterWithAddonParameters) Set(val *EnterpriseClusterWithAddonParameters) { v.value = val v.isSet = true } -func (v NullableEnterpriseClusterRegistrationResponse) IsSet() bool { +func (v NullableEnterpriseClusterWithAddonParameters) IsSet() bool { return v.isSet } -func (v *NullableEnterpriseClusterRegistrationResponse) Unset() { +func (v *NullableEnterpriseClusterWithAddonParameters) Unset() { v.value = nil v.isSet = false } -func NewNullableEnterpriseClusterRegistrationResponse(val *EnterpriseClusterRegistrationResponse) *NullableEnterpriseClusterRegistrationResponse { - return &NullableEnterpriseClusterRegistrationResponse{value: val, isSet: true} +func NewNullableEnterpriseClusterWithAddonParameters(val *EnterpriseClusterWithAddonParameters) *NullableEnterpriseClusterWithAddonParameters { + return &NullableEnterpriseClusterWithAddonParameters{value: val, isSet: true} } -func (v NullableEnterpriseClusterRegistrationResponse) MarshalJSON() ([]byte, error) { +func (v NullableEnterpriseClusterWithAddonParameters) MarshalJSON() ([]byte, error) { return json.Marshal(v.value) } -func (v *NullableEnterpriseClusterRegistrationResponse) UnmarshalJSON(src []byte) error { +func (v *NullableEnterpriseClusterWithAddonParameters) UnmarshalJSON(src []byte) error { v.isSet = true return json.Unmarshal(src, &v.value) } diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_registration_response_all_of.go b/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_with_addon_parameters_all_of.go similarity index 59% rename from app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_registration_response_all_of.go rename to app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_with_addon_parameters_all_of.go index 8f0d8a71..c7e46795 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_registration_response_all_of.go +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/model_enterprise_cluster_with_addon_parameters_all_of.go @@ -15,8 +15,8 @@ import ( "encoding/json" ) -// EnterpriseClusterRegistrationResponseAllOf struct for EnterpriseClusterRegistrationResponseAllOf -type EnterpriseClusterRegistrationResponseAllOf struct { +// EnterpriseClusterWithAddonParametersAllOf struct for EnterpriseClusterWithAddonParametersAllOf +type EnterpriseClusterWithAddonParametersAllOf struct { // Indicates whether Kafkas created on this data plane cluster have to be accessed via private network AccessKafkasViaPrivateNetwork bool `json:"access_kafkas_via_private_network"` // OCM cluster id of the registered Enterprise cluster @@ -26,26 +26,26 @@ type EnterpriseClusterRegistrationResponseAllOf struct { FleetshardParameters []FleetshardParameter `json:"fleetshard_parameters,omitempty"` } -// NewEnterpriseClusterRegistrationResponseAllOf instantiates a new EnterpriseClusterRegistrationResponseAllOf object +// NewEnterpriseClusterWithAddonParametersAllOf instantiates a new EnterpriseClusterWithAddonParametersAllOf object // This constructor will assign default values to properties that have it defined, // and makes sure properties required by API are set, but the set of arguments // will change when the set of required properties is changed -func NewEnterpriseClusterRegistrationResponseAllOf(accessKafkasViaPrivateNetwork bool) *EnterpriseClusterRegistrationResponseAllOf { - this := EnterpriseClusterRegistrationResponseAllOf{} +func NewEnterpriseClusterWithAddonParametersAllOf(accessKafkasViaPrivateNetwork bool) *EnterpriseClusterWithAddonParametersAllOf { + this := EnterpriseClusterWithAddonParametersAllOf{} this.AccessKafkasViaPrivateNetwork = accessKafkasViaPrivateNetwork return &this } -// NewEnterpriseClusterRegistrationResponseAllOfWithDefaults instantiates a new EnterpriseClusterRegistrationResponseAllOf object +// NewEnterpriseClusterWithAddonParametersAllOfWithDefaults instantiates a new EnterpriseClusterWithAddonParametersAllOf object // This constructor will only assign default values to properties that have it defined, // but it doesn't guarantee that properties required by API are set -func NewEnterpriseClusterRegistrationResponseAllOfWithDefaults() *EnterpriseClusterRegistrationResponseAllOf { - this := EnterpriseClusterRegistrationResponseAllOf{} +func NewEnterpriseClusterWithAddonParametersAllOfWithDefaults() *EnterpriseClusterWithAddonParametersAllOf { + this := EnterpriseClusterWithAddonParametersAllOf{} return &this } // GetAccessKafkasViaPrivateNetwork returns the AccessKafkasViaPrivateNetwork field value -func (o *EnterpriseClusterRegistrationResponseAllOf) GetAccessKafkasViaPrivateNetwork() bool { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetAccessKafkasViaPrivateNetwork() bool { if o == nil { var ret bool return ret @@ -56,7 +56,7 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetAccessKafkasViaPrivateNe // GetAccessKafkasViaPrivateNetworkOk returns a tuple with the AccessKafkasViaPrivateNetwork field value // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponseAllOf) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool) { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetAccessKafkasViaPrivateNetworkOk() (*bool, bool) { if o == nil { return nil, false } @@ -64,12 +64,12 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetAccessKafkasViaPrivateNe } // SetAccessKafkasViaPrivateNetwork sets field value -func (o *EnterpriseClusterRegistrationResponseAllOf) SetAccessKafkasViaPrivateNetwork(v bool) { +func (o *EnterpriseClusterWithAddonParametersAllOf) SetAccessKafkasViaPrivateNetwork(v bool) { o.AccessKafkasViaPrivateNetwork = v } // GetClusterId returns the ClusterId field value if set, zero value otherwise. -func (o *EnterpriseClusterRegistrationResponseAllOf) GetClusterId() string { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetClusterId() string { if o == nil || isNil(o.ClusterId) { var ret string return ret @@ -79,7 +79,7 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetClusterId() string { // GetClusterIdOk returns a tuple with the ClusterId field value if set, nil otherwise // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponseAllOf) GetClusterIdOk() (*string, bool) { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetClusterIdOk() (*string, bool) { if o == nil || isNil(o.ClusterId) { return nil, false } @@ -87,7 +87,7 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetClusterIdOk() (*string, } // HasClusterId returns a boolean if a field has been set. -func (o *EnterpriseClusterRegistrationResponseAllOf) HasClusterId() bool { +func (o *EnterpriseClusterWithAddonParametersAllOf) HasClusterId() bool { if o != nil && !isNil(o.ClusterId) { return true } @@ -96,12 +96,12 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) HasClusterId() bool { } // SetClusterId gets a reference to the given string and assigns it to the ClusterId field. -func (o *EnterpriseClusterRegistrationResponseAllOf) SetClusterId(v string) { +func (o *EnterpriseClusterWithAddonParametersAllOf) SetClusterId(v string) { o.ClusterId = &v } // GetStatus returns the Status field value if set, zero value otherwise. -func (o *EnterpriseClusterRegistrationResponseAllOf) GetStatus() string { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetStatus() string { if o == nil || isNil(o.Status) { var ret string return ret @@ -111,7 +111,7 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetStatus() string { // GetStatusOk returns a tuple with the Status field value if set, nil otherwise // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponseAllOf) GetStatusOk() (*string, bool) { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetStatusOk() (*string, bool) { if o == nil || isNil(o.Status) { return nil, false } @@ -119,7 +119,7 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetStatusOk() (*string, boo } // HasStatus returns a boolean if a field has been set. -func (o *EnterpriseClusterRegistrationResponseAllOf) HasStatus() bool { +func (o *EnterpriseClusterWithAddonParametersAllOf) HasStatus() bool { if o != nil && !isNil(o.Status) { return true } @@ -128,12 +128,12 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) HasStatus() bool { } // SetStatus gets a reference to the given string and assigns it to the Status field. -func (o *EnterpriseClusterRegistrationResponseAllOf) SetStatus(v string) { +func (o *EnterpriseClusterWithAddonParametersAllOf) SetStatus(v string) { o.Status = &v } // GetFleetshardParameters returns the FleetshardParameters field value if set, zero value otherwise. -func (o *EnterpriseClusterRegistrationResponseAllOf) GetFleetshardParameters() []FleetshardParameter { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetFleetshardParameters() []FleetshardParameter { if o == nil || isNil(o.FleetshardParameters) { var ret []FleetshardParameter return ret @@ -143,7 +143,7 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetFleetshardParameters() [ // GetFleetshardParametersOk returns a tuple with the FleetshardParameters field value if set, nil otherwise // and a boolean to check if the value has been set. -func (o *EnterpriseClusterRegistrationResponseAllOf) GetFleetshardParametersOk() ([]FleetshardParameter, bool) { +func (o *EnterpriseClusterWithAddonParametersAllOf) GetFleetshardParametersOk() ([]FleetshardParameter, bool) { if o == nil || isNil(o.FleetshardParameters) { return nil, false } @@ -151,7 +151,7 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) GetFleetshardParametersOk() } // HasFleetshardParameters returns a boolean if a field has been set. -func (o *EnterpriseClusterRegistrationResponseAllOf) HasFleetshardParameters() bool { +func (o *EnterpriseClusterWithAddonParametersAllOf) HasFleetshardParameters() bool { if o != nil && !isNil(o.FleetshardParameters) { return true } @@ -160,11 +160,11 @@ func (o *EnterpriseClusterRegistrationResponseAllOf) HasFleetshardParameters() b } // SetFleetshardParameters gets a reference to the given []FleetshardParameter and assigns it to the FleetshardParameters field. -func (o *EnterpriseClusterRegistrationResponseAllOf) SetFleetshardParameters(v []FleetshardParameter) { +func (o *EnterpriseClusterWithAddonParametersAllOf) SetFleetshardParameters(v []FleetshardParameter) { o.FleetshardParameters = v } -func (o EnterpriseClusterRegistrationResponseAllOf) MarshalJSON() ([]byte, error) { +func (o EnterpriseClusterWithAddonParametersAllOf) MarshalJSON() ([]byte, error) { toSerialize := map[string]interface{}{} if true { toSerialize["access_kafkas_via_private_network"] = o.AccessKafkasViaPrivateNetwork @@ -181,38 +181,38 @@ func (o EnterpriseClusterRegistrationResponseAllOf) MarshalJSON() ([]byte, error return json.Marshal(toSerialize) } -type NullableEnterpriseClusterRegistrationResponseAllOf struct { - value *EnterpriseClusterRegistrationResponseAllOf +type NullableEnterpriseClusterWithAddonParametersAllOf struct { + value *EnterpriseClusterWithAddonParametersAllOf isSet bool } -func (v NullableEnterpriseClusterRegistrationResponseAllOf) Get() *EnterpriseClusterRegistrationResponseAllOf { +func (v NullableEnterpriseClusterWithAddonParametersAllOf) Get() *EnterpriseClusterWithAddonParametersAllOf { return v.value } -func (v *NullableEnterpriseClusterRegistrationResponseAllOf) Set(val *EnterpriseClusterRegistrationResponseAllOf) { +func (v *NullableEnterpriseClusterWithAddonParametersAllOf) Set(val *EnterpriseClusterWithAddonParametersAllOf) { v.value = val v.isSet = true } -func (v NullableEnterpriseClusterRegistrationResponseAllOf) IsSet() bool { +func (v NullableEnterpriseClusterWithAddonParametersAllOf) IsSet() bool { return v.isSet } -func (v *NullableEnterpriseClusterRegistrationResponseAllOf) Unset() { +func (v *NullableEnterpriseClusterWithAddonParametersAllOf) Unset() { v.value = nil v.isSet = false } -func NewNullableEnterpriseClusterRegistrationResponseAllOf(val *EnterpriseClusterRegistrationResponseAllOf) *NullableEnterpriseClusterRegistrationResponseAllOf { - return &NullableEnterpriseClusterRegistrationResponseAllOf{value: val, isSet: true} +func NewNullableEnterpriseClusterWithAddonParametersAllOf(val *EnterpriseClusterWithAddonParametersAllOf) *NullableEnterpriseClusterWithAddonParametersAllOf { + return &NullableEnterpriseClusterWithAddonParametersAllOf{value: val, isSet: true} } -func (v NullableEnterpriseClusterRegistrationResponseAllOf) MarshalJSON() ([]byte, error) { +func (v NullableEnterpriseClusterWithAddonParametersAllOf) MarshalJSON() ([]byte, error) { return json.Marshal(v.value) } -func (v *NullableEnterpriseClusterRegistrationResponseAllOf) UnmarshalJSON(src []byte) error { +func (v *NullableEnterpriseClusterWithAddonParametersAllOf) UnmarshalJSON(src []byte) error { v.isSet = true return json.Unmarshal(src, &v.value) } diff --git a/app-services-sdk-go/kafkamgmt/apiv1/client/test/api_enterprise_dataplane_clusters_test.go b/app-services-sdk-go/kafkamgmt/apiv1/client/test/api_enterprise_dataplane_clusters_test.go index b9e44044..490d9408 100644 --- a/app-services-sdk-go/kafkamgmt/apiv1/client/test/api_enterprise_dataplane_clusters_test.go +++ b/app-services-sdk-go/kafkamgmt/apiv1/client/test/api_enterprise_dataplane_clusters_test.go @@ -36,6 +36,34 @@ func Test_kafkamgmtclient_EnterpriseDataplaneClustersApiService(t *testing.T) { }) + t.Run("Test EnterpriseDataplaneClustersApiService GetEnterpriseClusterById", func(t *testing.T) { + + t.Skip("skip test") // remove to run test + + var id string + + resp, httpRes, err := apiClient.EnterpriseDataplaneClustersApi.GetEnterpriseClusterById(context.Background(), id).Execute() + + require.Nil(t, err) + require.NotNil(t, resp) + assert.Equal(t, 200, httpRes.StatusCode) + + }) + + t.Run("Test EnterpriseDataplaneClustersApiService GetEnterpriseClusterWithAddonParameters", func(t *testing.T) { + + t.Skip("skip test") // remove to run test + + var id string + + resp, httpRes, err := apiClient.EnterpriseDataplaneClustersApi.GetEnterpriseClusterWithAddonParameters(context.Background(), id).Execute() + + require.Nil(t, err) + require.NotNil(t, resp) + assert.Equal(t, 200, httpRes.StatusCode) + + }) + t.Run("Test EnterpriseDataplaneClustersApiService GetEnterpriseOsdClusters", func(t *testing.T) { t.Skip("skip test") // remove to run test diff --git a/app-services-sdk-java/packages/kafka-management-sdk/.openapi-generator/FILES b/app-services-sdk-java/packages/kafka-management-sdk/.openapi-generator/FILES index 52dd802c..8ae02544 100644 --- a/app-services-sdk-java/packages/kafka-management-sdk/.openapi-generator/FILES +++ b/app-services-sdk-java/packages/kafka-management-sdk/.openapi-generator/FILES @@ -10,8 +10,8 @@ docs/EnterpriseCluster.md docs/EnterpriseClusterAllOf.md docs/EnterpriseClusterList.md docs/EnterpriseClusterListAllOf.md -docs/EnterpriseClusterRegistrationResponse.md -docs/EnterpriseClusterRegistrationResponseAllOf.md +docs/EnterpriseClusterWithAddonParameters.md +docs/EnterpriseClusterWithAddonParametersAllOf.md docs/EnterpriseDataplaneClustersApi.md docs/EnterpriseOsdClusterPayload.md docs/Error.md @@ -83,8 +83,8 @@ src/main/java/com/openshift/cloud/api/kas/models/EnterpriseCluster.java src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterAllOf.java src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterList.java src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterListAllOf.java -src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponse.java -src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponseAllOf.java +src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParameters.java +src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParametersAllOf.java src/main/java/com/openshift/cloud/api/kas/models/EnterpriseOsdClusterPayload.java src/main/java/com/openshift/cloud/api/kas/models/Error.java src/main/java/com/openshift/cloud/api/kas/models/ErrorList.java diff --git a/app-services-sdk-java/packages/kafka-management-sdk/api/openapi.yaml b/app-services-sdk-java/packages/kafka-management-sdk/api/openapi.yaml index 3935b42c..ed8e5810 100644 --- a/app-services-sdk-java/packages/kafka-management-sdk/api/openapi.yaml +++ b/app-services-sdk-java/packages/kafka-management-sdk/api/openapi.yaml @@ -1271,7 +1271,7 @@ paths: content: application/json: schema: - $ref: '#/components/schemas/EnterpriseClusterRegistrationResponse' + $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters' description: Enterprise cluster registered "400": content: @@ -1413,6 +1413,135 @@ paths: tags: - enterprise-dataplane-clusters x-accepts: application/json + get: + description: Returns enterprise data plane cluster by ID + operationId: getEnterpriseClusterById + parameters: + - description: ID of the enterprise data plane cluster + explode: false + in: path + name: id + required: true + schema: + type: string + style: simple + responses: + "200": + content: + application/json: + examples: + EnterpriseClusterExample: + $ref: '#/components/examples/EnterpriseClusterExample' + schema: + $ref: '#/components/schemas/EnterpriseCluster' + description: Found enterprise data plane cluster with provided ID + "401": + content: + application/json: + examples: + "401Example": + $ref: '#/components/examples/401Example' + schema: + $ref: '#/components/schemas/Error' + description: Auth token is invalid + "403": + content: + application/json: + examples: + "403Example": + $ref: '#/components/examples/403Example' + schema: + $ref: '#/components/schemas/Error' + description: User not authorized to access the service + "404": + content: + application/json: + examples: + "404Example": + $ref: '#/components/examples/404Example' + schema: + $ref: '#/components/schemas/Error' + description: No Enterprise data plane cluster with specified ID exists + "500": + content: + application/json: + examples: + "500Example": + $ref: '#/components/examples/500Example' + schema: + $ref: '#/components/schemas/Error' + description: Unexpected error occurred + security: + - Bearer: [] + tags: + - enterprise-dataplane-clusters + x-accepts: application/json + /api/kafkas_mgmt/v1/clusters/{id}/addon_parameters: + get: + description: Returns enterprise data plane cluster by ID along with its addon + parameters + operationId: getEnterpriseClusterWithAddonParameters + parameters: + - description: ID of the enterprise data plane cluster + explode: false + in: path + name: id + required: true + schema: + type: string + style: simple + responses: + "200": + content: + application/json: + examples: + EnterpriseClusterWithAddonsExample: + $ref: '#/components/examples/EnterpriseClusterWithAddonParametersExample' + schema: + $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters' + description: Returns the enterprise data plane cluster with addon parameters + for an existing enterprise data plane cluster with provided ID + "401": + content: + application/json: + examples: + "401Example": + $ref: '#/components/examples/401Example' + schema: + $ref: '#/components/schemas/Error' + description: Auth token is invalid + "403": + content: + application/json: + examples: + "403Example": + $ref: '#/components/examples/403Example' + schema: + $ref: '#/components/schemas/Error' + description: User not authorized to access the service + "404": + content: + application/json: + examples: + "404Example": + $ref: '#/components/examples/404Example' + schema: + $ref: '#/components/schemas/Error' + description: No Enterprise data plane cluster with specified ID exists + "500": + content: + application/json: + examples: + "500Example": + $ref: '#/components/examples/500Example' + schema: + $ref: '#/components/schemas/Error' + description: Unexpected error occurred + security: + - Bearer: [] + tags: + - enterprise-dataplane-clusters + x-accepts: application/json components: examples: USRegionExample: @@ -1432,6 +1561,13 @@ components: name: test_kafka plan: standard.x1 EnterpriseClusterExample: + value: + id: abcd1234ascd3456fdks9485lskd030h + kind: cluster + href: /api/kafkas_mgmr/v1/clusters/abcd1234ascd3456fdks9485lskd030h + status: ready + cluster_id: abcd1234ascd3456fdks9485lskd030h + EnterpriseClusterWithAddonParametersExample: value: id: abcd1234ascd3456fdks9485lskd030h kind: cluster @@ -1439,6 +1575,19 @@ components: status: ready cluster_id: abcd1234ascd3456fdks9485lskd030h access_kafkas_via_private_network: true + fleetshard_parameters: + - id: sso-auth-server-url + value: http://someurl/auth/realms/rhoas + - id: sso-client-id + value: kas-fleetshard-agent-1234abcd1234abcd1234abcd1234abcd + - id: sso-secret + value: some-secret + - id: cluster-id + value: 1234abcd1234abcd1234abcd1234abcd + - id: poll-interval + value: 15s + - id: resync-interval + value: 60s KafkaRequestExample: value: id: 1iSY6RQ3JKI8Q0OTmjQFd3ocFRg @@ -2402,11 +2551,11 @@ components: - cluster_ingress_dns_name - kafka_machine_pool_node_count type: object - EnterpriseClusterRegistrationResponse: + EnterpriseClusterWithAddonParameters: allOf: - $ref: '#/components/schemas/ObjectReference' - - $ref: '#/components/schemas/EnterpriseClusterRegistrationResponse_allOf' - description: Enterprise cluster registration endpoint response + - $ref: '#/components/schemas/EnterpriseClusterWithAddonParameters_allOf' + description: Enterprise cluster with addon parameters FleetshardParameter: description: Fleetshard parameter consumed by enterprise cluster properties: @@ -2700,7 +2849,7 @@ components: - $ref: '#/components/schemas/InstantQuery' type: array type: object - EnterpriseClusterRegistrationResponse_allOf: + EnterpriseClusterWithAddonParameters_allOf: properties: access_kafkas_via_private_network: description: Indicates whether Kafkas created on this data plane cluster diff --git a/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseClusterWithAddonParameters.md b/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseClusterWithAddonParameters.md new file mode 100644 index 00000000..d9ab02f2 --- /dev/null +++ b/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseClusterWithAddonParameters.md @@ -0,0 +1,20 @@ + + +# EnterpriseClusterWithAddonParameters + +Enterprise cluster with addon parameters + +## Properties + +Name | Type | Description | Notes +------------ | ------------- | ------------- | ------------- +**id** | **String** | | +**kind** | **String** | | +**href** | **String** | | +**accessKafkasViaPrivateNetwork** | **Boolean** | Indicates whether Kafkas created on this data plane cluster have to be accessed via private network | +**clusterId** | **String** | OCM cluster id of the registered Enterprise cluster | [optional] +**status** | **String** | status of registered Enterprise cluster | [optional] +**fleetshardParameters** | [**List<FleetshardParameter>**](FleetshardParameter.md) | | [optional] + + + diff --git a/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseClusterWithAddonParametersAllOf.md b/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseClusterWithAddonParametersAllOf.md new file mode 100644 index 00000000..2277e444 --- /dev/null +++ b/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseClusterWithAddonParametersAllOf.md @@ -0,0 +1,16 @@ + + +# EnterpriseClusterWithAddonParametersAllOf + + +## Properties + +Name | Type | Description | Notes +------------ | ------------- | ------------- | ------------- +**accessKafkasViaPrivateNetwork** | **Boolean** | Indicates whether Kafkas created on this data plane cluster have to be accessed via private network | +**clusterId** | **String** | OCM cluster id of the registered Enterprise cluster | [optional] +**status** | **String** | status of registered Enterprise cluster | [optional] +**fleetshardParameters** | [**List<FleetshardParameter>**](FleetshardParameter.md) | | [optional] + + + diff --git a/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseDataplaneClustersApi.md b/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseDataplaneClustersApi.md index bd28426b..3dccc7bd 100644 --- a/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseDataplaneClustersApi.md +++ b/app-services-sdk-java/packages/kafka-management-sdk/docs/EnterpriseDataplaneClustersApi.md @@ -5,6 +5,8 @@ All URIs are relative to *https://api.openshift.com* Method | HTTP request | Description ------------- | ------------- | ------------- [**deleteEnterpriseClusterById**](EnterpriseDataplaneClustersApi.md#deleteEnterpriseClusterById) | **DELETE** /api/kafkas_mgmt/v1/clusters/{id} | +[**getEnterpriseClusterById**](EnterpriseDataplaneClustersApi.md#getEnterpriseClusterById) | **GET** /api/kafkas_mgmt/v1/clusters/{id} | +[**getEnterpriseClusterWithAddonParameters**](EnterpriseDataplaneClustersApi.md#getEnterpriseClusterWithAddonParameters) | **GET** /api/kafkas_mgmt/v1/clusters/{id}/addon_parameters | [**getEnterpriseOsdClusters**](EnterpriseDataplaneClustersApi.md#getEnterpriseOsdClusters) | **GET** /api/kafkas_mgmt/v1/clusters | [**registerEnterpriseOsdCluster**](EnterpriseDataplaneClustersApi.md#registerEnterpriseOsdCluster) | **POST** /api/kafkas_mgmt/v1/clusters | @@ -88,6 +90,156 @@ Name | Type | Description | Notes | **500** | Unexpected error occurred | - | +## getEnterpriseClusterById + +> EnterpriseCluster getEnterpriseClusterById(id) + + + +Returns enterprise data plane cluster by ID + +### Example + +```java +// Import classes: +import com.openshift.cloud.api.kas.invoker.ApiClient; +import com.openshift.cloud.api.kas.invoker.ApiException; +import com.openshift.cloud.api.kas.invoker.Configuration; +import com.openshift.cloud.api.kas.invoker.auth.*; +import com.openshift.cloud.api.kas.invoker.models.*; +import com.openshift.cloud.api.kas.EnterpriseDataplaneClustersApi; + +public class Example { + public static void main(String[] args) { + ApiClient defaultClient = Configuration.getDefaultApiClient(); + defaultClient.setBasePath("https://api.openshift.com"); + + // Configure HTTP bearer authorization: Bearer + HttpBearerAuth Bearer = (HttpBearerAuth) defaultClient.getAuthentication("Bearer"); + Bearer.setBearerToken("BEARER TOKEN"); + + EnterpriseDataplaneClustersApi apiInstance = new EnterpriseDataplaneClustersApi(defaultClient); + String id = "id_example"; // String | ID of the enterprise data plane cluster + try { + EnterpriseCluster result = apiInstance.getEnterpriseClusterById(id); + System.out.println(result); + } catch (ApiException e) { + System.err.println("Exception when calling EnterpriseDataplaneClustersApi#getEnterpriseClusterById"); + System.err.println("Status code: " + e.getCode()); + System.err.println("Reason: " + e.getResponseBody()); + System.err.println("Response headers: " + e.getResponseHeaders()); + e.printStackTrace(); + } + } +} +``` + +### Parameters + + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- + **id** | **String**| ID of the enterprise data plane cluster | + +### Return type + +[**EnterpriseCluster**](EnterpriseCluster.md) + +### Authorization + +[Bearer](../README.md#Bearer) + +### HTTP request headers + +- **Content-Type**: Not defined +- **Accept**: application/json + + +### HTTP response details +| Status code | Description | Response headers | +|-------------|-------------|------------------| +| **200** | Found enterprise data plane cluster with provided ID | - | +| **401** | Auth token is invalid | - | +| **403** | User not authorized to access the service | - | +| **404** | No Enterprise data plane cluster with specified ID exists | - | +| **500** | Unexpected error occurred | - | + + +## getEnterpriseClusterWithAddonParameters + +> EnterpriseClusterWithAddonParameters getEnterpriseClusterWithAddonParameters(id) + + + +Returns enterprise data plane cluster by ID along with its addon parameters + +### Example + +```java +// Import classes: +import com.openshift.cloud.api.kas.invoker.ApiClient; +import com.openshift.cloud.api.kas.invoker.ApiException; +import com.openshift.cloud.api.kas.invoker.Configuration; +import com.openshift.cloud.api.kas.invoker.auth.*; +import com.openshift.cloud.api.kas.invoker.models.*; +import com.openshift.cloud.api.kas.EnterpriseDataplaneClustersApi; + +public class Example { + public static void main(String[] args) { + ApiClient defaultClient = Configuration.getDefaultApiClient(); + defaultClient.setBasePath("https://api.openshift.com"); + + // Configure HTTP bearer authorization: Bearer + HttpBearerAuth Bearer = (HttpBearerAuth) defaultClient.getAuthentication("Bearer"); + Bearer.setBearerToken("BEARER TOKEN"); + + EnterpriseDataplaneClustersApi apiInstance = new EnterpriseDataplaneClustersApi(defaultClient); + String id = "id_example"; // String | ID of the enterprise data plane cluster + try { + EnterpriseClusterWithAddonParameters result = apiInstance.getEnterpriseClusterWithAddonParameters(id); + System.out.println(result); + } catch (ApiException e) { + System.err.println("Exception when calling EnterpriseDataplaneClustersApi#getEnterpriseClusterWithAddonParameters"); + System.err.println("Status code: " + e.getCode()); + System.err.println("Reason: " + e.getResponseBody()); + System.err.println("Response headers: " + e.getResponseHeaders()); + e.printStackTrace(); + } + } +} +``` + +### Parameters + + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- + **id** | **String**| ID of the enterprise data plane cluster | + +### Return type + +[**EnterpriseClusterWithAddonParameters**](EnterpriseClusterWithAddonParameters.md) + +### Authorization + +[Bearer](../README.md#Bearer) + +### HTTP request headers + +- **Content-Type**: Not defined +- **Accept**: application/json + + +### HTTP response details +| Status code | Description | Response headers | +|-------------|-------------|------------------| +| **200** | Returns the enterprise data plane cluster with addon parameters for an existing enterprise data plane cluster with provided ID | - | +| **401** | Auth token is invalid | - | +| **403** | User not authorized to access the service | - | +| **404** | No Enterprise data plane cluster with specified ID exists | - | +| **500** | Unexpected error occurred | - | + + ## getEnterpriseOsdClusters > EnterpriseClusterList getEnterpriseOsdClusters() @@ -161,7 +313,7 @@ This endpoint does not need any parameter. ## registerEnterpriseOsdCluster -> EnterpriseClusterRegistrationResponse registerEnterpriseOsdCluster(enterpriseOsdClusterPayload) +> EnterpriseClusterWithAddonParameters registerEnterpriseOsdCluster(enterpriseOsdClusterPayload) @@ -190,7 +342,7 @@ public class Example { EnterpriseDataplaneClustersApi apiInstance = new EnterpriseDataplaneClustersApi(defaultClient); EnterpriseOsdClusterPayload enterpriseOsdClusterPayload = new EnterpriseOsdClusterPayload(); // EnterpriseOsdClusterPayload | Enterprise data plane cluster details try { - EnterpriseClusterRegistrationResponse result = apiInstance.registerEnterpriseOsdCluster(enterpriseOsdClusterPayload); + EnterpriseClusterWithAddonParameters result = apiInstance.registerEnterpriseOsdCluster(enterpriseOsdClusterPayload); System.out.println(result); } catch (ApiException e) { System.err.println("Exception when calling EnterpriseDataplaneClustersApi#registerEnterpriseOsdCluster"); @@ -212,7 +364,7 @@ Name | Type | Description | Notes ### Return type -[**EnterpriseClusterRegistrationResponse**](EnterpriseClusterRegistrationResponse.md) +[**EnterpriseClusterWithAddonParameters**](EnterpriseClusterWithAddonParameters.md) ### Authorization diff --git a/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/EnterpriseDataplaneClustersApi.java b/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/EnterpriseDataplaneClustersApi.java index f9f2e7c3..adf9d517 100644 --- a/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/EnterpriseDataplaneClustersApi.java +++ b/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/EnterpriseDataplaneClustersApi.java @@ -7,8 +7,9 @@ import javax.ws.rs.core.GenericType; +import com.openshift.cloud.api.kas.models.EnterpriseCluster; import com.openshift.cloud.api.kas.models.EnterpriseClusterList; -import com.openshift.cloud.api.kas.models.EnterpriseClusterRegistrationResponse; +import com.openshift.cloud.api.kas.models.EnterpriseClusterWithAddonParameters; import com.openshift.cloud.api.kas.models.EnterpriseOsdClusterPayload; import com.openshift.cloud.api.kas.models.Error; @@ -90,6 +91,94 @@ public Error deleteEnterpriseClusterById(Boolean async, String id, Boolean force GenericType localVarReturnType = new GenericType() {}; return apiClient.invokeAPI(localVarPath, "DELETE", localVarQueryParams, localVarPostBody, localVarHeaderParams, localVarCookieParams, localVarFormParams, localVarAccept, localVarContentType, localVarAuthNames, localVarReturnType); } + /** + * + * Returns enterprise data plane cluster by ID + * @param id ID of the enterprise data plane cluster (required) + * @return a {@code EnterpriseCluster} + * @throws ApiException if fails to make API call + */ + public EnterpriseCluster getEnterpriseClusterById(String id) throws ApiException { + Object localVarPostBody = null; + + // verify the required parameter 'id' is set + if (id == null) { + throw new ApiException(400, "Missing the required parameter 'id' when calling getEnterpriseClusterById"); + } + + // create path and map variables + String localVarPath = "/api/kafkas_mgmt/v1/clusters/{id}".replaceAll("\\{format\\}","json") + .replaceAll("\\{" + "id" + "\\}", apiClient.escapeString(id.toString())); + + // query params + List localVarQueryParams = new ArrayList(); + Map localVarHeaderParams = new HashMap(); + Map localVarCookieParams = new HashMap(); + Map localVarFormParams = new HashMap(); + + + + + + final String[] localVarAccepts = { + "application/json" + }; + final String localVarAccept = apiClient.selectHeaderAccept(localVarAccepts); + + final String[] localVarContentTypes = { + + }; + final String localVarContentType = apiClient.selectHeaderContentType(localVarContentTypes); + + String[] localVarAuthNames = new String[] { "Bearer" }; + + GenericType localVarReturnType = new GenericType() {}; + return apiClient.invokeAPI(localVarPath, "GET", localVarQueryParams, localVarPostBody, localVarHeaderParams, localVarCookieParams, localVarFormParams, localVarAccept, localVarContentType, localVarAuthNames, localVarReturnType); + } + /** + * + * Returns enterprise data plane cluster by ID along with its addon parameters + * @param id ID of the enterprise data plane cluster (required) + * @return a {@code EnterpriseClusterWithAddonParameters} + * @throws ApiException if fails to make API call + */ + public EnterpriseClusterWithAddonParameters getEnterpriseClusterWithAddonParameters(String id) throws ApiException { + Object localVarPostBody = null; + + // verify the required parameter 'id' is set + if (id == null) { + throw new ApiException(400, "Missing the required parameter 'id' when calling getEnterpriseClusterWithAddonParameters"); + } + + // create path and map variables + String localVarPath = "/api/kafkas_mgmt/v1/clusters/{id}/addon_parameters".replaceAll("\\{format\\}","json") + .replaceAll("\\{" + "id" + "\\}", apiClient.escapeString(id.toString())); + + // query params + List localVarQueryParams = new ArrayList(); + Map localVarHeaderParams = new HashMap(); + Map localVarCookieParams = new HashMap(); + Map localVarFormParams = new HashMap(); + + + + + + final String[] localVarAccepts = { + "application/json" + }; + final String localVarAccept = apiClient.selectHeaderAccept(localVarAccepts); + + final String[] localVarContentTypes = { + + }; + final String localVarContentType = apiClient.selectHeaderContentType(localVarContentTypes); + + String[] localVarAuthNames = new String[] { "Bearer" }; + + GenericType localVarReturnType = new GenericType() {}; + return apiClient.invokeAPI(localVarPath, "GET", localVarQueryParams, localVarPostBody, localVarHeaderParams, localVarCookieParams, localVarFormParams, localVarAccept, localVarContentType, localVarAuthNames, localVarReturnType); + } /** * * List all Enterprise data plane clusters @@ -131,10 +220,10 @@ public EnterpriseClusterList getEnterpriseOsdClusters() throws ApiException { * * Register enterprise data plane cluster * @param enterpriseOsdClusterPayload Enterprise data plane cluster details (required) - * @return a {@code EnterpriseClusterRegistrationResponse} + * @return a {@code EnterpriseClusterWithAddonParameters} * @throws ApiException if fails to make API call */ - public EnterpriseClusterRegistrationResponse registerEnterpriseOsdCluster(EnterpriseOsdClusterPayload enterpriseOsdClusterPayload) throws ApiException { + public EnterpriseClusterWithAddonParameters registerEnterpriseOsdCluster(EnterpriseOsdClusterPayload enterpriseOsdClusterPayload) throws ApiException { Object localVarPostBody = enterpriseOsdClusterPayload; // verify the required parameter 'enterpriseOsdClusterPayload' is set @@ -167,7 +256,7 @@ public EnterpriseClusterRegistrationResponse registerEnterpriseOsdCluster(Enterp String[] localVarAuthNames = new String[] { "Bearer" }; - GenericType localVarReturnType = new GenericType() {}; + GenericType localVarReturnType = new GenericType() {}; return apiClient.invokeAPI(localVarPath, "POST", localVarQueryParams, localVarPostBody, localVarHeaderParams, localVarCookieParams, localVarFormParams, localVarAccept, localVarContentType, localVarAuthNames, localVarReturnType); } } diff --git a/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponse.java b/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParameters.java similarity index 77% rename from app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponse.java rename to app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParameters.java index 8e48e1e3..65fcbd10 100644 --- a/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponse.java +++ b/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParameters.java @@ -20,7 +20,7 @@ import com.fasterxml.jackson.annotation.JsonCreator; import com.fasterxml.jackson.annotation.JsonTypeName; import com.fasterxml.jackson.annotation.JsonValue; -import com.openshift.cloud.api.kas.models.EnterpriseClusterRegistrationResponseAllOf; +import com.openshift.cloud.api.kas.models.EnterpriseClusterWithAddonParametersAllOf; import com.openshift.cloud.api.kas.models.FleetshardParameter; import com.openshift.cloud.api.kas.models.ObjectReference; import io.swagger.annotations.ApiModel; @@ -31,21 +31,21 @@ import com.fasterxml.jackson.annotation.JsonTypeName; /** - * Enterprise cluster registration endpoint response + * Enterprise cluster with addon parameters */ -@ApiModel(description = "Enterprise cluster registration endpoint response") +@ApiModel(description = "Enterprise cluster with addon parameters") @JsonPropertyOrder({ - EnterpriseClusterRegistrationResponse.JSON_PROPERTY_ID, - EnterpriseClusterRegistrationResponse.JSON_PROPERTY_KIND, - EnterpriseClusterRegistrationResponse.JSON_PROPERTY_HREF, - EnterpriseClusterRegistrationResponse.JSON_PROPERTY_ACCESS_KAFKAS_VIA_PRIVATE_NETWORK, - EnterpriseClusterRegistrationResponse.JSON_PROPERTY_CLUSTER_ID, - EnterpriseClusterRegistrationResponse.JSON_PROPERTY_STATUS, - EnterpriseClusterRegistrationResponse.JSON_PROPERTY_FLEETSHARD_PARAMETERS + EnterpriseClusterWithAddonParameters.JSON_PROPERTY_ID, + EnterpriseClusterWithAddonParameters.JSON_PROPERTY_KIND, + EnterpriseClusterWithAddonParameters.JSON_PROPERTY_HREF, + EnterpriseClusterWithAddonParameters.JSON_PROPERTY_ACCESS_KAFKAS_VIA_PRIVATE_NETWORK, + EnterpriseClusterWithAddonParameters.JSON_PROPERTY_CLUSTER_ID, + EnterpriseClusterWithAddonParameters.JSON_PROPERTY_STATUS, + EnterpriseClusterWithAddonParameters.JSON_PROPERTY_FLEETSHARD_PARAMETERS }) -@JsonTypeName("EnterpriseClusterRegistrationResponse") +@JsonTypeName("EnterpriseClusterWithAddonParameters") @javax.annotation.Generated(value = "org.openapitools.codegen.languages.JavaClientCodegen") -public class EnterpriseClusterRegistrationResponse { +public class EnterpriseClusterWithAddonParameters { public static final String JSON_PROPERTY_ID = "id"; private String id; @@ -67,10 +67,10 @@ public class EnterpriseClusterRegistrationResponse { public static final String JSON_PROPERTY_FLEETSHARD_PARAMETERS = "fleetshard_parameters"; private List fleetshardParameters = null; - public EnterpriseClusterRegistrationResponse() { + public EnterpriseClusterWithAddonParameters() { } - public EnterpriseClusterRegistrationResponse id(String id) { + public EnterpriseClusterWithAddonParameters id(String id) { this.id = id; return this; @@ -97,7 +97,7 @@ public void setId(String id) { } - public EnterpriseClusterRegistrationResponse kind(String kind) { + public EnterpriseClusterWithAddonParameters kind(String kind) { this.kind = kind; return this; @@ -124,7 +124,7 @@ public void setKind(String kind) { } - public EnterpriseClusterRegistrationResponse href(String href) { + public EnterpriseClusterWithAddonParameters href(String href) { this.href = href; return this; @@ -151,7 +151,7 @@ public void setHref(String href) { } - public EnterpriseClusterRegistrationResponse accessKafkasViaPrivateNetwork(Boolean accessKafkasViaPrivateNetwork) { + public EnterpriseClusterWithAddonParameters accessKafkasViaPrivateNetwork(Boolean accessKafkasViaPrivateNetwork) { this.accessKafkasViaPrivateNetwork = accessKafkasViaPrivateNetwork; return this; @@ -178,7 +178,7 @@ public void setAccessKafkasViaPrivateNetwork(Boolean accessKafkasViaPrivateNetwo } - public EnterpriseClusterRegistrationResponse clusterId(String clusterId) { + public EnterpriseClusterWithAddonParameters clusterId(String clusterId) { this.clusterId = clusterId; return this; @@ -205,7 +205,7 @@ public void setClusterId(String clusterId) { } - public EnterpriseClusterRegistrationResponse status(String status) { + public EnterpriseClusterWithAddonParameters status(String status) { this.status = status; return this; @@ -232,13 +232,13 @@ public void setStatus(String status) { } - public EnterpriseClusterRegistrationResponse fleetshardParameters(List fleetshardParameters) { + public EnterpriseClusterWithAddonParameters fleetshardParameters(List fleetshardParameters) { this.fleetshardParameters = fleetshardParameters; return this; } - public EnterpriseClusterRegistrationResponse addFleetshardParametersItem(FleetshardParameter fleetshardParametersItem) { + public EnterpriseClusterWithAddonParameters addFleetshardParametersItem(FleetshardParameter fleetshardParametersItem) { if (this.fleetshardParameters == null) { this.fleetshardParameters = new ArrayList<>(); } @@ -275,14 +275,14 @@ public boolean equals(Object o) { if (o == null || getClass() != o.getClass()) { return false; } - EnterpriseClusterRegistrationResponse enterpriseClusterRegistrationResponse = (EnterpriseClusterRegistrationResponse) o; - return Objects.equals(this.id, enterpriseClusterRegistrationResponse.id) && - Objects.equals(this.kind, enterpriseClusterRegistrationResponse.kind) && - Objects.equals(this.href, enterpriseClusterRegistrationResponse.href) && - Objects.equals(this.accessKafkasViaPrivateNetwork, enterpriseClusterRegistrationResponse.accessKafkasViaPrivateNetwork) && - Objects.equals(this.clusterId, enterpriseClusterRegistrationResponse.clusterId) && - Objects.equals(this.status, enterpriseClusterRegistrationResponse.status) && - Objects.equals(this.fleetshardParameters, enterpriseClusterRegistrationResponse.fleetshardParameters); + EnterpriseClusterWithAddonParameters enterpriseClusterWithAddonParameters = (EnterpriseClusterWithAddonParameters) o; + return Objects.equals(this.id, enterpriseClusterWithAddonParameters.id) && + Objects.equals(this.kind, enterpriseClusterWithAddonParameters.kind) && + Objects.equals(this.href, enterpriseClusterWithAddonParameters.href) && + Objects.equals(this.accessKafkasViaPrivateNetwork, enterpriseClusterWithAddonParameters.accessKafkasViaPrivateNetwork) && + Objects.equals(this.clusterId, enterpriseClusterWithAddonParameters.clusterId) && + Objects.equals(this.status, enterpriseClusterWithAddonParameters.status) && + Objects.equals(this.fleetshardParameters, enterpriseClusterWithAddonParameters.fleetshardParameters); } @Override @@ -293,7 +293,7 @@ public int hashCode() { @Override public String toString() { StringBuilder sb = new StringBuilder(); - sb.append("class EnterpriseClusterRegistrationResponse {\n"); + sb.append("class EnterpriseClusterWithAddonParameters {\n"); sb.append(" id: ").append(toIndentedString(id)).append("\n"); sb.append(" kind: ").append(toIndentedString(kind)).append("\n"); sb.append(" href: ").append(toIndentedString(href)).append("\n"); diff --git a/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponseAllOf.java b/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParametersAllOf.java similarity index 78% rename from app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponseAllOf.java rename to app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParametersAllOf.java index 49522f84..30a3206c 100644 --- a/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterRegistrationResponseAllOf.java +++ b/app-services-sdk-java/packages/kafka-management-sdk/src/main/java/com/openshift/cloud/api/kas/models/EnterpriseClusterWithAddonParametersAllOf.java @@ -29,17 +29,17 @@ import com.fasterxml.jackson.annotation.JsonTypeName; /** - * EnterpriseClusterRegistrationResponseAllOf + * EnterpriseClusterWithAddonParametersAllOf */ @JsonPropertyOrder({ - EnterpriseClusterRegistrationResponseAllOf.JSON_PROPERTY_ACCESS_KAFKAS_VIA_PRIVATE_NETWORK, - EnterpriseClusterRegistrationResponseAllOf.JSON_PROPERTY_CLUSTER_ID, - EnterpriseClusterRegistrationResponseAllOf.JSON_PROPERTY_STATUS, - EnterpriseClusterRegistrationResponseAllOf.JSON_PROPERTY_FLEETSHARD_PARAMETERS + EnterpriseClusterWithAddonParametersAllOf.JSON_PROPERTY_ACCESS_KAFKAS_VIA_PRIVATE_NETWORK, + EnterpriseClusterWithAddonParametersAllOf.JSON_PROPERTY_CLUSTER_ID, + EnterpriseClusterWithAddonParametersAllOf.JSON_PROPERTY_STATUS, + EnterpriseClusterWithAddonParametersAllOf.JSON_PROPERTY_FLEETSHARD_PARAMETERS }) -@JsonTypeName("EnterpriseClusterRegistrationResponse_allOf") +@JsonTypeName("EnterpriseClusterWithAddonParameters_allOf") @javax.annotation.Generated(value = "org.openapitools.codegen.languages.JavaClientCodegen") -public class EnterpriseClusterRegistrationResponseAllOf { +public class EnterpriseClusterWithAddonParametersAllOf { public static final String JSON_PROPERTY_ACCESS_KAFKAS_VIA_PRIVATE_NETWORK = "access_kafkas_via_private_network"; private Boolean accessKafkasViaPrivateNetwork; @@ -52,10 +52,10 @@ public class EnterpriseClusterRegistrationResponseAllOf { public static final String JSON_PROPERTY_FLEETSHARD_PARAMETERS = "fleetshard_parameters"; private List fleetshardParameters = null; - public EnterpriseClusterRegistrationResponseAllOf() { + public EnterpriseClusterWithAddonParametersAllOf() { } - public EnterpriseClusterRegistrationResponseAllOf accessKafkasViaPrivateNetwork(Boolean accessKafkasViaPrivateNetwork) { + public EnterpriseClusterWithAddonParametersAllOf accessKafkasViaPrivateNetwork(Boolean accessKafkasViaPrivateNetwork) { this.accessKafkasViaPrivateNetwork = accessKafkasViaPrivateNetwork; return this; @@ -82,7 +82,7 @@ public void setAccessKafkasViaPrivateNetwork(Boolean accessKafkasViaPrivateNetwo } - public EnterpriseClusterRegistrationResponseAllOf clusterId(String clusterId) { + public EnterpriseClusterWithAddonParametersAllOf clusterId(String clusterId) { this.clusterId = clusterId; return this; @@ -109,7 +109,7 @@ public void setClusterId(String clusterId) { } - public EnterpriseClusterRegistrationResponseAllOf status(String status) { + public EnterpriseClusterWithAddonParametersAllOf status(String status) { this.status = status; return this; @@ -136,13 +136,13 @@ public void setStatus(String status) { } - public EnterpriseClusterRegistrationResponseAllOf fleetshardParameters(List fleetshardParameters) { + public EnterpriseClusterWithAddonParametersAllOf fleetshardParameters(List fleetshardParameters) { this.fleetshardParameters = fleetshardParameters; return this; } - public EnterpriseClusterRegistrationResponseAllOf addFleetshardParametersItem(FleetshardParameter fleetshardParametersItem) { + public EnterpriseClusterWithAddonParametersAllOf addFleetshardParametersItem(FleetshardParameter fleetshardParametersItem) { if (this.fleetshardParameters == null) { this.fleetshardParameters = new ArrayList<>(); } @@ -179,11 +179,11 @@ public boolean equals(Object o) { if (o == null || getClass() != o.getClass()) { return false; } - EnterpriseClusterRegistrationResponseAllOf enterpriseClusterRegistrationResponseAllOf = (EnterpriseClusterRegistrationResponseAllOf) o; - return Objects.equals(this.accessKafkasViaPrivateNetwork, enterpriseClusterRegistrationResponseAllOf.accessKafkasViaPrivateNetwork) && - Objects.equals(this.clusterId, enterpriseClusterRegistrationResponseAllOf.clusterId) && - Objects.equals(this.status, enterpriseClusterRegistrationResponseAllOf.status) && - Objects.equals(this.fleetshardParameters, enterpriseClusterRegistrationResponseAllOf.fleetshardParameters); + EnterpriseClusterWithAddonParametersAllOf enterpriseClusterWithAddonParametersAllOf = (EnterpriseClusterWithAddonParametersAllOf) o; + return Objects.equals(this.accessKafkasViaPrivateNetwork, enterpriseClusterWithAddonParametersAllOf.accessKafkasViaPrivateNetwork) && + Objects.equals(this.clusterId, enterpriseClusterWithAddonParametersAllOf.clusterId) && + Objects.equals(this.status, enterpriseClusterWithAddonParametersAllOf.status) && + Objects.equals(this.fleetshardParameters, enterpriseClusterWithAddonParametersAllOf.fleetshardParameters); } @Override @@ -194,7 +194,7 @@ public int hashCode() { @Override public String toString() { StringBuilder sb = new StringBuilder(); - sb.append("class EnterpriseClusterRegistrationResponseAllOf {\n"); + sb.append("class EnterpriseClusterWithAddonParametersAllOf {\n"); sb.append(" accessKafkasViaPrivateNetwork: ").append(toIndentedString(accessKafkasViaPrivateNetwork)).append("\n"); sb.append(" clusterId: ").append(toIndentedString(clusterId)).append("\n"); sb.append(" status: ").append(toIndentedString(status)).append("\n"); diff --git a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/.openapi-generator/FILES b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/.openapi-generator/FILES index acb69945..96c64372 100644 --- a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/.openapi-generator/FILES +++ b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/.openapi-generator/FILES @@ -16,8 +16,8 @@ model/cloud-region.ts model/enterprise-cluster-all-of.ts model/enterprise-cluster-list-all-of.ts model/enterprise-cluster-list.ts -model/enterprise-cluster-registration-response-all-of.ts -model/enterprise-cluster-registration-response.ts +model/enterprise-cluster-with-addon-parameters-all-of.ts +model/enterprise-cluster-with-addon-parameters.ts model/enterprise-cluster.ts model/enterprise-osd-cluster-payload.ts model/error-list-all-of.ts diff --git a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/api/enterprise-dataplane-clusters-api.ts b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/api/enterprise-dataplane-clusters-api.ts index b2b084a4..5049bd91 100644 --- a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/api/enterprise-dataplane-clusters-api.ts +++ b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/api/enterprise-dataplane-clusters-api.ts @@ -21,9 +21,11 @@ import { DUMMY_BASE_URL, assertParamExists, setApiKeyToObject, setBasicAuthToObj // @ts-ignore import { BASE_PATH, COLLECTION_FORMATS, RequestArgs, BaseAPI, RequiredError } from '../base'; // @ts-ignore +import { EnterpriseCluster } from '../model'; +// @ts-ignore import { EnterpriseClusterList } from '../model'; // @ts-ignore -import { EnterpriseClusterRegistrationResponse } from '../model'; +import { EnterpriseClusterWithAddonParameters } from '../model'; // @ts-ignore import { EnterpriseOsdClusterPayload } from '../model'; /** @@ -72,6 +74,80 @@ export const EnterpriseDataplaneClustersApiAxiosParamCreator = function (configu + setSearchParams(localVarUrlObj, localVarQueryParameter); + let headersFromBaseOptions = baseOptions && baseOptions.headers ? baseOptions.headers : {}; + localVarRequestOptions.headers = {...localVarHeaderParameter, ...headersFromBaseOptions, ...options.headers}; + + return { + url: toPathString(localVarUrlObj), + options: localVarRequestOptions, + }; + }, + /** + * Returns enterprise data plane cluster by ID + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + */ + getEnterpriseClusterById: async (id: string, options: AxiosRequestConfig = {}): Promise => { + // verify required parameter 'id' is not null or undefined + assertParamExists('getEnterpriseClusterById', 'id', id) + const localVarPath = `/api/kafkas_mgmt/v1/clusters/{id}` + .replace(`{${"id"}}`, encodeURIComponent(String(id))); + // use dummy base URL string because the URL constructor only accepts absolute URLs. + const localVarUrlObj = new URL(localVarPath, DUMMY_BASE_URL); + let baseOptions; + if (configuration) { + baseOptions = configuration.baseOptions; + } + + const localVarRequestOptions = { method: 'GET', ...baseOptions, ...options}; + const localVarHeaderParameter = {} as any; + const localVarQueryParameter = {} as any; + + // authentication Bearer required + // http bearer authentication required + await setBearerAuthToObject(localVarHeaderParameter, configuration) + + + + setSearchParams(localVarUrlObj, localVarQueryParameter); + let headersFromBaseOptions = baseOptions && baseOptions.headers ? baseOptions.headers : {}; + localVarRequestOptions.headers = {...localVarHeaderParameter, ...headersFromBaseOptions, ...options.headers}; + + return { + url: toPathString(localVarUrlObj), + options: localVarRequestOptions, + }; + }, + /** + * Returns enterprise data plane cluster by ID along with its addon parameters + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + */ + getEnterpriseClusterWithAddonParameters: async (id: string, options: AxiosRequestConfig = {}): Promise => { + // verify required parameter 'id' is not null or undefined + assertParamExists('getEnterpriseClusterWithAddonParameters', 'id', id) + const localVarPath = `/api/kafkas_mgmt/v1/clusters/{id}/addon_parameters` + .replace(`{${"id"}}`, encodeURIComponent(String(id))); + // use dummy base URL string because the URL constructor only accepts absolute URLs. + const localVarUrlObj = new URL(localVarPath, DUMMY_BASE_URL); + let baseOptions; + if (configuration) { + baseOptions = configuration.baseOptions; + } + + const localVarRequestOptions = { method: 'GET', ...baseOptions, ...options}; + const localVarHeaderParameter = {} as any; + const localVarQueryParameter = {} as any; + + // authentication Bearer required + // http bearer authentication required + await setBearerAuthToObject(localVarHeaderParameter, configuration) + + + setSearchParams(localVarUrlObj, localVarQueryParameter); let headersFromBaseOptions = baseOptions && baseOptions.headers ? baseOptions.headers : {}; localVarRequestOptions.headers = {...localVarHeaderParameter, ...headersFromBaseOptions, ...options.headers}; @@ -175,6 +251,26 @@ export const EnterpriseDataplaneClustersApiFp = function(configuration?: Configu const localVarAxiosArgs = await localVarAxiosParamCreator.deleteEnterpriseClusterById(async, id, force, options); return createRequestFunction(localVarAxiosArgs, globalAxios, BASE_PATH, configuration); }, + /** + * Returns enterprise data plane cluster by ID + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + */ + async getEnterpriseClusterById(id: string, options?: AxiosRequestConfig): Promise<(axios?: AxiosInstance, basePath?: string) => AxiosPromise> { + const localVarAxiosArgs = await localVarAxiosParamCreator.getEnterpriseClusterById(id, options); + return createRequestFunction(localVarAxiosArgs, globalAxios, BASE_PATH, configuration); + }, + /** + * Returns enterprise data plane cluster by ID along with its addon parameters + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + */ + async getEnterpriseClusterWithAddonParameters(id: string, options?: AxiosRequestConfig): Promise<(axios?: AxiosInstance, basePath?: string) => AxiosPromise> { + const localVarAxiosArgs = await localVarAxiosParamCreator.getEnterpriseClusterWithAddonParameters(id, options); + return createRequestFunction(localVarAxiosArgs, globalAxios, BASE_PATH, configuration); + }, /** * List all Enterprise data plane clusters * @param {*} [options] Override http request option. @@ -190,7 +286,7 @@ export const EnterpriseDataplaneClustersApiFp = function(configuration?: Configu * @param {*} [options] Override http request option. * @throws {RequiredError} */ - async registerEnterpriseOsdCluster(enterpriseOsdClusterPayload: EnterpriseOsdClusterPayload, options?: AxiosRequestConfig): Promise<(axios?: AxiosInstance, basePath?: string) => AxiosPromise> { + async registerEnterpriseOsdCluster(enterpriseOsdClusterPayload: EnterpriseOsdClusterPayload, options?: AxiosRequestConfig): Promise<(axios?: AxiosInstance, basePath?: string) => AxiosPromise> { const localVarAxiosArgs = await localVarAxiosParamCreator.registerEnterpriseOsdCluster(enterpriseOsdClusterPayload, options); return createRequestFunction(localVarAxiosArgs, globalAxios, BASE_PATH, configuration); }, @@ -215,6 +311,24 @@ export const EnterpriseDataplaneClustersApiFactory = function (configuration?: C deleteEnterpriseClusterById(async: boolean, id: string, force?: boolean, options?: any): AxiosPromise { return localVarFp.deleteEnterpriseClusterById(async, id, force, options).then((request) => request(axios, basePath)); }, + /** + * Returns enterprise data plane cluster by ID + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + */ + getEnterpriseClusterById(id: string, options?: any): AxiosPromise { + return localVarFp.getEnterpriseClusterById(id, options).then((request) => request(axios, basePath)); + }, + /** + * Returns enterprise data plane cluster by ID along with its addon parameters + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + */ + getEnterpriseClusterWithAddonParameters(id: string, options?: any): AxiosPromise { + return localVarFp.getEnterpriseClusterWithAddonParameters(id, options).then((request) => request(axios, basePath)); + }, /** * List all Enterprise data plane clusters * @param {*} [options] Override http request option. @@ -229,7 +343,7 @@ export const EnterpriseDataplaneClustersApiFactory = function (configuration?: C * @param {*} [options] Override http request option. * @throws {RequiredError} */ - registerEnterpriseOsdCluster(enterpriseOsdClusterPayload: EnterpriseOsdClusterPayload, options?: any): AxiosPromise { + registerEnterpriseOsdCluster(enterpriseOsdClusterPayload: EnterpriseOsdClusterPayload, options?: any): AxiosPromise { return localVarFp.registerEnterpriseOsdCluster(enterpriseOsdClusterPayload, options).then((request) => request(axios, basePath)); }, }; @@ -252,6 +366,24 @@ export interface EnterpriseDataplaneClustersApiInterface { */ deleteEnterpriseClusterById(async: boolean, id: string, force?: boolean, options?: AxiosRequestConfig): AxiosPromise; + /** + * Returns enterprise data plane cluster by ID + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + * @memberof EnterpriseDataplaneClustersApiInterface + */ + getEnterpriseClusterById(id: string, options?: AxiosRequestConfig): AxiosPromise; + + /** + * Returns enterprise data plane cluster by ID along with its addon parameters + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + * @memberof EnterpriseDataplaneClustersApiInterface + */ + getEnterpriseClusterWithAddonParameters(id: string, options?: AxiosRequestConfig): AxiosPromise; + /** * List all Enterprise data plane clusters * @param {*} [options] Override http request option. @@ -267,7 +399,7 @@ export interface EnterpriseDataplaneClustersApiInterface { * @throws {RequiredError} * @memberof EnterpriseDataplaneClustersApiInterface */ - registerEnterpriseOsdCluster(enterpriseOsdClusterPayload: EnterpriseOsdClusterPayload, options?: AxiosRequestConfig): AxiosPromise; + registerEnterpriseOsdCluster(enterpriseOsdClusterPayload: EnterpriseOsdClusterPayload, options?: AxiosRequestConfig): AxiosPromise; } @@ -291,6 +423,28 @@ export class EnterpriseDataplaneClustersApi extends BaseAPI implements Enterpris return EnterpriseDataplaneClustersApiFp(this.configuration).deleteEnterpriseClusterById(async, id, force, options).then((request) => request(this.axios, this.basePath)); } + /** + * Returns enterprise data plane cluster by ID + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + * @memberof EnterpriseDataplaneClustersApi + */ + public getEnterpriseClusterById(id: string, options?: AxiosRequestConfig) { + return EnterpriseDataplaneClustersApiFp(this.configuration).getEnterpriseClusterById(id, options).then((request) => request(this.axios, this.basePath)); + } + + /** + * Returns enterprise data plane cluster by ID along with its addon parameters + * @param {string} id ID of the enterprise data plane cluster + * @param {*} [options] Override http request option. + * @throws {RequiredError} + * @memberof EnterpriseDataplaneClustersApi + */ + public getEnterpriseClusterWithAddonParameters(id: string, options?: AxiosRequestConfig) { + return EnterpriseDataplaneClustersApiFp(this.configuration).getEnterpriseClusterWithAddonParameters(id, options).then((request) => request(this.axios, this.basePath)); + } + /** * List all Enterprise data plane clusters * @param {*} [options] Override http request option. diff --git a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-registration-response-all-of.ts b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-with-addon-parameters-all-of.ts similarity index 74% rename from app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-registration-response-all-of.ts rename to app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-with-addon-parameters-all-of.ts index 21c51a53..7a43fa85 100644 --- a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-registration-response-all-of.ts +++ b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-with-addon-parameters-all-of.ts @@ -18,31 +18,31 @@ import { FleetshardParameter } from './fleetshard-parameter'; /** * * @export - * @interface EnterpriseClusterRegistrationResponseAllOf + * @interface EnterpriseClusterWithAddonParametersAllOf */ -export interface EnterpriseClusterRegistrationResponseAllOf { +export interface EnterpriseClusterWithAddonParametersAllOf { /** * Indicates whether Kafkas created on this data plane cluster have to be accessed via private network * @type {boolean} - * @memberof EnterpriseClusterRegistrationResponseAllOf + * @memberof EnterpriseClusterWithAddonParametersAllOf */ 'access_kafkas_via_private_network': boolean; /** * OCM cluster id of the registered Enterprise cluster * @type {string} - * @memberof EnterpriseClusterRegistrationResponseAllOf + * @memberof EnterpriseClusterWithAddonParametersAllOf */ 'cluster_id'?: string; /** * status of registered Enterprise cluster * @type {string} - * @memberof EnterpriseClusterRegistrationResponseAllOf + * @memberof EnterpriseClusterWithAddonParametersAllOf */ 'status'?: string; /** * * @type {Array} - * @memberof EnterpriseClusterRegistrationResponseAllOf + * @memberof EnterpriseClusterWithAddonParametersAllOf */ 'fleetshard_parameters'?: Array; } diff --git a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-registration-response.ts b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-with-addon-parameters.ts similarity index 61% rename from app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-registration-response.ts rename to app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-with-addon-parameters.ts index c9618cd9..4eac7a77 100644 --- a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-registration-response.ts +++ b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/enterprise-cluster-with-addon-parameters.ts @@ -13,15 +13,15 @@ */ -import { EnterpriseClusterRegistrationResponseAllOf } from './enterprise-cluster-registration-response-all-of'; +import { EnterpriseClusterWithAddonParametersAllOf } from './enterprise-cluster-with-addon-parameters-all-of'; import { FleetshardParameter } from './fleetshard-parameter'; import { ObjectReference } from './object-reference'; /** - * @type EnterpriseClusterRegistrationResponse - * Enterprise cluster registration endpoint response + * @type EnterpriseClusterWithAddonParameters + * Enterprise cluster with addon parameters * @export */ -export type EnterpriseClusterRegistrationResponse = EnterpriseClusterRegistrationResponseAllOf & ObjectReference; +export type EnterpriseClusterWithAddonParameters = EnterpriseClusterWithAddonParametersAllOf & ObjectReference; diff --git a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/index.ts b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/index.ts index a8e848fa..cfb3fd2f 100644 --- a/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/index.ts +++ b/app-services-sdk-js/packages/kafka-management-sdk/src/generated/model/index.ts @@ -8,8 +8,8 @@ export * from './enterprise-cluster'; export * from './enterprise-cluster-all-of'; export * from './enterprise-cluster-list'; export * from './enterprise-cluster-list-all-of'; -export * from './enterprise-cluster-registration-response'; -export * from './enterprise-cluster-registration-response-all-of'; +export * from './enterprise-cluster-with-addon-parameters'; +export * from './enterprise-cluster-with-addon-parameters-all-of'; export * from './enterprise-osd-cluster-payload'; export * from './error-list'; export * from './error-list-all-of'; diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/FILES b/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/FILES index 185e53f3..5bf27d50 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/FILES +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/FILES @@ -10,8 +10,8 @@ docs/EnterpriseCluster.md docs/EnterpriseClusterAllOf.md docs/EnterpriseClusterList.md docs/EnterpriseClusterListAllOf.md -docs/EnterpriseClusterRegistrationResponse.md -docs/EnterpriseClusterRegistrationResponseAllOf.md +docs/EnterpriseClusterWithAddonParameters.md +docs/EnterpriseClusterWithAddonParametersAllOf.md docs/EnterpriseDataplaneClustersApi.md docs/EnterpriseOsdClusterPayload.md docs/Error.md @@ -46,7 +46,6 @@ docs/SsoProvider.md docs/SsoProviderAllOf.md docs/SupportedKafkaBillingModel.md docs/SupportedKafkaInstanceType.md -docs/SupportedKafkaInstanceTypeSizesInner.md docs/SupportedKafkaInstanceTypesList.md docs/SupportedKafkaInstanceTypesListAllOf.md docs/SupportedKafkaSize.md @@ -76,8 +75,8 @@ rhoas_kafka_mgmt_sdk/model/enterprise_cluster.py rhoas_kafka_mgmt_sdk/model/enterprise_cluster_all_of.py rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list.py rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list_all_of.py -rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response.py -rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response_all_of.py +rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters.py +rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters_all_of.py rhoas_kafka_mgmt_sdk/model/enterprise_osd_cluster_payload.py rhoas_kafka_mgmt_sdk/model/error.py rhoas_kafka_mgmt_sdk/model/error_list.py @@ -109,7 +108,6 @@ rhoas_kafka_mgmt_sdk/model/sso_provider.py rhoas_kafka_mgmt_sdk/model/sso_provider_all_of.py rhoas_kafka_mgmt_sdk/model/supported_kafka_billing_model.py rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type.py -rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type_sizes_inner.py rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list.py rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list_all_of.py rhoas_kafka_mgmt_sdk/model/supported_kafka_size.py @@ -135,8 +133,8 @@ test/test_enterprise_cluster.py test/test_enterprise_cluster_all_of.py test/test_enterprise_cluster_list.py test/test_enterprise_cluster_list_all_of.py -test/test_enterprise_cluster_registration_response.py -test/test_enterprise_cluster_registration_response_all_of.py +test/test_enterprise_cluster_with_addon_parameters.py +test/test_enterprise_cluster_with_addon_parameters_all_of.py test/test_enterprise_dataplane_clusters_api.py test/test_enterprise_osd_cluster_payload.py test/test_error.py @@ -171,7 +169,6 @@ test/test_sso_provider.py test/test_sso_provider_all_of.py test/test_supported_kafka_billing_model.py test/test_supported_kafka_instance_type.py -test/test_supported_kafka_instance_type_sizes_inner.py test/test_supported_kafka_instance_types_list.py test/test_supported_kafka_instance_types_list_all_of.py test/test_supported_kafka_size.py diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/VERSION b/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/VERSION index 358e78e6..1e20ec35 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/VERSION +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/.openapi-generator/VERSION @@ -1 +1 @@ -6.1.0 \ No newline at end of file +5.4.0 \ No newline at end of file diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterRegistrationResponse.md b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterWithAddonParameters.md similarity index 91% rename from app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterRegistrationResponse.md rename to app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterWithAddonParameters.md index 7c5209a2..6354099d 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterRegistrationResponse.md +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterWithAddonParameters.md @@ -1,6 +1,6 @@ -# EnterpriseClusterRegistrationResponse +# EnterpriseClusterWithAddonParameters -Enterprise cluster registration endpoint response +Enterprise cluster with addon parameters ## Properties Name | Type | Description | Notes diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterRegistrationResponseAllOf.md b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterWithAddonParametersAllOf.md similarity index 95% rename from app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterRegistrationResponseAllOf.md rename to app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterWithAddonParametersAllOf.md index 07ac647b..05e11579 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterRegistrationResponseAllOf.md +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseClusterWithAddonParametersAllOf.md @@ -1,4 +1,4 @@ -# EnterpriseClusterRegistrationResponseAllOf +# EnterpriseClusterWithAddonParametersAllOf ## Properties diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseDataplaneClustersApi.md b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseDataplaneClustersApi.md index b80e8a53..e859d69b 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseDataplaneClustersApi.md +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/EnterpriseDataplaneClustersApi.md @@ -5,6 +5,8 @@ All URIs are relative to *https://api.openshift.com* Method | HTTP request | Description ------------- | ------------- | ------------- [**delete_enterprise_cluster_by_id**](EnterpriseDataplaneClustersApi.md#delete_enterprise_cluster_by_id) | **DELETE** /api/kafkas_mgmt/v1/clusters/{id} | +[**get_enterprise_cluster_by_id**](EnterpriseDataplaneClustersApi.md#get_enterprise_cluster_by_id) | **GET** /api/kafkas_mgmt/v1/clusters/{id} | +[**get_enterprise_cluster_with_addon_parameters**](EnterpriseDataplaneClustersApi.md#get_enterprise_cluster_with_addon_parameters) | **GET** /api/kafkas_mgmt/v1/clusters/{id}/addon_parameters | [**get_enterprise_osd_clusters**](EnterpriseDataplaneClustersApi.md#get_enterprise_osd_clusters) | **GET** /api/kafkas_mgmt/v1/clusters | [**register_enterprise_osd_cluster**](EnterpriseDataplaneClustersApi.md#register_enterprise_osd_cluster) | **POST** /api/kafkas_mgmt/v1/clusters | @@ -100,6 +102,168 @@ Name | Type | Description | Notes [[Back to top]](#) [[Back to API list]](../README.md#documentation-for-api-endpoints) [[Back to Model list]](../README.md#documentation-for-models) [[Back to README]](../README.md) +# **get_enterprise_cluster_by_id** +> EnterpriseCluster get_enterprise_cluster_by_id(id) + + + +Returns enterprise data plane cluster by ID + +### Example + +* Bearer (JWT) Authentication (Bearer): + +```python +import time +import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.api import enterprise_dataplane_clusters_api +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster import EnterpriseCluster +from rhoas_kafka_mgmt_sdk.model.error import Error +from pprint import pprint +# Defining the host is optional and defaults to https://api.openshift.com +# See configuration.py for a list of all supported configuration parameters. +configuration = rhoas_kafka_mgmt_sdk.Configuration( + host = "https://api.openshift.com" +) + +# The client must configure the authentication and authorization parameters +# in accordance with the API server security policy. +# Examples for each auth method are provided below, use the example that +# satisfies your auth use case. + +# Configure Bearer authorization (JWT): Bearer +configuration = rhoas_kafka_mgmt_sdk.Configuration( + access_token = 'YOUR_BEARER_TOKEN' +) + +# Enter a context with an instance of the API client +with rhoas_kafka_mgmt_sdk.ApiClient(configuration) as api_client: + # Create an instance of the API class + api_instance = enterprise_dataplane_clusters_api.EnterpriseDataplaneClustersApi(api_client) + id = "id_example" # str | ID of the enterprise data plane cluster + + # example passing only required values which don't have defaults set + try: + api_response = api_instance.get_enterprise_cluster_by_id(id) + pprint(api_response) + except rhoas_kafka_mgmt_sdk.ApiException as e: + print("Exception when calling EnterpriseDataplaneClustersApi->get_enterprise_cluster_by_id: %s\n" % e) +``` + + +### Parameters + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- + **id** | **str**| ID of the enterprise data plane cluster | + +### Return type + +[**EnterpriseCluster**](EnterpriseCluster.md) + +### Authorization + +[Bearer](../README.md#Bearer) + +### HTTP request headers + + - **Content-Type**: Not defined + - **Accept**: application/json + + +### HTTP response details + +| Status code | Description | Response headers | +|-------------|-------------|------------------| +**200** | Found enterprise data plane cluster with provided ID | - | +**401** | Auth token is invalid | - | +**403** | User not authorized to access the service | - | +**404** | No Enterprise data plane cluster with specified ID exists | - | +**500** | Unexpected error occurred | - | + +[[Back to top]](#) [[Back to API list]](../README.md#documentation-for-api-endpoints) [[Back to Model list]](../README.md#documentation-for-models) [[Back to README]](../README.md) + +# **get_enterprise_cluster_with_addon_parameters** +> EnterpriseClusterWithAddonParameters get_enterprise_cluster_with_addon_parameters(id) + + + +Returns enterprise data plane cluster by ID along with its addon parameters + +### Example + +* Bearer (JWT) Authentication (Bearer): + +```python +import time +import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.api import enterprise_dataplane_clusters_api +from rhoas_kafka_mgmt_sdk.model.error import Error +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters import EnterpriseClusterWithAddonParameters +from pprint import pprint +# Defining the host is optional and defaults to https://api.openshift.com +# See configuration.py for a list of all supported configuration parameters. +configuration = rhoas_kafka_mgmt_sdk.Configuration( + host = "https://api.openshift.com" +) + +# The client must configure the authentication and authorization parameters +# in accordance with the API server security policy. +# Examples for each auth method are provided below, use the example that +# satisfies your auth use case. + +# Configure Bearer authorization (JWT): Bearer +configuration = rhoas_kafka_mgmt_sdk.Configuration( + access_token = 'YOUR_BEARER_TOKEN' +) + +# Enter a context with an instance of the API client +with rhoas_kafka_mgmt_sdk.ApiClient(configuration) as api_client: + # Create an instance of the API class + api_instance = enterprise_dataplane_clusters_api.EnterpriseDataplaneClustersApi(api_client) + id = "id_example" # str | ID of the enterprise data plane cluster + + # example passing only required values which don't have defaults set + try: + api_response = api_instance.get_enterprise_cluster_with_addon_parameters(id) + pprint(api_response) + except rhoas_kafka_mgmt_sdk.ApiException as e: + print("Exception when calling EnterpriseDataplaneClustersApi->get_enterprise_cluster_with_addon_parameters: %s\n" % e) +``` + + +### Parameters + +Name | Type | Description | Notes +------------- | ------------- | ------------- | ------------- + **id** | **str**| ID of the enterprise data plane cluster | + +### Return type + +[**EnterpriseClusterWithAddonParameters**](EnterpriseClusterWithAddonParameters.md) + +### Authorization + +[Bearer](../README.md#Bearer) + +### HTTP request headers + + - **Content-Type**: Not defined + - **Accept**: application/json + + +### HTTP response details + +| Status code | Description | Response headers | +|-------------|-------------|------------------| +**200** | Returns the enterprise data plane cluster with addon parameters for an existing enterprise data plane cluster with provided ID | - | +**401** | Auth token is invalid | - | +**403** | User not authorized to access the service | - | +**404** | No Enterprise data plane cluster with specified ID exists | - | +**500** | Unexpected error occurred | - | + +[[Back to top]](#) [[Back to API list]](../README.md#documentation-for-api-endpoints) [[Back to Model list]](../README.md#documentation-for-models) [[Back to README]](../README.md) + # **get_enterprise_osd_clusters** > EnterpriseClusterList get_enterprise_osd_clusters() @@ -178,7 +342,7 @@ This endpoint does not need any parameter. [[Back to top]](#) [[Back to API list]](../README.md#documentation-for-api-endpoints) [[Back to Model list]](../README.md#documentation-for-models) [[Back to README]](../README.md) # **register_enterprise_osd_cluster** -> EnterpriseClusterRegistrationResponse register_enterprise_osd_cluster(enterprise_osd_cluster_payload) +> EnterpriseClusterWithAddonParameters register_enterprise_osd_cluster(enterprise_osd_cluster_payload) @@ -192,8 +356,8 @@ Register enterprise data plane cluster import time import rhoas_kafka_mgmt_sdk from rhoas_kafka_mgmt_sdk.api import enterprise_dataplane_clusters_api -from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response import EnterpriseClusterRegistrationResponse from rhoas_kafka_mgmt_sdk.model.error import Error +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters import EnterpriseClusterWithAddonParameters from rhoas_kafka_mgmt_sdk.model.enterprise_osd_cluster_payload import EnterpriseOsdClusterPayload from pprint import pprint # Defining the host is optional and defaults to https://api.openshift.com @@ -241,7 +405,7 @@ Name | Type | Description | Notes ### Return type -[**EnterpriseClusterRegistrationResponse**](EnterpriseClusterRegistrationResponse.md) +[**EnterpriseClusterWithAddonParameters**](EnterpriseClusterWithAddonParameters.md) ### Authorization diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/ServiceAccountListItem.md b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/ServiceAccountListItem.md index c4c2f357..b08552eb 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/ServiceAccountListItem.md +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/ServiceAccountListItem.md @@ -1,7 +1,6 @@ # ServiceAccountListItem - ## Properties Name | Type | Description | Notes ------------ | ------------- | ------------- | ------------- diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/SupportedKafkaInstanceType.md b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/SupportedKafkaInstanceType.md index 1291a115..4457141a 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/SupportedKafkaInstanceType.md +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/SupportedKafkaInstanceType.md @@ -8,7 +8,7 @@ Name | Type | Description | Notes **supported_billing_models** | [**[SupportedKafkaBillingModel]**](SupportedKafkaBillingModel.md) | A list of available kafka billing models for the instance type. Each kafka billing model item has a unique 'id' | **id** | **str** | Unique identifier of the Kafka instance type. | [optional] **display_name** | **str** | Human readable name of the supported Kafka instance type | [optional] -**sizes** | [**[SupportedKafkaInstanceTypeSizesInner]**](SupportedKafkaInstanceTypeSizesInner.md) | A list of Kafka instance sizes available for this instance type | [optional] +**sizes** | **[bool, date, datetime, dict, float, int, list, str, none_type]** | A list of Kafka instance sizes available for this instance type | [optional] **any string name** | **bool, date, datetime, dict, float, int, list, str, none_type** | any string name can be used but the value must be the correct type | [optional] [[Back to Model list]](../README.md#documentation-for-models) [[Back to API list]](../README.md#documentation-for-api-endpoints) [[Back to README]](../README.md) diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/SupportedKafkaInstanceTypeSizesInner.md b/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/SupportedKafkaInstanceTypeSizesInner.md deleted file mode 100644 index 3022f63d..00000000 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/docs/SupportedKafkaInstanceTypeSizesInner.md +++ /dev/null @@ -1,29 +0,0 @@ -# SupportedKafkaInstanceTypeSizesInner - - -## Properties -Name | Type | Description | Notes ------------- | ------------- | ------------- | ------------- -**id** | **str** | Unique identifier of this Kafka instance size. | [optional] -**display_name** | **str** | Display name of this Kafka instance size. | [optional] -**ingress_throughput_per_sec** | **SupportedKafkaSizeBytesValueItem** | | [optional] -**egress_throughput_per_sec** | **SupportedKafkaSizeBytesValueItem** | | [optional] -**total_max_connections** | **int** | Maximum amount of total connections available to this Kafka instance size. | [optional] -**max_data_retention_size** | **SupportedKafkaSizeBytesValueItem** | | [optional] -**max_partitions** | **int** | Maximum amount of total partitions available to this Kafka instance size. | [optional] -**max_data_retention_period** | **str** | Maximum data retention period available to this Kafka instance size. | [optional] -**max_connection_attempts_per_sec** | **int** | Maximium connection attempts per second available to this Kafka instance size. | [optional] -**max_message_size** | **SupportedKafkaSizeBytesValueItem** | | [optional] -**min_in_sync_replicas** | **int** | Minimum number of in-sync replicas. | [optional] -**replication_factor** | **int** | Replication factor available to this Kafka instance size. | [optional] -**supported_az_modes** | **[str]** | List of Availability Zone modes that this Kafka instance size supports. The possible values are \"single\", \"multi\". | [optional] -**lifespan_seconds** | **int, none_type** | The limit lifespan of the kafka instance in seconds. If not specified then the instance never expires. | [optional] -**quota_consumed** | **int** | Quota consumed by this Kafka instance size. | [optional] -**quota_type** | **str** | Quota type used by this Kafka instance size. This is now deprecated, please refer to supported_billing_models at instance-type level instead. | [optional] -**capacity_consumed** | **int** | Data plane cluster capacity consumed by this Kafka instance size. | [optional] -**maturity_status** | **str** | Maturity level of the size. Can be \"stable\" or \"preview\". | [optional] -**any string name** | **bool, date, datetime, dict, float, int, list, str, none_type** | any string name can be used but the value must be the correct type | [optional] - -[[Back to Model list]](../README.md#documentation-for-models) [[Back to API list]](../README.md#documentation-for-api-endpoints) [[Back to README]](../README.md) - - diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/default_api.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/default_api.py index ca389929..3dc3d8ec 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/default_api.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/default_api.py @@ -784,10 +784,6 @@ def create_kafka( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -819,7 +815,6 @@ def create_kafka( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['_async'] = \ _async kwargs['kafka_request_payload'] = \ @@ -871,10 +866,6 @@ def delete_kafka_by_id( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -906,7 +897,6 @@ def delete_kafka_by_id( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id kwargs['_async'] = \ @@ -956,10 +946,6 @@ def federate_metrics( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -991,7 +977,6 @@ def federate_metrics( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.federate_metrics_endpoint.call_with_http_info(**kwargs) @@ -1041,10 +1026,6 @@ def get_cloud_provider_regions( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1076,7 +1057,6 @@ def get_cloud_provider_regions( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.get_cloud_provider_regions_endpoint.call_with_http_info(**kwargs) @@ -1123,10 +1103,6 @@ def get_cloud_providers( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1158,7 +1134,6 @@ def get_cloud_providers( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) return self.get_cloud_providers_endpoint.call_with_http_info(**kwargs) def get_instance_types_by_cloud_provider_and_region( @@ -1206,10 +1181,6 @@ def get_instance_types_by_cloud_provider_and_region( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1241,7 +1212,6 @@ def get_instance_types_by_cloud_provider_and_region( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['cloud_provider'] = \ cloud_provider kwargs['cloud_region'] = \ @@ -1291,10 +1261,6 @@ def get_kafka_by_id( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1326,7 +1292,6 @@ def get_kafka_by_id( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.get_kafka_by_id_endpoint.call_with_http_info(**kwargs) @@ -1375,10 +1340,6 @@ def get_kafkas( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1410,7 +1371,6 @@ def get_kafkas( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) return self.get_kafkas_endpoint.call_with_http_info(**kwargs) def get_metrics_by_instant_query( @@ -1457,10 +1417,6 @@ def get_metrics_by_instant_query( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1492,7 +1448,6 @@ def get_metrics_by_instant_query( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.get_metrics_by_instant_query_endpoint.call_with_http_info(**kwargs) @@ -1545,10 +1500,6 @@ def get_metrics_by_range_query( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1580,7 +1531,6 @@ def get_metrics_by_range_query( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id kwargs['duration'] = \ @@ -1629,10 +1579,6 @@ def get_version_metadata( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1664,7 +1610,6 @@ def get_version_metadata( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) return self.get_version_metadata_endpoint.call_with_http_info(**kwargs) def update_kafka_by_id( @@ -1712,10 +1657,6 @@ def update_kafka_by_id( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -1747,7 +1688,6 @@ def update_kafka_by_id( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id kwargs['kafka_update_request'] = \ diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/enterprise_dataplane_clusters_api.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/enterprise_dataplane_clusters_api.py index 63c1b2d5..3cb51c0f 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/enterprise_dataplane_clusters_api.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/enterprise_dataplane_clusters_api.py @@ -22,8 +22,9 @@ none_type, validate_and_convert_types ) +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster import EnterpriseCluster from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_list import EnterpriseClusterList -from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response import EnterpriseClusterRegistrationResponse +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters import EnterpriseClusterWithAddonParameters from rhoas_kafka_mgmt_sdk.model.enterprise_osd_cluster_payload import EnterpriseOsdClusterPayload from rhoas_kafka_mgmt_sdk.model.error import Error @@ -101,6 +102,108 @@ def __init__(self, api_client=None): }, api_client=api_client ) + self.get_enterprise_cluster_by_id_endpoint = _Endpoint( + settings={ + 'response_type': (EnterpriseCluster,), + 'auth': [ + 'Bearer' + ], + 'endpoint_path': '/api/kafkas_mgmt/v1/clusters/{id}', + 'operation_id': 'get_enterprise_cluster_by_id', + 'http_method': 'GET', + 'servers': None, + }, + params_map={ + 'all': [ + 'id', + ], + 'required': [ + 'id', + ], + 'nullable': [ + ], + 'enum': [ + ], + 'validation': [ + ] + }, + root_map={ + 'validations': { + }, + 'allowed_values': { + }, + 'openapi_types': { + 'id': + (str,), + }, + 'attribute_map': { + 'id': 'id', + }, + 'location_map': { + 'id': 'path', + }, + 'collection_format_map': { + } + }, + headers_map={ + 'accept': [ + 'application/json' + ], + 'content_type': [], + }, + api_client=api_client + ) + self.get_enterprise_cluster_with_addon_parameters_endpoint = _Endpoint( + settings={ + 'response_type': (EnterpriseClusterWithAddonParameters,), + 'auth': [ + 'Bearer' + ], + 'endpoint_path': '/api/kafkas_mgmt/v1/clusters/{id}/addon_parameters', + 'operation_id': 'get_enterprise_cluster_with_addon_parameters', + 'http_method': 'GET', + 'servers': None, + }, + params_map={ + 'all': [ + 'id', + ], + 'required': [ + 'id', + ], + 'nullable': [ + ], + 'enum': [ + ], + 'validation': [ + ] + }, + root_map={ + 'validations': { + }, + 'allowed_values': { + }, + 'openapi_types': { + 'id': + (str,), + }, + 'attribute_map': { + 'id': 'id', + }, + 'location_map': { + 'id': 'path', + }, + 'collection_format_map': { + } + }, + headers_map={ + 'accept': [ + 'application/json' + ], + 'content_type': [], + }, + api_client=api_client + ) self.get_enterprise_osd_clusters_endpoint = _Endpoint( settings={ 'response_type': (EnterpriseClusterList,), @@ -147,7 +250,7 @@ def __init__(self, api_client=None): ) self.register_enterprise_osd_cluster_endpoint = _Endpoint( settings={ - 'response_type': (EnterpriseClusterRegistrationResponse,), + 'response_type': (EnterpriseClusterWithAddonParameters,), 'auth': [ 'Bearer' ], @@ -243,10 +346,6 @@ def delete_enterprise_cluster_by_id( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -278,13 +377,168 @@ def delete_enterprise_cluster_by_id( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['_async'] = \ _async kwargs['id'] = \ id return self.delete_enterprise_cluster_by_id_endpoint.call_with_http_info(**kwargs) + def get_enterprise_cluster_by_id( + self, + id, + **kwargs + ): + """get_enterprise_cluster_by_id # noqa: E501 + + Returns enterprise data plane cluster by ID # noqa: E501 + This method makes a synchronous HTTP request by default. To make an + asynchronous HTTP request, please pass async_req=True + + >>> thread = api.get_enterprise_cluster_by_id(id, async_req=True) + >>> result = thread.get() + + Args: + id (str): ID of the enterprise data plane cluster + + Keyword Args: + _return_http_data_only (bool): response data without head status + code and headers. Default is True. + _preload_content (bool): if False, the urllib3.HTTPResponse object + will be returned without reading/decoding response data. + Default is True. + _request_timeout (int/float/tuple): timeout setting for this request. If + one number provided, it will be total request timeout. It can also + be a pair (tuple) of (connection, read) timeouts. + Default is None. + _check_input_type (bool): specifies if type checking + should be done one the data sent to the server. + Default is True. + _check_return_type (bool): specifies if type checking + should be done one the data received from the server. + Default is True. + _spec_property_naming (bool): True if the variable names in the input data + are serialized names, as specified in the OpenAPI document. + False if the variable names in the input data + are pythonic names, e.g. snake case (default) + _content_type (str/None): force body content-type. + Default is None and content-type will be predicted by allowed + content-types and body. + _host_index (int/None): specifies the index of the server + that we want to use. + Default is read from the configuration. + async_req (bool): execute request asynchronously + + Returns: + EnterpriseCluster + If the method is called asynchronously, returns the request + thread. + """ + kwargs['async_req'] = kwargs.get( + 'async_req', False + ) + kwargs['_return_http_data_only'] = kwargs.get( + '_return_http_data_only', True + ) + kwargs['_preload_content'] = kwargs.get( + '_preload_content', True + ) + kwargs['_request_timeout'] = kwargs.get( + '_request_timeout', None + ) + kwargs['_check_input_type'] = kwargs.get( + '_check_input_type', True + ) + kwargs['_check_return_type'] = kwargs.get( + '_check_return_type', True + ) + kwargs['_spec_property_naming'] = kwargs.get( + '_spec_property_naming', False + ) + kwargs['_content_type'] = kwargs.get( + '_content_type') + kwargs['_host_index'] = kwargs.get('_host_index') + kwargs['id'] = \ + id + return self.get_enterprise_cluster_by_id_endpoint.call_with_http_info(**kwargs) + + def get_enterprise_cluster_with_addon_parameters( + self, + id, + **kwargs + ): + """get_enterprise_cluster_with_addon_parameters # noqa: E501 + + Returns enterprise data plane cluster by ID along with its addon parameters # noqa: E501 + This method makes a synchronous HTTP request by default. To make an + asynchronous HTTP request, please pass async_req=True + + >>> thread = api.get_enterprise_cluster_with_addon_parameters(id, async_req=True) + >>> result = thread.get() + + Args: + id (str): ID of the enterprise data plane cluster + + Keyword Args: + _return_http_data_only (bool): response data without head status + code and headers. Default is True. + _preload_content (bool): if False, the urllib3.HTTPResponse object + will be returned without reading/decoding response data. + Default is True. + _request_timeout (int/float/tuple): timeout setting for this request. If + one number provided, it will be total request timeout. It can also + be a pair (tuple) of (connection, read) timeouts. + Default is None. + _check_input_type (bool): specifies if type checking + should be done one the data sent to the server. + Default is True. + _check_return_type (bool): specifies if type checking + should be done one the data received from the server. + Default is True. + _spec_property_naming (bool): True if the variable names in the input data + are serialized names, as specified in the OpenAPI document. + False if the variable names in the input data + are pythonic names, e.g. snake case (default) + _content_type (str/None): force body content-type. + Default is None and content-type will be predicted by allowed + content-types and body. + _host_index (int/None): specifies the index of the server + that we want to use. + Default is read from the configuration. + async_req (bool): execute request asynchronously + + Returns: + EnterpriseClusterWithAddonParameters + If the method is called asynchronously, returns the request + thread. + """ + kwargs['async_req'] = kwargs.get( + 'async_req', False + ) + kwargs['_return_http_data_only'] = kwargs.get( + '_return_http_data_only', True + ) + kwargs['_preload_content'] = kwargs.get( + '_preload_content', True + ) + kwargs['_request_timeout'] = kwargs.get( + '_request_timeout', None + ) + kwargs['_check_input_type'] = kwargs.get( + '_check_input_type', True + ) + kwargs['_check_return_type'] = kwargs.get( + '_check_return_type', True + ) + kwargs['_spec_property_naming'] = kwargs.get( + '_spec_property_naming', False + ) + kwargs['_content_type'] = kwargs.get( + '_content_type') + kwargs['_host_index'] = kwargs.get('_host_index') + kwargs['id'] = \ + id + return self.get_enterprise_cluster_with_addon_parameters_endpoint.call_with_http_info(**kwargs) + def get_enterprise_osd_clusters( self, **kwargs @@ -325,10 +579,6 @@ def get_enterprise_osd_clusters( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -360,7 +610,6 @@ def get_enterprise_osd_clusters( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) return self.get_enterprise_osd_clusters_endpoint.call_with_http_info(**kwargs) def register_enterprise_osd_cluster( @@ -406,14 +655,10 @@ def register_enterprise_osd_cluster( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: - EnterpriseClusterRegistrationResponse + EnterpriseClusterWithAddonParameters If the method is called asynchronously, returns the request thread. """ @@ -441,7 +686,6 @@ def register_enterprise_osd_cluster( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['enterprise_osd_cluster_payload'] = \ enterprise_osd_cluster_payload return self.register_enterprise_osd_cluster_endpoint.call_with_http_info(**kwargs) diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/errors_api.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/errors_api.py index 4f529ee3..dc4c4fc7 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/errors_api.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/errors_api.py @@ -172,10 +172,6 @@ def get_error_by_id( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -207,7 +203,6 @@ def get_error_by_id( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.get_error_by_id_endpoint.call_with_http_info(**kwargs) @@ -252,10 +247,6 @@ def get_errors( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -287,6 +278,5 @@ def get_errors( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) return self.get_errors_endpoint.call_with_http_info(**kwargs) diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/security_api.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/security_api.py index c19d85ce..6c25e3ff 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/security_api.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api/security_api.py @@ -380,10 +380,6 @@ def create_service_account( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -415,7 +411,6 @@ def create_service_account( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['service_account_request'] = \ service_account_request return self.create_service_account_endpoint.call_with_http_info(**kwargs) @@ -463,10 +458,6 @@ def delete_service_account_by_id( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -498,7 +489,6 @@ def delete_service_account_by_id( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.delete_service_account_by_id_endpoint.call_with_http_info(**kwargs) @@ -546,10 +536,6 @@ def get_service_account_by_id( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -581,7 +567,6 @@ def get_service_account_by_id( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.get_service_account_by_id_endpoint.call_with_http_info(**kwargs) @@ -627,10 +612,6 @@ def get_service_accounts( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -662,7 +643,6 @@ def get_service_accounts( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) return self.get_service_accounts_endpoint.call_with_http_info(**kwargs) def get_sso_providers( @@ -705,10 +685,6 @@ def get_sso_providers( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -740,7 +716,6 @@ def get_sso_providers( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) return self.get_sso_providers_endpoint.call_with_http_info(**kwargs) def reset_service_account_creds( @@ -786,10 +761,6 @@ def reset_service_account_creds( _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. - _request_auths (list): set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - Default is None async_req (bool): execute request asynchronously Returns: @@ -821,7 +792,6 @@ def reset_service_account_creds( kwargs['_content_type'] = kwargs.get( '_content_type') kwargs['_host_index'] = kwargs.get('_host_index') - kwargs['_request_auths'] = kwargs.get('_request_auths', None) kwargs['id'] = \ id return self.reset_service_account_creds_endpoint.call_with_http_info(**kwargs) diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api_client.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api_client.py index 1851b518..d10e0ebe 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api_client.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/api_client.py @@ -133,8 +133,7 @@ def __call_api( _request_timeout: typing.Optional[typing.Union[int, float, typing.Tuple]] = None, _host: typing.Optional[str] = None, _check_type: typing.Optional[bool] = None, - _content_type: typing.Optional[str] = None, - _request_auths: typing.Optional[typing.List[typing.Dict[str, typing.Any]]] = None + _content_type: typing.Optional[str] = None ): config = self.configuration @@ -176,7 +175,7 @@ def __call_api( post_params.extend(self.files_parameters(files)) if header_params['Content-Type'].startswith("multipart"): post_params = self.parameters_to_multipart(post_params, - (dict)) + (dict) ) # body if body: @@ -184,8 +183,7 @@ def __call_api( # auth setting self.update_params_for_auth(header_params, query_params, - auth_settings, resource_path, method, body, - request_auths=_request_auths) + auth_settings, resource_path, method, body) # request url if _host is None: @@ -249,14 +247,13 @@ def parameters_to_multipart(self, params, collection_types): if collection_types is None: collection_types = (dict) for k, v in params.items() if isinstance(params, dict) else params: # noqa: E501 - if isinstance( - v, collection_types): # v is instance of collection_type, formatting as application/json - v = json.dumps(v, ensure_ascii=False).encode("utf-8") - field = RequestField(k, v) - field.make_multipart(content_type="application/json; charset=utf-8") - new_params.append(field) + if isinstance(v, collection_types): # v is instance of collection_type, formatting as application/json + v = json.dumps(v, ensure_ascii=False).encode("utf-8") + field = RequestField(k, v) + field.make_multipart(content_type="application/json; charset=utf-8") + new_params.append(field) else: - new_params.append((k, v)) + new_params.append((k, v)) return new_params @classmethod @@ -275,10 +272,8 @@ def sanitize_for_serialization(cls, obj): """ if isinstance(obj, (ModelNormal, ModelComposed)): return { - key: cls.sanitize_for_serialization(val) for key, - val in model_to_dict( - obj, - serialize=True).items()} + key: cls.sanitize_for_serialization(val) for key, val in model_to_dict(obj, serialize=True).items() + } elif isinstance(obj, io.IOBase): return cls.get_file_data_and_close_file(obj) elif isinstance(obj, (str, int, float, none_type, bool)): @@ -291,9 +286,7 @@ def sanitize_for_serialization(cls, obj): return [cls.sanitize_for_serialization(item) for item in obj] if isinstance(obj, dict): return {key: cls.sanitize_for_serialization(val) for key, val in obj.items()} - raise ApiValueError( - 'Unable to prepare type {} for serialization'.format( - obj.__class__.__name__)) + raise ApiValueError('Unable to prepare type {} for serialization'.format(obj.__class__.__name__)) def deserialize(self, response, response_type, _check_type): """Deserializes response into an object. @@ -358,8 +351,7 @@ def call_api( _preload_content: bool = True, _request_timeout: typing.Optional[typing.Union[int, float, typing.Tuple]] = None, _host: typing.Optional[str] = None, - _check_type: typing.Optional[bool] = None, - _request_auths: typing.Optional[typing.List[typing.Dict[str, typing.Any]]] = None + _check_type: typing.Optional[bool] = None ): """Makes the HTTP request (synchronous) and returns deserialized data. @@ -407,10 +399,6 @@ def call_api( :param _check_type: boolean describing if the data back from the server should have its type checked. :type _check_type: bool, optional - :param _request_auths: set to override the auth_settings for an a single - request; this effectively ignores the authentication - in the spec for a single request. - :type _request_auths: list, optional :return: If async_req parameter is True, the request will be called asynchronously. @@ -425,7 +413,7 @@ def call_api( response_type, auth_settings, _return_http_data_only, collection_formats, _preload_content, _request_timeout, _host, - _check_type, _request_auths=_request_auths) + _check_type) return self.pool.apply_async(self.__call_api, (resource_path, method, path_params, @@ -438,7 +426,7 @@ def call_api( collection_formats, _preload_content, _request_timeout, - _host, _check_type, None, _request_auths)) + _host, _check_type)) def request(self, method, url, query_params=None, headers=None, post_params=None, body=None, _preload_content=True, @@ -537,9 +525,7 @@ def get_file_data_and_close_file(file_instance: io.IOBase) -> bytes: file_instance.close() return file_data - def files_parameters(self, - files: typing.Optional[typing.Dict[str, - typing.List[io.IOBase]]] = None): + def files_parameters(self, files: typing.Optional[typing.Dict[str, typing.List[io.IOBase]]] = None): """Builds form parameters. :param files: None or a dict with key=param_name and @@ -597,7 +583,7 @@ def select_header_content_type(self, content_types, method=None, body=None): :return: Content-Type (e.g. application/json). """ if not content_types: - return None + return 'application/json' content_types = [x.lower() for x in content_types] @@ -612,7 +598,7 @@ def select_header_content_type(self, content_types, method=None, body=None): return content_types[0] def update_params_for_auth(self, headers, queries, auth_settings, - resource_path, method, body, request_auths=None): + resource_path, method, body): """Updates header and query params based on authentication setting. :param headers: Header parameters dict to be updated. @@ -622,36 +608,24 @@ def update_params_for_auth(self, headers, queries, auth_settings, :param method: A string representation of the HTTP request method. :param body: A object representing the body of the HTTP request. The object type is the return value of _encoder.default(). - :param request_auths: if set, the provided settings will - override the token in the configuration. """ if not auth_settings: return - if request_auths: - for auth_setting in request_auths: - self._apply_auth_params( - headers, queries, resource_path, method, body, auth_setting) - return - for auth in auth_settings: auth_setting = self.configuration.auth_settings().get(auth) if auth_setting: - self._apply_auth_params( - headers, queries, resource_path, method, body, auth_setting) - - def _apply_auth_params(self, headers, queries, resource_path, method, body, auth_setting): - if auth_setting['in'] == 'cookie': - headers['Cookie'] = auth_setting['key'] + "=" + auth_setting['value'] - elif auth_setting['in'] == 'header': - if auth_setting['type'] != 'http-signature': - headers[auth_setting['key']] = auth_setting['value'] - elif auth_setting['in'] == 'query': - queries.append((auth_setting['key'], auth_setting['value'])) - else: - raise ApiValueError( - 'Authentication token must be in `query` or `header`' - ) + if auth_setting['in'] == 'cookie': + headers['Cookie'] = auth_setting['value'] + elif auth_setting['in'] == 'header': + if auth_setting['type'] != 'http-signature': + headers[auth_setting['key']] = auth_setting['value'] + elif auth_setting['in'] == 'query': + queries.append((auth_setting['key'], auth_setting['value'])) + else: + raise ApiValueError( + 'Authentication token must be in `query` or `header`' + ) class Endpoint(object): @@ -701,8 +675,7 @@ def __init__(self, settings=None, params_map=None, root_map=None, '_check_input_type', '_check_return_type', '_content_type', - '_spec_property_naming', - '_request_auths' + '_spec_property_naming' ]) self.params_map['nullable'].extend(['_request_timeout']) self.validations = root_map['validations'] @@ -717,8 +690,7 @@ def __init__(self, settings=None, params_map=None, root_map=None, '_check_input_type': (bool,), '_check_return_type': (bool,), '_spec_property_naming': (bool,), - '_content_type': (none_type, str), - '_request_auths': (none_type, list) + '_content_type': (none_type, str) } self.openapi_types.update(extra_types) self.attribute_map = root_map['attribute_map'] @@ -871,11 +843,10 @@ def call_with_http_info(self, **kwargs): content_type_headers_list = self.headers_map['content_type'] if content_type_headers_list: if params['body'] != "": - content_types_list = self.api_client.select_header_content_type( + header_list = self.api_client.select_header_content_type( content_type_headers_list, self.settings['http_method'], params['body']) - if content_types_list: - params['header']['Content-Type'] = content_types_list + params['header']['Content-Type'] = header_list return self.api_client.call_api( self.settings['endpoint_path'], self.settings['http_method'], @@ -893,5 +864,4 @@ def call_with_http_info(self, **kwargs): _preload_content=kwargs['_preload_content'], _request_timeout=kwargs['_request_timeout'], _host=_host, - _request_auths=kwargs['_request_auths'], collection_formats=params['collection_format']) diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/apis/__init__.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/apis/__init__.py index f20ef0dc..14810ea0 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/apis/__init__.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/apis/__init__.py @@ -6,7 +6,7 @@ # raise a `RecursionError`. # In order to avoid this, import only the API that you directly need like: # -# from rhoas_kafka_mgmt_sdk.api.default_api import DefaultApi +# from .api.default_api import DefaultApi # # or import this package, but before doing it, use: # diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/exceptions.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/exceptions.py index 0bbf3bce..e1bf9288 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/exceptions.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/exceptions.py @@ -9,6 +9,7 @@ """ + class OpenApiException(Exception): """The base exception class for all OpenAPIExceptions""" @@ -112,7 +113,7 @@ def __init__(self, status=None, reason=None, http_resp=None): def __str__(self): """Custom error messages for exception""" - error_message = "Status Code: {0}\n"\ + error_message = "({0})\n"\ "Reason: {1}\n".format(self.status, self.reason) if self.headers: error_message += "HTTP response headers: {0}\n".format( diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/__init__.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/__init__.py index f18eedae..cfe32b78 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/__init__.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/__init__.py @@ -2,4 +2,4 @@ # reference which would not work in python2 # do not import all models into this module because that uses a lot of memory and stack frames # if you need the ability to import all models from one package, import them with -# from rhoas_kafka_mgmt_sdk.models import ModelA, ModelB +# from {{packageName}.models import ModelA, ModelB diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider.py index ad6234ab..147ed66e 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider.py @@ -153,7 +153,7 @@ def _from_openapi_data(cls, enabled, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -161,18 +161,14 @@ def _from_openapi_data(cls, enabled, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -252,18 +248,14 @@ def __init__(self, enabled, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list.py index db21b66d..793217b7 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list.py @@ -31,8 +31,10 @@ def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.cloud_provider import CloudProvider from rhoas_kafka_mgmt_sdk.model.cloud_provider_list_all_of import CloudProviderListAllOf from rhoas_kafka_mgmt_sdk.model.list import List + globals()['CloudProvider'] = CloudProvider globals()['CloudProviderListAllOf'] = CloudProviderListAllOf globals()['List'] = List @@ -165,18 +167,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -271,18 +269,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list_all_of.py index e6926d4d..abc750f2 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_provider_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.cloud_provider import CloudProvider + globals()['CloudProvider'] = CloudProvider + class CloudProviderListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'items': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 } @@ -139,7 +145,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -147,18 +153,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -231,18 +233,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region.py index ebef0657..ec9de2ba 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region.py @@ -160,7 +160,7 @@ def _from_openapi_data(cls, capacity, *args, **kwargs): # noqa: E501 enabled = kwargs.get('enabled', False) _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -168,18 +168,14 @@ def _from_openapi_data(cls, capacity, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -261,18 +257,14 @@ def __init__(self, capacity, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list.py index 944d9b9d..8a8e98a3 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list.py @@ -31,8 +31,10 @@ def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.cloud_region import CloudRegion from rhoas_kafka_mgmt_sdk.model.cloud_region_list_all_of import CloudRegionListAllOf from rhoas_kafka_mgmt_sdk.model.list import List + globals()['CloudRegion'] = CloudRegion globals()['CloudRegionListAllOf'] = CloudRegionListAllOf globals()['List'] = List @@ -165,18 +167,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -271,18 +269,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list_all_of.py index ff947927..3e7c5d45 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/cloud_region_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.cloud_region import CloudRegion + globals()['CloudRegion'] = CloudRegion + class CloudRegionListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'items': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 } @@ -139,7 +145,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -147,18 +153,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -231,18 +233,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster.py index 0af23c79..e0297e3f 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster.py @@ -168,18 +168,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -275,18 +271,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_all_of.py index 95553e6e..8204468d 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_all_of.py @@ -147,7 +147,7 @@ def _from_openapi_data(cls, access_kafkas_via_private_network, *args, **kwargs): """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -155,18 +155,14 @@ def _from_openapi_data(cls, access_kafkas_via_private_network, *args, **kwargs): self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -244,18 +240,14 @@ def __init__(self, access_kafkas_via_private_network, *args, **kwargs): # noqa: _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list.py index da305d56..827ab60b 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list.py @@ -31,8 +31,10 @@ def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.enterprise_cluster import EnterpriseCluster from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_list_all_of import EnterpriseClusterListAllOf from rhoas_kafka_mgmt_sdk.model.list import List + globals()['EnterpriseCluster'] = EnterpriseCluster globals()['EnterpriseClusterListAllOf'] = EnterpriseClusterListAllOf globals()['List'] = List @@ -165,18 +167,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -271,18 +269,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list_all_of.py index 9e71d30e..f5cf8bd5 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.enterprise_cluster import EnterpriseCluster + globals()['EnterpriseCluster'] = EnterpriseCluster + class EnterpriseClusterListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'items': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 } @@ -139,7 +145,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -147,18 +153,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -231,18 +233,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters.py similarity index 89% rename from app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response.py rename to app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters.py index 86428cfd..e3713e4e 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters.py @@ -31,13 +31,15 @@ def lazy_import(): - from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response_all_of import EnterpriseClusterRegistrationResponseAllOf + from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters_all_of import EnterpriseClusterWithAddonParametersAllOf + from rhoas_kafka_mgmt_sdk.model.fleetshard_parameter import FleetshardParameter from rhoas_kafka_mgmt_sdk.model.object_reference import ObjectReference - globals()['EnterpriseClusterRegistrationResponseAllOf'] = EnterpriseClusterRegistrationResponseAllOf + globals()['EnterpriseClusterWithAddonParametersAllOf'] = EnterpriseClusterWithAddonParametersAllOf + globals()['FleetshardParameter'] = FleetshardParameter globals()['ObjectReference'] = ObjectReference -class EnterpriseClusterRegistrationResponse(ModelComposed): +class EnterpriseClusterWithAddonParameters(ModelComposed): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech @@ -120,7 +122,7 @@ def discriminator(): @classmethod @convert_js_args_to_python_args def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 - """EnterpriseClusterRegistrationResponse - a model defined in OpenAPI + """EnterpriseClusterWithAddonParameters - a model defined in OpenAPI Keyword Args: id (str): @@ -171,18 +173,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -230,7 +228,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 @convert_js_args_to_python_args def __init__(self, *args, **kwargs): # noqa: E501 - """EnterpriseClusterRegistrationResponse - a model defined in OpenAPI + """EnterpriseClusterWithAddonParameters - a model defined in OpenAPI Keyword Args: id (str): @@ -279,18 +277,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -339,7 +333,7 @@ def _composed_schemas(): 'anyOf': [ ], 'allOf': [ - EnterpriseClusterRegistrationResponseAllOf, + EnterpriseClusterWithAddonParametersAllOf, ObjectReference, ], 'oneOf': [ diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters_all_of.py similarity index 90% rename from app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response_all_of.py rename to app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters_all_of.py index 79c2dbf9..d2574192 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_registration_response_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_cluster_with_addon_parameters_all_of.py @@ -30,8 +30,12 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.fleetshard_parameter import FleetshardParameter + globals()['FleetshardParameter'] = FleetshardParameter -class EnterpriseClusterRegistrationResponseAllOf(ModelNormal): + +class EnterpriseClusterWithAddonParametersAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'access_kafkas_via_private_network': (bool,), # noqa: E501 'cluster_id': (str,), # noqa: E501 @@ -108,7 +114,7 @@ def discriminator(): @classmethod @convert_js_args_to_python_args def _from_openapi_data(cls, access_kafkas_via_private_network, *args, **kwargs): # noqa: E501 - """EnterpriseClusterRegistrationResponseAllOf - a model defined in OpenAPI + """EnterpriseClusterWithAddonParametersAllOf - a model defined in OpenAPI Args: access_kafkas_via_private_network (bool): Indicates whether Kafkas created on this data plane cluster have to be accessed via private network @@ -150,7 +156,7 @@ def _from_openapi_data(cls, access_kafkas_via_private_network, *args, **kwargs): """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -158,18 +164,14 @@ def _from_openapi_data(cls, access_kafkas_via_private_network, *args, **kwargs): self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -200,7 +202,7 @@ def _from_openapi_data(cls, access_kafkas_via_private_network, *args, **kwargs): @convert_js_args_to_python_args def __init__(self, access_kafkas_via_private_network, *args, **kwargs): # noqa: E501 - """EnterpriseClusterRegistrationResponseAllOf - a model defined in OpenAPI + """EnterpriseClusterWithAddonParametersAllOf - a model defined in OpenAPI Args: access_kafkas_via_private_network (bool): Indicates whether Kafkas created on this data plane cluster have to be accessed via private network @@ -248,18 +250,14 @@ def __init__(self, access_kafkas_via_private_network, *args, **kwargs): # noqa: _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_osd_cluster_payload.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_osd_cluster_payload.py index 320d99fd..a6af6ad9 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_osd_cluster_payload.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/enterprise_osd_cluster_payload.py @@ -153,7 +153,7 @@ def _from_openapi_data(cls, access_kafkas_via_private_network, cluster_id, clust """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -161,18 +161,14 @@ def _from_openapi_data(cls, access_kafkas_via_private_network, cluster_id, clust self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -256,18 +252,14 @@ def __init__(self, access_kafkas_via_private_network, cluster_id, cluster_extern _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error.py index a1482eb8..547f7b82 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error.py @@ -156,7 +156,7 @@ def _from_openapi_data(cls, reason, id, kind, href, code, *args, **kwargs): # n """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -164,18 +164,14 @@ def _from_openapi_data(cls, reason, id, kind, href, code, *args, **kwargs): # n self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -260,18 +256,14 @@ def __init__(self, reason, id, kind, href, code, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list.py index 073e68ff..c312c383 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list.py @@ -167,18 +167,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -273,18 +269,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list_all_of.py index 48ef3c5e..a2b84b00 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/error_list_all_of.py @@ -145,7 +145,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -153,18 +153,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -237,18 +233,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/fleetshard_parameter.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/fleetshard_parameter.py index acd3fd6e..87eeb9e1 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/fleetshard_parameter.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/fleetshard_parameter.py @@ -142,7 +142,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -150,18 +150,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -235,18 +231,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/instant_query.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/instant_query.py index 3f552f7b..207570cc 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/instant_query.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/instant_query.py @@ -147,7 +147,7 @@ def _from_openapi_data(cls, value, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -155,18 +155,14 @@ def _from_openapi_data(cls, value, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -244,18 +240,14 @@ def __init__(self, value, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request.py index 3a87bf07..93beef17 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request.py @@ -248,18 +248,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -381,18 +377,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_all_of.py index 4dab9c18..28cd8a50 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_all_of.py @@ -231,7 +231,7 @@ def _from_openapi_data(cls, multi_az, reauthentication_enabled, *args, **kwargs) """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -239,18 +239,14 @@ def _from_openapi_data(cls, multi_az, reauthentication_enabled, *args, **kwargs) self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -355,18 +351,14 @@ def __init__(self, multi_az, reauthentication_enabled, *args, **kwargs): # noqa _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list.py index 41848514..0d1e107a 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list.py @@ -31,8 +31,10 @@ def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.kafka_request import KafkaRequest from rhoas_kafka_mgmt_sdk.model.kafka_request_list_all_of import KafkaRequestListAllOf from rhoas_kafka_mgmt_sdk.model.list import List + globals()['KafkaRequest'] = KafkaRequest globals()['KafkaRequestListAllOf'] = KafkaRequestListAllOf globals()['List'] = List @@ -165,18 +167,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -271,18 +269,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list_all_of.py index 02f65034..58f15461 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.kafka_request import KafkaRequest + globals()['KafkaRequest'] = KafkaRequest + class KafkaRequestListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'items': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 } @@ -139,7 +145,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -147,18 +153,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -231,18 +233,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_payload.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_payload.py index 0b499230..0f127be5 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_payload.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_request_payload.py @@ -162,7 +162,7 @@ def _from_openapi_data(cls, name, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -170,18 +170,14 @@ def _from_openapi_data(cls, name, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -264,18 +260,14 @@ def __init__(self, name, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_update_request.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_update_request.py index 15f55bed..7b13db40 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_update_request.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/kafka_update_request.py @@ -142,7 +142,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -150,18 +150,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -235,18 +231,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/list.py index 1980d029..7be4dfc1 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/list.py @@ -159,7 +159,7 @@ def _from_openapi_data(cls, kind, page, size, total, items, *args, **kwargs): # """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -167,18 +167,14 @@ def _from_openapi_data(cls, kind, page, size, total, items, *args, **kwargs): # self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -262,18 +258,14 @@ def __init__(self, kind, page, size, total, items, *args, **kwargs): # noqa: E5 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list.py index 6847cca6..0a0768f6 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list.py @@ -31,7 +31,9 @@ def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.instant_query import InstantQuery from rhoas_kafka_mgmt_sdk.model.metrics_instant_query_list_all_of import MetricsInstantQueryListAllOf + globals()['InstantQuery'] = InstantQuery globals()['MetricsInstantQueryListAllOf'] = MetricsInstantQueryListAllOf @@ -157,18 +159,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -261,18 +259,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list_all_of.py index e00b4fd8..179de51c 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_instant_query_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.instant_query import InstantQuery + globals()['InstantQuery'] = InstantQuery + class MetricsInstantQueryListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'kind': (str,), # noqa: E501 'id': (str,), # noqa: E501 @@ -145,7 +151,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -153,18 +159,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -239,18 +241,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list.py index dc500baf..9eddbff1 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list.py @@ -32,7 +32,9 @@ def lazy_import(): from rhoas_kafka_mgmt_sdk.model.metrics_range_query_list_all_of import MetricsRangeQueryListAllOf + from rhoas_kafka_mgmt_sdk.model.range_query import RangeQuery globals()['MetricsRangeQueryListAllOf'] = MetricsRangeQueryListAllOf + globals()['RangeQuery'] = RangeQuery class MetricsRangeQueryList(ModelComposed): @@ -157,18 +159,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -261,18 +259,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list_all_of.py index 8046afc2..22b8a8e7 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/metrics_range_query_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.range_query import RangeQuery + globals()['RangeQuery'] = RangeQuery + class MetricsRangeQueryListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'kind': (str,), # noqa: E501 'id': (str,), # noqa: E501 @@ -145,7 +151,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -153,18 +159,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -239,18 +241,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/object_reference.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/object_reference.py index 72f681ee..63d412d7 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/object_reference.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/object_reference.py @@ -147,7 +147,7 @@ def _from_openapi_data(cls, id, kind, href, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -155,18 +155,14 @@ def _from_openapi_data(cls, id, kind, href, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -246,18 +242,14 @@ def __init__(self, id, kind, href, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/range_query.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/range_query.py index 034f7227..026d57a8 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/range_query.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/range_query.py @@ -148,7 +148,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -156,18 +156,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -241,18 +237,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/region_capacity_list_item.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/region_capacity_list_item.py index 33215bd6..b89ca6f4 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/region_capacity_list_item.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/region_capacity_list_item.py @@ -144,7 +144,7 @@ def _from_openapi_data(cls, instance_type, available_sizes, *args, **kwargs): # """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -152,18 +152,14 @@ def _from_openapi_data(cls, instance_type, available_sizes, *args, **kwargs): # self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -241,18 +237,14 @@ def __init__(self, instance_type, available_sizes, *args, **kwargs): # noqa: E5 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account.py index a0e92862..839f69c2 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account.py @@ -162,8 +162,8 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 Animal class but this time we won't travel through its discriminator because we passed in _visited_composed_classes = (Animal,) - name (str): . [optional] # noqa: E501 - description (str): . [optional] # noqa: E501 + name (str): [optional] # noqa: E501 + description (str): [optional] # noqa: E501 client_id (str): [optional] # noqa: E501 client_secret (str): [optional] # noqa: E501 owner (str): [optional] # noqa: E501 @@ -180,18 +180,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -275,8 +271,8 @@ def __init__(self, *args, **kwargs): # noqa: E501 Animal class but this time we won't travel through its discriminator because we passed in _visited_composed_classes = (Animal,) - name (str): . [optional] # noqa: E501 - description (str): . [optional] # noqa: E501 + name (str): [optional] # noqa: E501 + description (str): [optional] # noqa: E501 client_id (str): [optional] # noqa: E501 client_secret (str): [optional] # noqa: E501 owner (str): [optional] # noqa: E501 @@ -291,18 +287,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_all_of.py index 6461076c..f2600693 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_all_of.py @@ -150,8 +150,8 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 through its discriminator because we passed in _visited_composed_classes = (Animal,) id (str): server generated unique id of the service account. [optional] # noqa: E501 - name (str): . [optional] # noqa: E501 - description (str): . [optional] # noqa: E501 + name (str): [optional] # noqa: E501 + description (str): [optional] # noqa: E501 client_id (str): [optional] # noqa: E501 client_secret (str): [optional] # noqa: E501 owner (str): [optional] # noqa: E501 @@ -160,7 +160,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -168,18 +168,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -243,8 +239,8 @@ def __init__(self, *args, **kwargs): # noqa: E501 through its discriminator because we passed in _visited_composed_classes = (Animal,) id (str): server generated unique id of the service account. [optional] # noqa: E501 - name (str): . [optional] # noqa: E501 - description (str): . [optional] # noqa: E501 + name (str): [optional] # noqa: E501 + description (str): [optional] # noqa: E501 client_id (str): [optional] # noqa: E501 client_secret (str): [optional] # noqa: E501 owner (str): [optional] # noqa: E501 @@ -259,18 +255,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list.py index d521298c..7722f1f7 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list.py @@ -32,7 +32,9 @@ def lazy_import(): from rhoas_kafka_mgmt_sdk.model.service_account_list_all_of import ServiceAccountListAllOf + from rhoas_kafka_mgmt_sdk.model.service_account_list_item import ServiceAccountListItem globals()['ServiceAccountListAllOf'] = ServiceAccountListAllOf + globals()['ServiceAccountListItem'] = ServiceAccountListItem class ServiceAccountList(ModelComposed): @@ -154,18 +156,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -257,18 +255,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_all_of.py index e04caf44..a4a6d1fb 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.service_account_list_item import ServiceAccountListItem + globals()['ServiceAccountListItem'] = ServiceAccountListItem + class ServiceAccountListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'kind': (str,), # noqa: E501 'items': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 @@ -144,7 +150,7 @@ def _from_openapi_data(cls, kind, items, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -152,18 +158,14 @@ def _from_openapi_data(cls, kind, items, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -241,18 +243,14 @@ def __init__(self, kind, items, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item.py index 7a0da5b4..b897e12e 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item.py @@ -177,18 +177,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -287,18 +283,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item_all_of.py index fbdee08f..103ede0e 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_list_item_all_of.py @@ -157,7 +157,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -165,18 +165,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -255,18 +251,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_request.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_request.py index 5abd5d34..997ead15 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_request.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/service_account_request.py @@ -144,7 +144,7 @@ def _from_openapi_data(cls, name, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -152,18 +152,14 @@ def _from_openapi_data(cls, name, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -240,18 +236,14 @@ def __init__(self, name, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider.py index b0ac2325..3f1f6ac2 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider.py @@ -174,18 +174,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -283,18 +279,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider_all_of.py index 6a016270..c3aba32e 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/sso_provider_all_of.py @@ -151,7 +151,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -159,18 +159,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -247,18 +243,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_billing_model.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_billing_model.py index 0e289f9a..9a09b893 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_billing_model.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_billing_model.py @@ -162,7 +162,7 @@ def _from_openapi_data(cls, id, ams_resource, ams_product, ams_billing_models, * """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -170,18 +170,14 @@ def _from_openapi_data(cls, id, ams_resource, ams_product, ams_billing_models, * self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -263,18 +259,14 @@ def __init__(self, id, ams_resource, ams_product, ams_billing_models, *args, **k _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type.py index 5ecf659f..0aa70d4c 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type.py @@ -32,9 +32,9 @@ def lazy_import(): from rhoas_kafka_mgmt_sdk.model.supported_kafka_billing_model import SupportedKafkaBillingModel - from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type_sizes_inner import SupportedKafkaInstanceTypeSizesInner + from rhoas_kafka_mgmt_sdk.model.supported_kafka_size import SupportedKafkaSize globals()['SupportedKafkaBillingModel'] = SupportedKafkaBillingModel - globals()['SupportedKafkaInstanceTypeSizesInner'] = SupportedKafkaInstanceTypeSizesInner + globals()['SupportedKafkaSize'] = SupportedKafkaSize class SupportedKafkaInstanceType(ModelNormal): @@ -96,7 +96,7 @@ def openapi_types(): 'supported_billing_models': ([SupportedKafkaBillingModel],), # noqa: E501 'id': (str,), # noqa: E501 'display_name': (str,), # noqa: E501 - 'sizes': ([SupportedKafkaInstanceTypeSizesInner],), # noqa: E501 + 'sizes': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 } @cached_property @@ -157,11 +157,11 @@ def _from_openapi_data(cls, supported_billing_models, *args, **kwargs): # noqa: _visited_composed_classes = (Animal,) id (str): Unique identifier of the Kafka instance type.. [optional] # noqa: E501 display_name (str): Human readable name of the supported Kafka instance type. [optional] # noqa: E501 - sizes ([SupportedKafkaInstanceTypeSizesInner]): A list of Kafka instance sizes available for this instance type. [optional] # noqa: E501 + sizes ([bool, date, datetime, dict, float, int, list, str, none_type]): A list of Kafka instance sizes available for this instance type. [optional] # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -169,18 +169,14 @@ def _from_openapi_data(cls, supported_billing_models, *args, **kwargs): # noqa: self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -249,7 +245,7 @@ def __init__(self, supported_billing_models, *args, **kwargs): # noqa: E501 _visited_composed_classes = (Animal,) id (str): Unique identifier of the Kafka instance type.. [optional] # noqa: E501 display_name (str): Human readable name of the supported Kafka instance type. [optional] # noqa: E501 - sizes ([SupportedKafkaInstanceTypeSizesInner]): A list of Kafka instance sizes available for this instance type. [optional] # noqa: E501 + sizes ([bool, date, datetime, dict, float, int, list, str, none_type]): A list of Kafka instance sizes available for this instance type. [optional] # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) @@ -259,18 +255,14 @@ def __init__(self, supported_billing_models, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type_sizes_inner.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type_sizes_inner.py deleted file mode 100644 index 6b160177..00000000 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_type_sizes_inner.py +++ /dev/null @@ -1,390 +0,0 @@ -""" - Kafka Management API - - Kafka Management API is a REST API to manage Kafka instances # noqa: E501 - - The version of the OpenAPI document: 1.14.0 - Contact: rhosak-support@redhat.com - Generated by: https://openapi-generator.tech -""" - - -import re # noqa: F401 -import sys # noqa: F401 - -from rhoas_kafka_mgmt_sdk.model_utils import ( # noqa: F401 - ApiTypeError, - ModelComposed, - ModelNormal, - ModelSimple, - cached_property, - change_keys_js_to_python, - convert_js_args_to_python_args, - date, - datetime, - file_type, - none_type, - validate_get_composed_info, - OpenApiModel -) -from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError - - -def lazy_import(): - from rhoas_kafka_mgmt_sdk.model.supported_kafka_size import SupportedKafkaSize - from rhoas_kafka_mgmt_sdk.model.supported_kafka_size_bytes_value_item import SupportedKafkaSizeBytesValueItem - globals()['SupportedKafkaSize'] = SupportedKafkaSize - globals()['SupportedKafkaSizeBytesValueItem'] = SupportedKafkaSizeBytesValueItem - - -class SupportedKafkaInstanceTypeSizesInner(ModelComposed): - """NOTE: This class is auto generated by OpenAPI Generator. - Ref: https://openapi-generator.tech - - Do not edit the class manually. - - Attributes: - allowed_values (dict): The key is the tuple path to the attribute - and the for var_name this is (var_name,). The value is a dict - with a capitalized key describing the allowed value and an allowed - value. These dicts store the allowed enum values. - attribute_map (dict): The key is attribute name - and the value is json key in definition. - discriminator_value_class_map (dict): A dict to go from the discriminator - variable value to the discriminator class name. - validations (dict): The key is the tuple path to the attribute - and the for var_name this is (var_name,). The value is a dict - that stores validations for max_length, min_length, max_items, - min_items, exclusive_maximum, inclusive_maximum, exclusive_minimum, - inclusive_minimum, and regex. - additional_properties_type (tuple): A tuple of classes accepted - as additional properties values. - """ - - allowed_values = { - } - - validations = { - } - - @cached_property - def additional_properties_type(): - """ - This must be a method because a model may have properties that are - of type self, this must run after the class is loaded - """ - lazy_import() - return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 - - _nullable = False - - @cached_property - def openapi_types(): - """ - This must be a method because a model may have properties that are - of type self, this must run after the class is loaded - - Returns - openapi_types (dict): The key is attribute name - and the value is attribute type. - """ - lazy_import() - return { - 'id': (str,), # noqa: E501 - 'display_name': (str,), # noqa: E501 - 'ingress_throughput_per_sec': (SupportedKafkaSizeBytesValueItem,), # noqa: E501 - 'egress_throughput_per_sec': (SupportedKafkaSizeBytesValueItem,), # noqa: E501 - 'total_max_connections': (int,), # noqa: E501 - 'max_data_retention_size': (SupportedKafkaSizeBytesValueItem,), # noqa: E501 - 'max_partitions': (int,), # noqa: E501 - 'max_data_retention_period': (str,), # noqa: E501 - 'max_connection_attempts_per_sec': (int,), # noqa: E501 - 'max_message_size': (SupportedKafkaSizeBytesValueItem,), # noqa: E501 - 'min_in_sync_replicas': (int,), # noqa: E501 - 'replication_factor': (int,), # noqa: E501 - 'supported_az_modes': ([str],), # noqa: E501 - 'lifespan_seconds': (int, none_type,), # noqa: E501 - 'quota_consumed': (int,), # noqa: E501 - 'quota_type': (str,), # noqa: E501 - 'capacity_consumed': (int,), # noqa: E501 - 'maturity_status': (str,), # noqa: E501 - } - - @cached_property - def discriminator(): - return None - - - attribute_map = { - 'id': 'id', # noqa: E501 - 'display_name': 'display_name', # noqa: E501 - 'ingress_throughput_per_sec': 'ingress_throughput_per_sec', # noqa: E501 - 'egress_throughput_per_sec': 'egress_throughput_per_sec', # noqa: E501 - 'total_max_connections': 'total_max_connections', # noqa: E501 - 'max_data_retention_size': 'max_data_retention_size', # noqa: E501 - 'max_partitions': 'max_partitions', # noqa: E501 - 'max_data_retention_period': 'max_data_retention_period', # noqa: E501 - 'max_connection_attempts_per_sec': 'max_connection_attempts_per_sec', # noqa: E501 - 'max_message_size': 'max_message_size', # noqa: E501 - 'min_in_sync_replicas': 'min_in_sync_replicas', # noqa: E501 - 'replication_factor': 'replication_factor', # noqa: E501 - 'supported_az_modes': 'supported_az_modes', # noqa: E501 - 'lifespan_seconds': 'lifespan_seconds', # noqa: E501 - 'quota_consumed': 'quota_consumed', # noqa: E501 - 'quota_type': 'quota_type', # noqa: E501 - 'capacity_consumed': 'capacity_consumed', # noqa: E501 - 'maturity_status': 'maturity_status', # noqa: E501 - } - - read_only_vars = { - } - - @classmethod - @convert_js_args_to_python_args - def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 - """SupportedKafkaInstanceTypeSizesInner - a model defined in OpenAPI - - Keyword Args: - _check_type (bool): if True, values for parameters in openapi_types - will be type checked and a TypeError will be - raised if the wrong type is input. - Defaults to True - _path_to_item (tuple/list): This is a list of keys or values to - drill down to the model in received_data - when deserializing a response - _spec_property_naming (bool): True if the variable names in the input data - are serialized names, as specified in the OpenAPI document. - False if the variable names in the input data - are pythonic names, e.g. snake case (default) - _configuration (Configuration): the instance to use when - deserializing a file_type parameter. - If passed, type conversion is attempted - If omitted no type conversion is done. - _visited_composed_classes (tuple): This stores a tuple of - classes that we have traveled through so that - if we see that class again we will not use its - discriminator again. - When traveling through a discriminator, the - composed schema that is - is traveled through is added to this set. - For example if Animal has a discriminator - petType and we pass in "Dog", and the class Dog - allOf includes Animal, we move through Animal - once using the discriminator, and pick Dog. - Then in Dog, we will make an instance of the - Animal class but this time we won't travel - through its discriminator because we passed in - _visited_composed_classes = (Animal,) - id (str): Unique identifier of this Kafka instance size.. [optional] # noqa: E501 - display_name (str): Display name of this Kafka instance size.. [optional] # noqa: E501 - ingress_throughput_per_sec (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - egress_throughput_per_sec (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - total_max_connections (int): Maximum amount of total connections available to this Kafka instance size.. [optional] # noqa: E501 - max_data_retention_size (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - max_partitions (int): Maximum amount of total partitions available to this Kafka instance size.. [optional] # noqa: E501 - max_data_retention_period (str): Maximum data retention period available to this Kafka instance size.. [optional] # noqa: E501 - max_connection_attempts_per_sec (int): Maximium connection attempts per second available to this Kafka instance size.. [optional] # noqa: E501 - max_message_size (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - min_in_sync_replicas (int): Minimum number of in-sync replicas.. [optional] # noqa: E501 - replication_factor (int): Replication factor available to this Kafka instance size.. [optional] # noqa: E501 - supported_az_modes ([str]): List of Availability Zone modes that this Kafka instance size supports. The possible values are \"single\", \"multi\".. [optional] # noqa: E501 - lifespan_seconds (int, none_type): The limit lifespan of the kafka instance in seconds. If not specified then the instance never expires.. [optional] # noqa: E501 - quota_consumed (int): Quota consumed by this Kafka instance size.. [optional] # noqa: E501 - quota_type (str): Quota type used by this Kafka instance size. This is now deprecated, please refer to supported_billing_models at instance-type level instead.. [optional] # noqa: E501 - capacity_consumed (int): Data plane cluster capacity consumed by this Kafka instance size.. [optional] # noqa: E501 - maturity_status (str): Maturity level of the size. Can be \"stable\" or \"preview\".. [optional] # noqa: E501 - """ - - _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', False) - _path_to_item = kwargs.pop('_path_to_item', ()) - _configuration = kwargs.pop('_configuration', None) - _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) - - self = super(OpenApiModel, cls).__new__(cls) - - if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) - - self._data_store = {} - self._check_type = _check_type - self._spec_property_naming = _spec_property_naming - self._path_to_item = _path_to_item - self._configuration = _configuration - self._visited_composed_classes = _visited_composed_classes + (self.__class__,) - - constant_args = { - '_check_type': _check_type, - '_path_to_item': _path_to_item, - '_spec_property_naming': _spec_property_naming, - '_configuration': _configuration, - '_visited_composed_classes': self._visited_composed_classes, - } - composed_info = validate_get_composed_info( - constant_args, kwargs, self) - self._composed_instances = composed_info[0] - self._var_name_to_model_instances = composed_info[1] - self._additional_properties_model_instances = composed_info[2] - discarded_args = composed_info[3] - - for var_name, var_value in kwargs.items(): - if var_name in discarded_args and \ - self._configuration is not None and \ - self._configuration.discard_unknown_keys and \ - self._additional_properties_model_instances: - # discard variable. - continue - setattr(self, var_name, var_value) - - return self - - required_properties = set([ - '_data_store', - '_check_type', - '_spec_property_naming', - '_path_to_item', - '_configuration', - '_visited_composed_classes', - '_composed_instances', - '_var_name_to_model_instances', - '_additional_properties_model_instances', - ]) - - @convert_js_args_to_python_args - def __init__(self, *args, **kwargs): # noqa: E501 - """SupportedKafkaInstanceTypeSizesInner - a model defined in OpenAPI - - Keyword Args: - _check_type (bool): if True, values for parameters in openapi_types - will be type checked and a TypeError will be - raised if the wrong type is input. - Defaults to True - _path_to_item (tuple/list): This is a list of keys or values to - drill down to the model in received_data - when deserializing a response - _spec_property_naming (bool): True if the variable names in the input data - are serialized names, as specified in the OpenAPI document. - False if the variable names in the input data - are pythonic names, e.g. snake case (default) - _configuration (Configuration): the instance to use when - deserializing a file_type parameter. - If passed, type conversion is attempted - If omitted no type conversion is done. - _visited_composed_classes (tuple): This stores a tuple of - classes that we have traveled through so that - if we see that class again we will not use its - discriminator again. - When traveling through a discriminator, the - composed schema that is - is traveled through is added to this set. - For example if Animal has a discriminator - petType and we pass in "Dog", and the class Dog - allOf includes Animal, we move through Animal - once using the discriminator, and pick Dog. - Then in Dog, we will make an instance of the - Animal class but this time we won't travel - through its discriminator because we passed in - _visited_composed_classes = (Animal,) - id (str): Unique identifier of this Kafka instance size.. [optional] # noqa: E501 - display_name (str): Display name of this Kafka instance size.. [optional] # noqa: E501 - ingress_throughput_per_sec (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - egress_throughput_per_sec (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - total_max_connections (int): Maximum amount of total connections available to this Kafka instance size.. [optional] # noqa: E501 - max_data_retention_size (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - max_partitions (int): Maximum amount of total partitions available to this Kafka instance size.. [optional] # noqa: E501 - max_data_retention_period (str): Maximum data retention period available to this Kafka instance size.. [optional] # noqa: E501 - max_connection_attempts_per_sec (int): Maximium connection attempts per second available to this Kafka instance size.. [optional] # noqa: E501 - max_message_size (SupportedKafkaSizeBytesValueItem): [optional] # noqa: E501 - min_in_sync_replicas (int): Minimum number of in-sync replicas.. [optional] # noqa: E501 - replication_factor (int): Replication factor available to this Kafka instance size.. [optional] # noqa: E501 - supported_az_modes ([str]): List of Availability Zone modes that this Kafka instance size supports. The possible values are \"single\", \"multi\".. [optional] # noqa: E501 - lifespan_seconds (int, none_type): The limit lifespan of the kafka instance in seconds. If not specified then the instance never expires.. [optional] # noqa: E501 - quota_consumed (int): Quota consumed by this Kafka instance size.. [optional] # noqa: E501 - quota_type (str): Quota type used by this Kafka instance size. This is now deprecated, please refer to supported_billing_models at instance-type level instead.. [optional] # noqa: E501 - capacity_consumed (int): Data plane cluster capacity consumed by this Kafka instance size.. [optional] # noqa: E501 - maturity_status (str): Maturity level of the size. Can be \"stable\" or \"preview\".. [optional] # noqa: E501 - """ - - _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', False) - _path_to_item = kwargs.pop('_path_to_item', ()) - _configuration = kwargs.pop('_configuration', None) - _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) - - if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) - - self._data_store = {} - self._check_type = _check_type - self._spec_property_naming = _spec_property_naming - self._path_to_item = _path_to_item - self._configuration = _configuration - self._visited_composed_classes = _visited_composed_classes + (self.__class__,) - - constant_args = { - '_check_type': _check_type, - '_path_to_item': _path_to_item, - '_spec_property_naming': _spec_property_naming, - '_configuration': _configuration, - '_visited_composed_classes': self._visited_composed_classes, - } - composed_info = validate_get_composed_info( - constant_args, kwargs, self) - self._composed_instances = composed_info[0] - self._var_name_to_model_instances = composed_info[1] - self._additional_properties_model_instances = composed_info[2] - discarded_args = composed_info[3] - - for var_name, var_value in kwargs.items(): - if var_name in discarded_args and \ - self._configuration is not None and \ - self._configuration.discard_unknown_keys and \ - self._additional_properties_model_instances: - # discard variable. - continue - setattr(self, var_name, var_value) - if var_name in self.read_only_vars: - raise ApiAttributeError(f"`{var_name}` is a read-only attribute. Use `from_openapi_data` to instantiate " - f"class with read only attributes.") - - @cached_property - def _composed_schemas(): - # we need this here to make our import statements work - # we must store _composed_schemas in here so the code is only run - # when we invoke this method. If we kept this at the class - # level we would get an error because the class level - # code would be run when this module is imported, and these composed - # classes don't exist yet because their module has not finished - # loading - lazy_import() - return { - 'anyOf': [ - ], - 'allOf': [ - SupportedKafkaSize, - ], - 'oneOf': [ - ], - } diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list.py index 95590162..0c39ca13 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list.py @@ -31,7 +31,9 @@ def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type import SupportedKafkaInstanceType from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_types_list_all_of import SupportedKafkaInstanceTypesListAllOf + globals()['SupportedKafkaInstanceType'] = SupportedKafkaInstanceType globals()['SupportedKafkaInstanceTypesListAllOf'] = SupportedKafkaInstanceTypesListAllOf @@ -151,18 +153,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -253,18 +251,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list_all_of.py index 9a2694ef..f43beaae 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_instance_types_list_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type import SupportedKafkaInstanceType + globals()['SupportedKafkaInstanceType'] = SupportedKafkaInstanceType + class SupportedKafkaInstanceTypesListAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'instance_types': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 } @@ -139,7 +145,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -147,18 +153,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -231,18 +233,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size.py index 8721ace3..d16780ea 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size.py @@ -196,7 +196,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -204,18 +204,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -305,18 +301,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size_bytes_value_item.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size_bytes_value_item.py index 3ba09aee..0286ba51 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size_bytes_value_item.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/supported_kafka_size_bytes_value_item.py @@ -139,7 +139,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -147,18 +147,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -231,18 +227,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/values.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/values.py index 8617fe8f..7a6c84c7 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/values.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/values.py @@ -144,7 +144,7 @@ def _from_openapi_data(cls, value, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -152,18 +152,14 @@ def _from_openapi_data(cls, value, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -240,18 +236,14 @@ def __init__(self, value, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata.py index 1a7a03a2..4785c50a 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata.py @@ -165,18 +165,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -271,18 +267,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata_all_of.py index 0b8f43f8..009bbe89 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model/version_metadata_all_of.py @@ -30,6 +30,10 @@ from rhoas_kafka_mgmt_sdk.exceptions import ApiAttributeError +def lazy_import(): + from rhoas_kafka_mgmt_sdk.model.object_reference import ObjectReference + globals()['ObjectReference'] = ObjectReference + class VersionMetadataAllOf(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. @@ -67,6 +71,7 @@ def additional_properties_type(): This must be a method because a model may have properties that are of type self, this must run after the class is loaded """ + lazy_import() return (bool, date, datetime, dict, float, int, list, str, none_type,) # noqa: E501 _nullable = False @@ -81,6 +86,7 @@ def openapi_types(): openapi_types (dict): The key is attribute name and the value is attribute type. """ + lazy_import() return { 'server_version': (str,), # noqa: E501 'collections': ([bool, date, datetime, dict, float, int, list, str, none_type],), # noqa: E501 @@ -142,7 +148,7 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 """ _check_type = kwargs.pop('_check_type', True) - _spec_property_naming = kwargs.pop('_spec_property_naming', True) + _spec_property_naming = kwargs.pop('_spec_property_naming', False) _path_to_item = kwargs.pop('_path_to_item', ()) _configuration = kwargs.pop('_configuration', None) _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) @@ -150,18 +156,14 @@ def _from_openapi_data(cls, *args, **kwargs): # noqa: E501 self = super(OpenApiModel, cls).__new__(cls) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type @@ -235,18 +237,14 @@ def __init__(self, *args, **kwargs): # noqa: E501 _visited_composed_classes = kwargs.pop('_visited_composed_classes', ()) if args: - for arg in args: - if isinstance(arg, dict): - kwargs.update(arg) - else: - raise ApiTypeError( - "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( - args, - self.__class__.__name__, - ), - path_to_item=_path_to_item, - valid_classes=(self.__class__,), - ) + raise ApiTypeError( + "Invalid positional arguments=%s passed to %s. Remove those invalid positional arguments." % ( + args, + self.__class__.__name__, + ), + path_to_item=_path_to_item, + valid_classes=(self.__class__,), + ) self._data_store = {} self._check_type = _check_type diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model_utils.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model_utils.py index e3d01e2a..1057628c 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model_utils.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/model_utils.py @@ -17,7 +17,6 @@ import pprint import re import tempfile -import uuid from dateutil.parser import parse @@ -43,9 +42,7 @@ def wrapped_init(_self, *args, **kwargs): """ spec_property_naming = kwargs.get('_spec_property_naming', False) if spec_property_naming: - kwargs = change_keys_js_to_python( - kwargs, _self if isinstance( - _self, type) else _self.__class__) + kwargs = change_keys_js_to_python(kwargs, _self if isinstance(_self, type) else _self.__class__) return fn(_self, *args, **kwargs) return wrapped_init @@ -70,7 +67,6 @@ def __get__(self, instance, cls=None): PRIMITIVE_TYPES = (list, float, int, bool, datetime, date, str, file_type) - def allows_single_value_input(cls): """ This function returns True if the input composed schema model or any @@ -95,7 +91,6 @@ def allows_single_value_input(cls): return any(allows_single_value_input(c) for c in cls._composed_schemas['oneOf']) return False - def composed_model_input_classes(cls): """ This function returns a list of the possible models that can be accepted as @@ -198,7 +193,7 @@ def __copy__(self): if self.get("_spec_property_naming", False): return cls._new_from_openapi_data(**self.__dict__) else: - return cls.__new__(cls, **self.__dict__) + return new_cls.__new__(cls, **self.__dict__) def __deepcopy__(self, memo): cls = self.__class__ @@ -206,7 +201,7 @@ def __deepcopy__(self, memo): if self.get("_spec_property_naming", False): new_inst = cls._new_from_openapi_data() else: - new_inst = cls.__new__(cls, **self.__dict__) + new_inst = cls.__new__(cls) for k, v in self.__dict__.items(): setattr(new_inst, k, deepcopy(v, memo)) @@ -229,6 +224,7 @@ def __new__(cls, *args, **kwargs): oneof_instance = get_oneof_instance(cls, model_kwargs, kwargs, model_arg=arg) return oneof_instance + visited_composed_classes = kwargs.get('_visited_composed_classes', ()) if ( cls.discriminator is None or @@ -275,7 +271,7 @@ def __new__(cls, *args, **kwargs): # call itself and update the list of visited classes, and the initial # value must be an empty list. Hence not using 'visited_composed_classes' new_cls = get_discriminator_class( - cls, discr_propertyname_py, discr_value, []) + cls, discr_propertyname_py, discr_value, []) if new_cls is None: path_to_item = kwargs.get('_path_to_item', ()) disc_prop_value = kwargs.get( @@ -330,6 +326,7 @@ def __new__(cls, *args, **kwargs): return new_inst + @classmethod @convert_js_args_to_python_args def _new_from_openapi_data(cls, *args, **kwargs): @@ -348,6 +345,7 @@ def _new_from_openapi_data(cls, *args, **kwargs): oneof_instance = get_oneof_instance(cls, model_kwargs, kwargs, model_arg=arg) return oneof_instance + visited_composed_classes = kwargs.get('_visited_composed_classes', ()) if ( cls.discriminator is None or @@ -394,7 +392,7 @@ def _new_from_openapi_data(cls, *args, **kwargs): # call itself and update the list of visited classes, and the initial # value must be an empty list. Hence not using 'visited_composed_classes' new_cls = get_discriminator_class( - cls, discr_propertyname_py, discr_value, []) + cls, discr_propertyname_py, discr_value, []) if new_cls is None: path_to_item = kwargs.get('_path_to_item', ()) disc_prop_value = kwargs.get( @@ -439,6 +437,7 @@ def _new_from_openapi_data(cls, *args, **kwargs): # new_cls it will not include the allOf validations in self self_inst = cls._from_openapi_data(*args, **kwargs) + new_inst = new_cls._new_from_openapi_data(*args, **kwargs) return new_inst @@ -739,8 +738,7 @@ def __eq__(self, other): UPCONVERSION_TYPE_PAIRS = ( (str, datetime), (str, date), - # A float may be serialized as an integer, e.g. '3' is a valid serialized float. - (int, float), + (int, float), # A float may be serialized as an integer, e.g. '3' is a valid serialized float. (list, ModelComposed), (dict, ModelComposed), (str, ModelComposed), @@ -893,8 +891,8 @@ def is_json_validation_enabled(schema_keyword, configuration=None): """ return (configuration is None or - not hasattr(configuration, '_disabled_client_side_validations') or - schema_keyword not in configuration._disabled_client_side_validations) + not hasattr(configuration, '_disabled_client_side_validations') or + schema_keyword not in configuration._disabled_client_side_validations) def check_validations( @@ -1032,9 +1030,9 @@ def check_validations( not re.search(current_validations['regex']['pattern'], input_values, flags=flags)): err_msg = r"Invalid value for `%s`, must match regular expression `%s`" % ( - input_variable_path[0], - current_validations['regex']['pattern'] - ) + input_variable_path[0], + current_validations['regex']['pattern'] + ) if flags != 0: # Don't print the regex flags if the flags are not # specified in the OAS document. @@ -1125,7 +1123,6 @@ def remove_uncoercible(required_types_classes, current_item, spec_property_namin results_classes.append(required_type_class) return results_classes - def get_discriminated_classes(cls): """ Returns all the classes that a discriminator converts to @@ -1324,7 +1321,7 @@ def get_discriminator_class(model_class, # Descendant example: mammal -> whale/zebra/Pig -> BasquePig/DanishPig # if we try to make BasquePig from mammal, we need to travel through # the oneOf descendant discriminators to find BasquePig - descendant_classes = model_class._composed_schemas.get('oneOf', ()) + \ + descendant_classes = model_class._composed_schemas.get('oneOf', ()) + \ model_class._composed_schemas.get('anyOf', ()) ancestor_classes = model_class._composed_schemas.get('allOf', ()) possible_classes = descendant_classes + ancestor_classes @@ -1332,7 +1329,7 @@ def get_discriminator_class(model_class, # Check if the schema has inherited discriminators. if hasattr(cls, 'discriminator') and cls.discriminator is not None: used_model_class = get_discriminator_class( - cls, discr_name, discr_value, cls_visited) + cls, discr_name, discr_value, cls_visited) if used_model_class is not None: return used_model_class return used_model_class @@ -1403,13 +1400,7 @@ def deserialize_file(response_data, configuration, content_disposition=None): if content_disposition: filename = re.search(r'filename=[\'"]?([^\'"\s]+)[\'"]?', - content_disposition, - flags=re.I) - if filename is not None: - filename = filename.group(1) - else: - filename = "default_" + str(uuid.uuid4()) - + content_disposition).group(1) path = os.path.join(os.path.dirname(path), filename) with open(path, "wb") as f: @@ -1499,11 +1490,9 @@ def is_type_nullable(input_type): if issubclass(input_type, ModelComposed): # If oneOf/anyOf, check if the 'null' type is one of the allowed types. for t in input_type._composed_schemas.get('oneOf', ()): - if is_type_nullable(t): - return True + if is_type_nullable(t): return True for t in input_type._composed_schemas.get('anyOf', ()): - if is_type_nullable(t): - return True + if is_type_nullable(t): return True return False @@ -1518,7 +1507,7 @@ def is_valid_type(input_class_simple, valid_classes): bool """ if issubclass(input_class_simple, OpenApiModel) and \ - valid_classes == (bool, date, datetime, dict, float, int, list, str, none_type,): + valid_classes == (bool, date, datetime, dict, float, int, list, str, none_type,): return True valid_type = input_class_simple in valid_classes if not valid_type and ( @@ -1576,9 +1565,7 @@ def validate_and_convert_types(input_value, required_types_mixed, path_to_item, input_class_simple = get_simple_class(input_value) valid_type = is_valid_type(input_class_simple, valid_classes) if not valid_type: - if (configuration - or (input_class_simple == dict - and dict not in valid_classes)): + if configuration: # if input_value is not valid_type try to convert it converted_instance = attempt_convert_item( input_value, @@ -1671,11 +1658,7 @@ def model_to_dict(model_instance, serialize=True): attribute_map """ result = {} - - def extract_item(item): return ( - item[0], model_to_dict( - item[1], serialize=serialize)) if hasattr( - item[1], '_data_store') else item + extract_item = lambda item: (item[0], model_to_dict(item[1], serialize=serialize)) if hasattr(item[1], '_data_store') else item model_instances = [model_instance] if model_instance._composed_schemas: @@ -1695,24 +1678,24 @@ def extract_item(item): return ( except KeyError: used_fallback_python_attribute_names.add(attr) if isinstance(value, list): - if not value: - # empty list or None - result[attr] = value - else: - res = [] - for v in value: - if isinstance(v, PRIMITIVE_TYPES) or v is None: - res.append(v) - elif isinstance(v, ModelSimple): - res.append(v.value) - elif isinstance(v, dict): - res.append(dict(map( - extract_item, - v.items() - ))) - else: - res.append(model_to_dict(v, serialize=serialize)) - result[attr] = res + if not value: + # empty list or None + result[attr] = value + else: + res = [] + for v in value: + if isinstance(v, PRIMITIVE_TYPES) or v is None: + res.append(v) + elif isinstance(v, ModelSimple): + res.append(v.value) + elif isinstance(v, dict): + res.append(dict(map( + extract_item, + v.items() + ))) + else: + res.append(model_to_dict(v, serialize=serialize)) + result[attr] = res elif isinstance(value, dict): result[attr] = dict(map( extract_item, @@ -1864,15 +1847,13 @@ def get_oneof_instance(cls, model_kwargs, constant_kwargs, model_arg=None): try: if not single_value_input: if constant_kwargs.get('_spec_property_naming'): - oneof_instance = oneof_class._from_openapi_data( - **model_kwargs, **constant_kwargs) + oneof_instance = oneof_class._from_openapi_data(**model_kwargs, **constant_kwargs) else: oneof_instance = oneof_class(**model_kwargs, **constant_kwargs) else: if issubclass(oneof_class, ModelSimple): if constant_kwargs.get('_spec_property_naming'): - oneof_instance = oneof_class._from_openapi_data( - model_arg, **constant_kwargs) + oneof_instance = oneof_class._from_openapi_data(model_arg, **constant_kwargs) else: oneof_instance = oneof_class(model_arg, **constant_kwargs) elif oneof_class in PRIMITIVE_TYPES: @@ -2047,13 +2028,11 @@ def validate_get_composed_info(constant_args, model_args, self): var_name_to_model_instances = {} for prop_name in model_args: if prop_name not in discarded_args: - var_name_to_model_instances[prop_name] = [self] + list( - filter( - lambda x: prop_name in x.openapi_types, composed_instances)) + var_name_to_model_instances[prop_name] = [self] + composed_instances return [ - composed_instances, - var_name_to_model_instances, - additional_properties_model_instances, - discarded_args + composed_instances, + var_name_to_model_instances, + additional_properties_model_instances, + discarded_args ] diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/models/__init__.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/models/__init__.py index eb94320a..6ae8be70 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/models/__init__.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/rhoas_kafka_mgmt_sdk/models/__init__.py @@ -19,8 +19,8 @@ from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_all_of import EnterpriseClusterAllOf from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_list import EnterpriseClusterList from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_list_all_of import EnterpriseClusterListAllOf -from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response import EnterpriseClusterRegistrationResponse -from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response_all_of import EnterpriseClusterRegistrationResponseAllOf +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters import EnterpriseClusterWithAddonParameters +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters_all_of import EnterpriseClusterWithAddonParametersAllOf from rhoas_kafka_mgmt_sdk.model.enterprise_osd_cluster_payload import EnterpriseOsdClusterPayload from rhoas_kafka_mgmt_sdk.model.error import Error from rhoas_kafka_mgmt_sdk.model.error_list import ErrorList @@ -52,7 +52,6 @@ from rhoas_kafka_mgmt_sdk.model.sso_provider_all_of import SsoProviderAllOf from rhoas_kafka_mgmt_sdk.model.supported_kafka_billing_model import SupportedKafkaBillingModel from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type import SupportedKafkaInstanceType -from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type_sizes_inner import SupportedKafkaInstanceTypeSizesInner from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_types_list import SupportedKafkaInstanceTypesList from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_types_list_all_of import SupportedKafkaInstanceTypesListAllOf from rhoas_kafka_mgmt_sdk.model.supported_kafka_size import SupportedKafkaSize diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list.py index 7b9ac59d..ff7fa065 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list.py @@ -13,8 +13,10 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.cloud_provider import CloudProvider from rhoas_kafka_mgmt_sdk.model.cloud_provider_list_all_of import CloudProviderListAllOf from rhoas_kafka_mgmt_sdk.model.list import List +globals()['CloudProvider'] = CloudProvider globals()['CloudProviderListAllOf'] = CloudProviderListAllOf globals()['List'] = List from rhoas_kafka_mgmt_sdk.model.cloud_provider_list import CloudProviderList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list_all_of.py index 41c7075b..7c6dd301 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_provider_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.cloud_provider import CloudProvider +globals()['CloudProvider'] = CloudProvider from rhoas_kafka_mgmt_sdk.model.cloud_provider_list_all_of import CloudProviderListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list.py index 266facf0..4898c430 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list.py @@ -13,8 +13,10 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.cloud_region import CloudRegion from rhoas_kafka_mgmt_sdk.model.cloud_region_list_all_of import CloudRegionListAllOf from rhoas_kafka_mgmt_sdk.model.list import List +globals()['CloudRegion'] = CloudRegion globals()['CloudRegionListAllOf'] = CloudRegionListAllOf globals()['List'] = List from rhoas_kafka_mgmt_sdk.model.cloud_region_list import CloudRegionList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list_all_of.py index 8c0d64fa..9b42a7b3 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_cloud_region_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.cloud_region import CloudRegion +globals()['CloudRegion'] = CloudRegion from rhoas_kafka_mgmt_sdk.model.cloud_region_list_all_of import CloudRegionListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list.py index 4ef1ac65..7b7c2765 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list.py @@ -13,8 +13,10 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster import EnterpriseCluster from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_list_all_of import EnterpriseClusterListAllOf from rhoas_kafka_mgmt_sdk.model.list import List +globals()['EnterpriseCluster'] = EnterpriseCluster globals()['EnterpriseClusterListAllOf'] = EnterpriseClusterListAllOf globals()['List'] = List from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_list import EnterpriseClusterList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list_all_of.py index 9ef1b6ab..e7438ecc 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster import EnterpriseCluster +globals()['EnterpriseCluster'] = EnterpriseCluster from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_list_all_of import EnterpriseClusterListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_registration_response.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_registration_response.py deleted file mode 100644 index b0072dd4..00000000 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_registration_response.py +++ /dev/null @@ -1,40 +0,0 @@ -""" - Kafka Management API - - Kafka Management API is a REST API to manage Kafka instances # noqa: E501 - - The version of the OpenAPI document: 1.14.0 - Contact: rhosak-support@redhat.com - Generated by: https://openapi-generator.tech -""" - - -import sys -import unittest - -import rhoas_kafka_mgmt_sdk -from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response_all_of import EnterpriseClusterRegistrationResponseAllOf -from rhoas_kafka_mgmt_sdk.model.object_reference import ObjectReference -globals()['EnterpriseClusterRegistrationResponseAllOf'] = EnterpriseClusterRegistrationResponseAllOf -globals()['ObjectReference'] = ObjectReference -from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response import EnterpriseClusterRegistrationResponse - - -class TestEnterpriseClusterRegistrationResponse(unittest.TestCase): - """EnterpriseClusterRegistrationResponse unit test stubs""" - - def setUp(self): - pass - - def tearDown(self): - pass - - def testEnterpriseClusterRegistrationResponse(self): - """Test EnterpriseClusterRegistrationResponse""" - # FIXME: construct object with mandatory attributes with example values - # model = EnterpriseClusterRegistrationResponse() # noqa: E501 - pass - - -if __name__ == '__main__': - unittest.main() diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_registration_response_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_registration_response_all_of.py deleted file mode 100644 index e758b900..00000000 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_registration_response_all_of.py +++ /dev/null @@ -1,36 +0,0 @@ -""" - Kafka Management API - - Kafka Management API is a REST API to manage Kafka instances # noqa: E501 - - The version of the OpenAPI document: 1.14.0 - Contact: rhosak-support@redhat.com - Generated by: https://openapi-generator.tech -""" - - -import sys -import unittest - -import rhoas_kafka_mgmt_sdk -from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_registration_response_all_of import EnterpriseClusterRegistrationResponseAllOf - - -class TestEnterpriseClusterRegistrationResponseAllOf(unittest.TestCase): - """EnterpriseClusterRegistrationResponseAllOf unit test stubs""" - - def setUp(self): - pass - - def tearDown(self): - pass - - def testEnterpriseClusterRegistrationResponseAllOf(self): - """Test EnterpriseClusterRegistrationResponseAllOf""" - # FIXME: construct object with mandatory attributes with example values - # model = EnterpriseClusterRegistrationResponseAllOf() # noqa: E501 - pass - - -if __name__ == '__main__': - unittest.main() diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_with_addon_parameters.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_with_addon_parameters.py new file mode 100644 index 00000000..532a4952 --- /dev/null +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_with_addon_parameters.py @@ -0,0 +1,42 @@ +""" + Kafka Management API + + Kafka Management API is a REST API to manage Kafka instances # noqa: E501 + + The version of the OpenAPI document: 1.14.0 + Contact: rhosak-support@redhat.com + Generated by: https://openapi-generator.tech +""" + + +import sys +import unittest + +import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters_all_of import EnterpriseClusterWithAddonParametersAllOf +from rhoas_kafka_mgmt_sdk.model.fleetshard_parameter import FleetshardParameter +from rhoas_kafka_mgmt_sdk.model.object_reference import ObjectReference +globals()['EnterpriseClusterWithAddonParametersAllOf'] = EnterpriseClusterWithAddonParametersAllOf +globals()['FleetshardParameter'] = FleetshardParameter +globals()['ObjectReference'] = ObjectReference +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters import EnterpriseClusterWithAddonParameters + + +class TestEnterpriseClusterWithAddonParameters(unittest.TestCase): + """EnterpriseClusterWithAddonParameters unit test stubs""" + + def setUp(self): + pass + + def tearDown(self): + pass + + def testEnterpriseClusterWithAddonParameters(self): + """Test EnterpriseClusterWithAddonParameters""" + # FIXME: construct object with mandatory attributes with example values + # model = EnterpriseClusterWithAddonParameters() # noqa: E501 + pass + + +if __name__ == '__main__': + unittest.main() diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_with_addon_parameters_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_with_addon_parameters_all_of.py new file mode 100644 index 00000000..240ff5d3 --- /dev/null +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_cluster_with_addon_parameters_all_of.py @@ -0,0 +1,38 @@ +""" + Kafka Management API + + Kafka Management API is a REST API to manage Kafka instances # noqa: E501 + + The version of the OpenAPI document: 1.14.0 + Contact: rhosak-support@redhat.com + Generated by: https://openapi-generator.tech +""" + + +import sys +import unittest + +import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.fleetshard_parameter import FleetshardParameter +globals()['FleetshardParameter'] = FleetshardParameter +from rhoas_kafka_mgmt_sdk.model.enterprise_cluster_with_addon_parameters_all_of import EnterpriseClusterWithAddonParametersAllOf + + +class TestEnterpriseClusterWithAddonParametersAllOf(unittest.TestCase): + """EnterpriseClusterWithAddonParametersAllOf unit test stubs""" + + def setUp(self): + pass + + def tearDown(self): + pass + + def testEnterpriseClusterWithAddonParametersAllOf(self): + """Test EnterpriseClusterWithAddonParametersAllOf""" + # FIXME: construct object with mandatory attributes with example values + # model = EnterpriseClusterWithAddonParametersAllOf() # noqa: E501 + pass + + +if __name__ == '__main__': + unittest.main() diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_dataplane_clusters_api.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_dataplane_clusters_api.py index 20287124..8e733a92 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_dataplane_clusters_api.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_enterprise_dataplane_clusters_api.py @@ -30,6 +30,18 @@ def test_delete_enterprise_cluster_by_id(self): """ pass + def test_get_enterprise_cluster_by_id(self): + """Test case for get_enterprise_cluster_by_id + + """ + pass + + def test_get_enterprise_cluster_with_addon_parameters(self): + """Test case for get_enterprise_cluster_with_addon_parameters + + """ + pass + def test_get_enterprise_osd_clusters(self): """Test case for get_enterprise_osd_clusters diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list.py index dff9f224..c8a5df15 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list.py @@ -13,8 +13,10 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.kafka_request import KafkaRequest from rhoas_kafka_mgmt_sdk.model.kafka_request_list_all_of import KafkaRequestListAllOf from rhoas_kafka_mgmt_sdk.model.list import List +globals()['KafkaRequest'] = KafkaRequest globals()['KafkaRequestListAllOf'] = KafkaRequestListAllOf globals()['List'] = List from rhoas_kafka_mgmt_sdk.model.kafka_request_list import KafkaRequestList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list_all_of.py index 87fbad39..6185c464 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_kafka_request_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.kafka_request import KafkaRequest +globals()['KafkaRequest'] = KafkaRequest from rhoas_kafka_mgmt_sdk.model.kafka_request_list_all_of import KafkaRequestListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list.py index de1804dd..40f095aa 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list.py @@ -13,7 +13,9 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.instant_query import InstantQuery from rhoas_kafka_mgmt_sdk.model.metrics_instant_query_list_all_of import MetricsInstantQueryListAllOf +globals()['InstantQuery'] = InstantQuery globals()['MetricsInstantQueryListAllOf'] = MetricsInstantQueryListAllOf from rhoas_kafka_mgmt_sdk.model.metrics_instant_query_list import MetricsInstantQueryList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list_all_of.py index 4637c11f..feec3a4e 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_instant_query_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.instant_query import InstantQuery +globals()['InstantQuery'] = InstantQuery from rhoas_kafka_mgmt_sdk.model.metrics_instant_query_list_all_of import MetricsInstantQueryListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list.py index 9f2e10f9..8f6d1097 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list.py @@ -14,7 +14,9 @@ import rhoas_kafka_mgmt_sdk from rhoas_kafka_mgmt_sdk.model.metrics_range_query_list_all_of import MetricsRangeQueryListAllOf +from rhoas_kafka_mgmt_sdk.model.range_query import RangeQuery globals()['MetricsRangeQueryListAllOf'] = MetricsRangeQueryListAllOf +globals()['RangeQuery'] = RangeQuery from rhoas_kafka_mgmt_sdk.model.metrics_range_query_list import MetricsRangeQueryList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list_all_of.py index 18d80bdf..203b4e12 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_metrics_range_query_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.range_query import RangeQuery +globals()['RangeQuery'] = RangeQuery from rhoas_kafka_mgmt_sdk.model.metrics_range_query_list_all_of import MetricsRangeQueryListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list.py index 34507862..47c544bf 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list.py @@ -14,7 +14,9 @@ import rhoas_kafka_mgmt_sdk from rhoas_kafka_mgmt_sdk.model.service_account_list_all_of import ServiceAccountListAllOf +from rhoas_kafka_mgmt_sdk.model.service_account_list_item import ServiceAccountListItem globals()['ServiceAccountListAllOf'] = ServiceAccountListAllOf +globals()['ServiceAccountListItem'] = ServiceAccountListItem from rhoas_kafka_mgmt_sdk.model.service_account_list import ServiceAccountList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list_all_of.py index f864083e..db564501 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_service_account_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.service_account_list_item import ServiceAccountListItem +globals()['ServiceAccountListItem'] = ServiceAccountListItem from rhoas_kafka_mgmt_sdk.model.service_account_list_all_of import ServiceAccountListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_type.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_type.py index 0895aef6..6c2c514a 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_type.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_type.py @@ -14,9 +14,9 @@ import rhoas_kafka_mgmt_sdk from rhoas_kafka_mgmt_sdk.model.supported_kafka_billing_model import SupportedKafkaBillingModel -from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type_sizes_inner import SupportedKafkaInstanceTypeSizesInner +from rhoas_kafka_mgmt_sdk.model.supported_kafka_size import SupportedKafkaSize globals()['SupportedKafkaBillingModel'] = SupportedKafkaBillingModel -globals()['SupportedKafkaInstanceTypeSizesInner'] = SupportedKafkaInstanceTypeSizesInner +globals()['SupportedKafkaSize'] = SupportedKafkaSize from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type import SupportedKafkaInstanceType diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_type_sizes_inner.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_type_sizes_inner.py deleted file mode 100644 index e02b8c0c..00000000 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_type_sizes_inner.py +++ /dev/null @@ -1,40 +0,0 @@ -""" - Kafka Management API - - Kafka Management API is a REST API to manage Kafka instances # noqa: E501 - - The version of the OpenAPI document: 1.14.0 - Contact: rhosak-support@redhat.com - Generated by: https://openapi-generator.tech -""" - - -import sys -import unittest - -import rhoas_kafka_mgmt_sdk -from rhoas_kafka_mgmt_sdk.model.supported_kafka_size import SupportedKafkaSize -from rhoas_kafka_mgmt_sdk.model.supported_kafka_size_bytes_value_item import SupportedKafkaSizeBytesValueItem -globals()['SupportedKafkaSize'] = SupportedKafkaSize -globals()['SupportedKafkaSizeBytesValueItem'] = SupportedKafkaSizeBytesValueItem -from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type_sizes_inner import SupportedKafkaInstanceTypeSizesInner - - -class TestSupportedKafkaInstanceTypeSizesInner(unittest.TestCase): - """SupportedKafkaInstanceTypeSizesInner unit test stubs""" - - def setUp(self): - pass - - def tearDown(self): - pass - - def testSupportedKafkaInstanceTypeSizesInner(self): - """Test SupportedKafkaInstanceTypeSizesInner""" - # FIXME: construct object with mandatory attributes with example values - # model = SupportedKafkaInstanceTypeSizesInner() # noqa: E501 - pass - - -if __name__ == '__main__': - unittest.main() diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list.py index 0f20302b..99fdd7cc 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list.py @@ -13,7 +13,9 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type import SupportedKafkaInstanceType from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_types_list_all_of import SupportedKafkaInstanceTypesListAllOf +globals()['SupportedKafkaInstanceType'] = SupportedKafkaInstanceType globals()['SupportedKafkaInstanceTypesListAllOf'] = SupportedKafkaInstanceTypesListAllOf from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_types_list import SupportedKafkaInstanceTypesList diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list_all_of.py index 13f7e118..4fd43b10 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_supported_kafka_instance_types_list_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_type import SupportedKafkaInstanceType +globals()['SupportedKafkaInstanceType'] = SupportedKafkaInstanceType from rhoas_kafka_mgmt_sdk.model.supported_kafka_instance_types_list_all_of import SupportedKafkaInstanceTypesListAllOf diff --git a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_version_metadata_all_of.py b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_version_metadata_all_of.py index ec5d5278..203d42e2 100644 --- a/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_version_metadata_all_of.py +++ b/app-services-sdk-python/sdks/kafka_mgmt_sdk/test/test_version_metadata_all_of.py @@ -13,6 +13,8 @@ import unittest import rhoas_kafka_mgmt_sdk +from rhoas_kafka_mgmt_sdk.model.object_reference import ObjectReference +globals()['ObjectReference'] = ObjectReference from rhoas_kafka_mgmt_sdk.model.version_metadata_all_of import VersionMetadataAllOf