diff --git a/providers/src/databricks_account/v00.00.00000/provider.yaml b/providers/src/databricks_account/v00.00.00000/provider.yaml index 3ee68a22..e703b8af 100644 --- a/providers/src/databricks_account/v00.00.00000/provider.yaml +++ b/providers/src/databricks_account/v00.00.00000/provider.yaml @@ -1,77 +1,103 @@ -id: databricks_account -name: databricks_account -version: v00.00.00000 -providerServices: - billing: - id: billing:v00.00.00000 - name: billing - preferred: true - service: - $ref: databricks_account/v00.00.00000/services/billing.yaml - title: Account Billing - version: v00.00.00000 - description: Account Billing - iam: - id: iam:v00.00.00000 - name: iam - preferred: true - service: - $ref: databricks_account/v00.00.00000/services/iam.yaml - title: Identity and Access Management - version: v00.00.00000 - description: Identity and Access Management - logging: - id: logging:v00.00.00000 - name: logging - preferred: true - service: - $ref: databricks_account/v00.00.00000/services/logging.yaml - title: Log Delivery - version: v00.00.00000 - description: Log Delivery - oauth: - id: oauth:v00.00.00000 - name: oauth - preferred: true - service: - $ref: databricks_account/v00.00.00000/services/oauth.yaml - title: OAuth Integrations - version: v00.00.00000 - description: OAuth Integrations - provisioning: - id: provisioning:v00.00.00000 - name: provisioning - preferred: true - service: - $ref: databricks_account/v00.00.00000/services/provisioning.yaml - title: Account Provisioning - version: v00.00.00000 - description: Account Provisioning - settings: - id: settings:v00.00.00000 - name: settings - preferred: true - service: - $ref: databricks_account/v00.00.00000/services/settings.yaml - title: Account Settings - version: v00.00.00000 - description: Account Settings - unity_catalog: - id: unity_catalog:v00.00.00000 - name: unity_catalog - preferred: true - service: - $ref: databricks_account/v00.00.00000/services/unity_catalog.yaml - title: Unity Catalog - version: v00.00.00000 - description: Unity Catalog -config: - auth: - type: oauth2 - client_id_env_var: DATABRICKS_CLIENT_ID - client_secret_env_var: DATABRICKS_CLIENT_SECRET - grant_type: client_credentials - token_url: https://accounts.cloud.databricks.com/oidc/accounts/{{ .__env__DATABRICKS_ACCOUNT_ID - }}/v1/token - scopes: - - all-apis +id: databricks_account +name: databricks_account +version: v00.00.00000 +providerServices: + billing: + id: billing:v00.00.00000 + name: billing + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/billing.yaml + title: Databricks Billing API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks billing service (account-level + APIs), generated from the Databricks Python SDK. + catalog: + id: catalog:v00.00.00000 + name: catalog + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/catalog.yaml + title: Databricks Catalog API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks catalog service (account-level + APIs), generated from the Databricks Python SDK. + iam: + id: iam:v00.00.00000 + name: iam + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/iam.yaml + title: Databricks Iam API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks iam service (account-level APIs), + generated from the Databricks Python SDK. + iamv2: + id: iamv2:v00.00.00000 + name: iamv2 + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/iamv2.yaml + title: Databricks Iamv2 API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks iamv2 service (account-level + APIs), generated from the Databricks Python SDK. + oauth2: + id: oauth2:v00.00.00000 + name: oauth2 + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/oauth2.yaml + title: Databricks Oauth2 API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks oauth2 service (account-level + APIs), generated from the Databricks Python SDK. + provisioning: + id: provisioning:v00.00.00000 + name: provisioning + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/provisioning.yaml + title: Databricks Provisioning API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks provisioning service + (account-level APIs), generated from the Databricks Python SDK. + settings: + id: settings:v00.00.00000 + name: settings + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/settings.yaml + title: Databricks Settings API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks settings service (account-level + APIs), generated from the Databricks Python SDK. + settingsv2: + id: settingsv2:v00.00.00000 + name: settingsv2 + preferred: true + service: + $ref: databricks_account/v00.00.00000/services/settingsv2.yaml + title: Databricks Settingsv2 API (account) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks settingsv2 service (account-level + APIs), generated from the Databricks Python SDK. +config: + auth: + type: oauth2 + client_id_env_var: DATABRICKS_CLIENT_ID + client_secret_env_var: DATABRICKS_CLIENT_SECRET + grant_type: client_credentials + token_url: >- + https://accounts.cloud.databricks.com/oidc/accounts/{{ + .__env__DATABRICKS_ACCOUNT_ID }}/v1/token + scopes: + - all-apis diff --git a/providers/src/databricks_account/v00.00.00000/services/billing.yaml b/providers/src/databricks_account/v00.00.00000/services/billing.yaml index e10286e9..33c0a1ac 100644 --- a/providers/src/databricks_account/v00.00.00000/services/billing.yaml +++ b/providers/src/databricks_account/v00.00.00000/services/billing.yaml @@ -1,758 +1,1654 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Billing API - description: Account Billing -servers: -- url: https://accounts.cloud.databricks.com -paths: - /api/2.0/accounts/{account_id}/usage/download: - get: - operationId: billable-usage-download - externalDocs: - url: https://docs.databricks.com/api/account/billableusage/download - x-stackQL-resource: billable_usage - x-stackQL-method: download - x-stackQL-verb: exec - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: start_month - required: true - description: string - in: query - - name: end_month - required: true - description: string - in: query - - name: personal_data - in: query - responses: - '200': - description: Billable usage data was returned successfully. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Returns billable usage logs in CSV format for the specified account - and date range. For the data schema, see - /api/2.1/accounts/{account_id}/budgets: - post: - operationId: budgets-create - externalDocs: - url: https://docs.databricks.com/api/account/budgets/create - x-stackQL-resource: budgets - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - budget: - type: object - properties: - budget_configuration_id: - type: string - account_id: - type: string - create_time: - type: integer - update_time: - type: integer - alert_configurations: - type: array - items: - type: object - properties: - alert_configuration_id: - type: string - time_period: - type: string - trigger_type: - type: string - quantity_type: - type: string - quantity_threshold: - type: string - action_configurations: - type: array - items: - type: object - properties: - action_configuration_id: - type: string - action_type: - type: string - target: - type: string - filter: - type: object - properties: - workspace_id: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: string - display_name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - budget: - type: required - example: - budget: - account_id: 449e7a5c-69d3-4b8a-aaaf-5c9b713ebc65 - alert_configurations: - - time_period: MONTH - trigger_type: CUMULATIVE_SPENDING_EXCEEDED - quantity_type: LIST_PRICE_DOLLARS_USD - quantity_threshold: string - action_configurations: - - action_type: EMAIL_NOTIFICATION - target: string - filter: - workspace_id: - operator: IN - values: - - 0 - tags: - - key: string - value: - operator: IN - values: - - string - display_name: string - description: Create a new budget configuration for an account. For full details, - see - get: - operationId: budgets-list - externalDocs: - url: https://docs.databricks.com/api/account/budgets/list - x-stackQL-resource: budgets - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - budgets: - type: array - items: - type: object - properties: - budget_configuration_id: - type: string - account_id: - type: string - create_time: - type: integer - update_time: - type: integer - alert_configurations: - type: array - items: - type: object - properties: - alert_configuration_id: - type: string - time_period: - type: string - trigger_type: - type: string - quantity_type: - type: string - quantity_threshold: - type: string - action_configurations: - type: array - items: - type: object - properties: - action_configuration_id: - type: string - action_type: - type: string - target: - type: string - filter: - type: object - properties: - workspace_id: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: string - display_name: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.budgets - description: Gets all budgets associated with this account. - /api/2.1/accounts/{account_id}/budgets/{budget_id}: - delete: - operationId: budgets-delete - externalDocs: - url: https://docs.databricks.com/api/account/budgets/delete - x-stackQL-resource: budgets - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: budget_id - required: true - description: uuid - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a budget configuration for an account. Both account and - budget configuration are specified by ID. This cannot be undone. - get: - operationId: budgets-get - externalDocs: - url: https://docs.databricks.com/api/account/budgets/get - x-stackQL-resource: budgets - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: budget_id - required: true - description: uuid - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - budget: - type: object - properties: - budget_configuration_id: - type: string - account_id: - type: string - create_time: - type: integer - update_time: - type: integer - alert_configurations: - type: array - items: - type: object - properties: - alert_configuration_id: - type: string - time_period: - type: string - trigger_type: - type: string - quantity_type: - type: string - quantity_threshold: - type: string - action_configurations: - type: array - items: - type: object - properties: - action_configuration_id: - type: string - action_type: - type: string - target: - type: string - filter: - type: object - properties: - workspace_id: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: string - display_name: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.budget - description: Gets a budget configuration for an account. Both account and budget - configuration are specified by ID. - put: - operationId: budgets-update - externalDocs: - url: https://docs.databricks.com/api/account/budgets/update - x-stackQL-resource: budgets - x-stackQL-method: update - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: budget_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - budget: - type: object - properties: - budget_configuration_id: - type: string - account_id: - type: string - create_time: - type: integer - update_time: - type: integer - alert_configurations: - type: array - items: - type: object - properties: - alert_configuration_id: - type: string - time_period: - type: string - trigger_type: - type: string - quantity_type: - type: string - quantity_threshold: - type: string - action_configurations: - type: array - items: - type: object - properties: - action_configuration_id: - type: string - action_type: - type: string - target: - type: string - filter: - type: object - properties: - workspace_id: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: object - properties: - operator: - type: string - values: - type: array - items: - type: string - display_name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - budget: - type: required - example: - budget: - budget_configuration_id: dc0b6e4b-23ef-471d-adf9-1d38d9199f98 - account_id: 449e7a5c-69d3-4b8a-aaaf-5c9b713ebc65 - alert_configurations: - - alert_configuration_id: 60deed03-5f33-4eda-ac0b-03b20fccc5d1 - time_period: MONTH - trigger_type: CUMULATIVE_SPENDING_EXCEEDED - quantity_type: LIST_PRICE_DOLLARS_USD - quantity_threshold: string - action_configurations: - - action_configuration_id: 4079440f-2386-43ee-bb12-2fb4137f0080 - action_type: EMAIL_NOTIFICATION - target: string - filter: - workspace_id: - operator: IN - values: - - 0 - tags: - - key: string - value: - operator: IN - values: - - string - display_name: string - description: Updates a budget configuration for an account. Both account and - budget configuration are specified by ID. - /api/2.0/accounts/{account_id}/dashboard: - post: - operationId: usage-dashboards-create - externalDocs: - url: https://docs.databricks.com/api/account/usagedashboards/create - x-stackQL-resource: usage_dashboards - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - dashboard_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - workspace_id: - type: integer - format: int64 - dashboard_type: - type: string - example: - workspace_id: 0 - dashboard_type: USAGE_DASHBOARD_TYPE_WORKSPACE - description: Create a usage dashboard specified by workspaceId, accountId, and - dashboard type. - get: - operationId: usage-dashboards-get - externalDocs: - url: https://docs.databricks.com/api/account/usagedashboards/get - x-stackQL-resource: usage_dashboards - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - - name: workspace_id - in: query - - name: dashboard_type - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - dashboard_id: - type: string - dashboard_url: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - description: Get a usage dashboard specified by workspaceId, accountId, and - dashboard type. -components: - x-stackQL-resources: - billable_usage: - id: databricks_account.billing.billable_usage - name: billable_usage - title: Billable_usage - methods: - download: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1usage~1download/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/billable_usage/methods/download' - budgets: - id: databricks_account.billing.budgets - name: budgets - title: Budgets - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.budgets - delete: - operation: - $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets~1{budget_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets~1{budget_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.budget - update: - operation: - $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets~1{budget_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/budgets/methods/get' - - $ref: '#/components/x-stackQL-resources/budgets/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/budgets/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/budgets/methods/update' - delete: - - $ref: '#/components/x-stackQL-resources/budgets/methods/delete' - exec: [] - usage_dashboards: - id: databricks_account.billing.usage_dashboards - name: usage_dashboards - title: Usage_dashboards - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1dashboard/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1dashboard/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/usage_dashboards/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/usage_dashboards/methods/create' - update: [] - replace: [] - delete: [] - exec: [] +openapi: 3.0.0 +info: + title: Databricks Billing API (account) + description: OpenAPI specification for the Databricks billing service (account-level APIs), generated from the Databricks + Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.billing +servers: +- url: https://accounts.cloud.databricks.com +paths: + /api/2.0/accounts/{account_id}/usage/download: + get: + operationId: billable_usage_download + summary: Returns billable usage logs in CSV format for the specified account and date range. For the data + tags: + - billing + - billable_usage + description: "Returns billable usage logs in CSV format for the specified account and date range. For the data\nschema,\ + \ see:\n\n- AWS: [CSV file schema]. - GCP: [CSV file schema].\n\nNote that this method might take multiple minutes\ + \ to complete.\n\n**Warning**: Depending on the queried date range, the number of workspaces in the account, the size\ + \ of\nthe response and the internet speed of the caller, this API may hit a timeout after a few minutes. If\nyou experience\ + \ this, try to mitigate by calling the API with narrower date ranges.\n\n[CSV file schema]: https://docs.gcp.databricks.com/administration-guide/account-settings/usage-analysis.html#csv-file-schema\n\ + \n:param start_month: str\n Format specification for month in the format `YYYY-MM`. This is used to specify billable\ + \ usage\n `start_month` and `end_month` properties. **Note**: Billable usage logs are unavailable before March\n\ + \ 2019 (`2019-03`).\n:param end_month: str\n Format: `YYYY-MM`. Last month to return billable usage logs for. This\ + \ field is required.\n:param personal_data: bool (optional)\n Specify whether to include personally identifiable\ + \ information in the billable usage logs, for\n example the email addresses of cluster creators. Handle this information\ + \ with care. Defaults to\n false.\n\n:returns: :class:`DownloadResponse`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: start_month + in: query + required: true + schema: + type: string + description: 'Format specification for month in the format `YYYY-MM`. This is used to specify billable usage `start_month` + and `end_month` properties. **Note**: Billable usage logs are unavailable before March 2019 (`2019-03`).' + - name: end_month + in: query + required: true + schema: + type: string + description: 'Format: `YYYY-MM`. Last month to return billable usage logs for. This field is required.' + - name: personal_data + in: query + required: false + schema: + type: string + description: Specify whether to include personally identifiable information in the billable usage logs, for example + the email addresses of cluster creators. Handle this information with care. Defaults to false. + responses: + '200': + description: Success + content: + text/plain: + schema: + type: object + properties: + contents: + type: string + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BillableUsageAPI + /api/2.1/accounts/{account_id}/budget-policies: + post: + operationId: budget_policy_create + summary: Creates a new policy. + tags: + - billing + - budget_policy + description: "Creates a new policy.\n\n:param policy: :class:`BudgetPolicy` (optional)\n The policy to create. `policy_id`\ + \ needs to be empty as it will be generated `policy_name` must be\n provided, custom_tags may need to be provided\ + \ depending on the cloud provider. All other fields are\n optional.\n:param request_id: str (optional)\n A unique\ + \ identifier for this request. Restricted to 36 ASCII characters. A random UUID is\n recommended. This request is\ + \ only idempotent if a `request_id` is provided.\n\n:returns: :class:`BudgetPolicy`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + policy: + type: string + description: The policy to create. `policy_id` needs to be empty as it will be generated `policy_name` must + be provided, custom_tags may need to be provided depending on the cloud provider. All other fields are + optional. + request_id: + type: string + description: A unique identifier for this request. Restricted to 36 ASCII characters. A random UUID is recommended. + This request is only idempotent if a `request_id` is provided. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BudgetPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetPolicyAPI + get: + operationId: budget_policy_list + summary: Lists all policies. Policies are returned in the alphabetically ascending order of their names. + tags: + - billing + - budget_policy + description: "Lists all policies. Policies are returned in the alphabetically ascending order of their names.\n\n:param\ + \ filter_by: :class:`Filter` (optional)\n A filter to apply to the list of policies.\n:param page_size: int (optional)\n\ + \ The maximum number of budget policies to return. If unspecified, at most 100 budget policies will be\n returned.\ + \ The maximum value is 1000; values above 1000 will be coerced to 1000.\n:param page_token: str (optional)\n A page\ + \ token, received from a previous `ListServerlessPolicies` call. Provide this to retrieve the\n subsequent page.\ + \ If unspecified, the first page will be returned.\n\n When paginating, all other parameters provided to `ListServerlessPoliciesRequest`\ + \ must match the\n call that provided the page token.\n:param sort_spec: :class:`SortSpec` (optional)\n The sort\ + \ specification.\n\n:returns: Iterator over :class:`BudgetPolicy`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: filter_by + in: query + required: false + schema: + type: string + description: A filter to apply to the list of policies. + - name: page_size + in: query + required: false + schema: + type: string + description: The maximum number of budget policies to return. If unspecified, at most 100 budget policies will be + returned. The maximum value is 1000; values above 1000 will be coerced to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: A page token, received from a previous `ListServerlessPolicies` call. Provide this to retrieve the subsequent + page. If unspecified, the first page will be returned. When paginating, all other parameters provided to `ListServerlessPoliciesRequest` + must match the call that provided the page token. + - name: sort_spec + in: query + required: false + schema: + type: string + description: The sort specification. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListBudgetPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetPolicyAPI + /api/2.1/accounts/{account_id}/budget-policies/{policy_id}: + delete: + operationId: budget_policy_delete + summary: Deletes a policy + tags: + - billing + - budget_policy + description: "Deletes a policy\n\n:param policy_id: str\n The Id of the policy." + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: policy_id + in: path + required: true + schema: + type: string + description: The Id of the policy. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetPolicyAPI + get: + operationId: budget_policy_get + summary: Retrieves a policy by it's ID. + tags: + - billing + - budget_policy + description: "Retrieves a policy by it's ID.\n\n:param policy_id: str\n The Id of the policy.\n\n:returns: :class:`BudgetPolicy`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: policy_id + in: path + required: true + schema: + type: string + description: The Id of the policy. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BudgetPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetPolicyAPI + patch: + operationId: budget_policy_update + summary: Updates a policy + tags: + - billing + - budget_policy + description: "Updates a policy\n\n:param policy_id: str\n The Id of the policy. This field is generated by Databricks\ + \ and globally unique.\n:param policy: :class:`BudgetPolicy`\n The policy to update. `creator_user_id` cannot be\ + \ specified in the request. All other fields must be\n specified even if not changed. The `policy_id` is used to\ + \ identify the policy to update.\n:param limit_config: :class:`LimitConfig` (optional)\n DEPRECATED. This is redundant\ + \ field as LimitConfig is part of the BudgetPolicy\n\n:returns: :class:`BudgetPolicy`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: policy_id + in: path + required: true + schema: + type: string + description: The Id of the policy. This field is generated by Databricks and globally unique. + - name: limit_config + in: query + required: false + schema: + type: string + description: DEPRECATED. This is redundant field as LimitConfig is part of the BudgetPolicy + requestBody: + content: + application/json: + schema: + type: object + properties: + policy: + type: string + description: The policy to update. `creator_user_id` cannot be specified in the request. All other fields + must be specified even if not changed. The `policy_id` is used to identify the policy to update. + required: + - policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BudgetPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetPolicyAPI + /api/2.1/accounts/{account_id}/budgets: + post: + operationId: budgets_create + summary: Create a new budget configuration for an account. For full details, see + tags: + - billing + - budgets + description: "Create a new budget configuration for an account. For full details, see\nhttps://docs.databricks.com/en/admin/account-settings/budgets.html.\n\ + \n:param budget: :class:`CreateBudgetConfigurationBudget`\n Properties of the new budget configuration.\n\n:returns:\ + \ :class:`CreateBudgetConfigurationResponse`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + budget: + type: string + description: Properties of the new budget configuration. + required: + - budget + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateBudgetConfigurationResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetsAPI + get: + operationId: budgets_list + summary: Gets all budgets associated with this account. + tags: + - billing + - budgets + description: "Gets all budgets associated with this account.\n\n:param page_token: str (optional)\n A page token received\ + \ from a previous get all budget configurations call. This token can be used to\n retrieve the subsequent page. Requests\ + \ first page if absent.\n\n:returns: Iterator over :class:`BudgetConfiguration`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: A page token received from a previous get all budget configurations call. This token can be used to retrieve + the subsequent page. Requests first page if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListBudgetConfigurationsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetsAPI + /api/2.1/accounts/{account_id}/budgets/{budget_id}: + delete: + operationId: budgets_delete + summary: Deletes a budget configuration for an account. Both account and budget configuration are specified by + tags: + - billing + - budgets + description: "Deletes a budget configuration for an account. Both account and budget configuration are specified by\n\ + ID. This cannot be undone.\n\n:param budget_id: str\n The Databricks budget configuration ID." + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: budget_id + in: path + required: true + schema: + type: string + description: The Databricks budget configuration ID. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetsAPI + get: + operationId: budgets_get + summary: Gets a budget configuration for an account. Both account and budget configuration are specified by ID. + tags: + - billing + - budgets + description: "Gets a budget configuration for an account. Both account and budget configuration are specified by ID.\n\ + \n:param budget_id: str\n The budget configuration ID\n\n:returns: :class:`GetBudgetConfigurationResponse`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: budget_id + in: path + required: true + schema: + type: string + description: The budget configuration ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetBudgetConfigurationResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetsAPI + put: + operationId: budgets_update + summary: Updates a budget configuration for an account. Both account and budget configuration are specified by + tags: + - billing + - budgets + description: "Updates a budget configuration for an account. Both account and budget configuration are specified by\n\ + ID.\n\n:param budget_id: str\n The Databricks budget configuration ID.\n:param budget: :class:`UpdateBudgetConfigurationBudget`\n\ + \ The updated budget. This will overwrite the budget specified by the budget ID.\n\n:returns: :class:`UpdateBudgetConfigurationResponse`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: budget_id + in: path + required: true + schema: + type: string + description: The Databricks budget configuration ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + budget: + type: string + description: The updated budget. This will overwrite the budget specified by the budget ID. + required: + - budget + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateBudgetConfigurationResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: BudgetsAPI + /api/2.0/accounts/{account_id}/log-delivery: + post: + operationId: log_delivery_create + summary: Creates a new Databricks log delivery configuration to enable delivery of the specified type of logs + tags: + - billing + - log_delivery + description: 'Creates a new Databricks log delivery configuration to enable delivery of the specified type of logs + + to your storage location. This requires that you already created a [credential + + object](:method:Credentials/Create) (which encapsulates a cross-account service IAM role) and a + + [storage configuration object](:method:Storage/Create) (which encapsulates an S3 bucket). + + + For full details, including the required IAM role policies and bucket policies, see [Deliver and + + access billable usage logs] or [Configure audit logging]. + + + **Note**: There is a limit on the number of log delivery configurations available per account (each + + limit applies separately to each log type including billable usage and audit logs). You can create a + + maximum of two enabled account-level delivery configurations (configurations without a workspace + + filter) per type. Additionally, you can create two enabled workspace-level delivery configurations per + + workspace for each log type, which means that the same workspace ID can occur in the workspace filter + + for no more than two delivery configurations per log type. + + + You cannot delete a log delivery configuration, but you can disable it (see [Enable or disable log + + delivery configuration](:method:LogDelivery/PatchStatus)). + + + [Configure audit logging]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html + + [Deliver and access billable usage logs]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html + + + :param log_delivery_configuration: :class:`CreateLogDeliveryConfigurationParams` + + + :returns: :class:`WrappedLogDeliveryConfiguration`' + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + log_delivery_configuration: + type: string + description: ':returns: :class:`WrappedLogDeliveryConfiguration`' + required: + - log_delivery_configuration + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WrappedLogDeliveryConfiguration' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LogDeliveryAPI + get: + operationId: log_delivery_list + summary: Gets all Databricks log delivery configurations associated with an account specified by ID. + tags: + - billing + - log_delivery + description: "Gets all Databricks log delivery configurations associated with an account specified by ID.\n\n:param\ + \ credentials_id: str (optional)\n The Credentials id to filter the search results with\n:param page_token: str (optional)\n\ + \ A page token received from a previous get all budget configurations call. This token can be used to\n retrieve\ + \ the subsequent page. Requests first page if absent.\n:param status: :class:`LogDeliveryConfigStatus` (optional)\n\ + \ The log delivery status to filter the search results with\n:param storage_configuration_id: str (optional)\n The\ + \ Storage Configuration id to filter the search results with\n\n:returns: Iterator over :class:`LogDeliveryConfiguration`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: credentials_id + in: query + required: false + schema: + type: string + description: The Credentials id to filter the search results with + - name: page_token + in: query + required: false + schema: + type: string + description: A page token received from a previous get all budget configurations call. This token can be used to retrieve + the subsequent page. Requests first page if absent. + - name: status + in: query + required: false + schema: + type: string + description: The log delivery status to filter the search results with + - name: storage_configuration_id + in: query + required: false + schema: + type: string + description: The Storage Configuration id to filter the search results with + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WrappedLogDeliveryConfigurations' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LogDeliveryAPI + /api/2.0/accounts/{account_id}/log-delivery/{log_delivery_configuration_id}: + get: + operationId: log_delivery_get + summary: Gets a Databricks log delivery configuration object for an account, both specified by ID. + tags: + - billing + - log_delivery + description: "Gets a Databricks log delivery configuration object for an account, both specified by ID.\n\n:param log_delivery_configuration_id:\ + \ str\n The log delivery configuration id of customer\n\n:returns: :class:`GetLogDeliveryConfigurationResponse`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: log_delivery_configuration_id + in: path + required: true + schema: + type: string + description: The log delivery configuration id of customer + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetLogDeliveryConfigurationResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LogDeliveryAPI + patch: + operationId: log_delivery_patch_status + summary: Enables or disables a log delivery configuration. Deletion of delivery configurations is not + tags: + - billing + - log_delivery + description: "Enables or disables a log delivery configuration. Deletion of delivery configurations is not\nsupported,\ + \ so disable log delivery configurations that are no longer needed. Note that you can't\nre-enable a delivery configuration\ + \ if this would violate the delivery configuration limits described\nunder [Create log delivery](:method:LogDelivery/Create).\n\ + \n:param log_delivery_configuration_id: str\n The log delivery configuration id of customer\n:param status: :class:`LogDeliveryConfigStatus`\n\ + \ Status of log delivery configuration. Set to `ENABLED` (enabled) or `DISABLED` (disabled). Defaults\n to `ENABLED`.\ + \ You can [enable or disable the\n configuration](#operation/patch-log-delivery-config-status) later. Deletion of\ + \ a configuration is\n not supported, so disable a log delivery configuration that is no longer needed." + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: log_delivery_configuration_id + in: path + required: true + schema: + type: string + description: The log delivery configuration id of customer + requestBody: + content: + application/json: + schema: + type: object + properties: + status: + type: string + description: Status of log delivery configuration. Set to `ENABLED` (enabled) or `DISABLED` (disabled). + Defaults to `ENABLED`. You can [enable or disable the configuration](#operation/patch-log-delivery-config-status) + later. Deletion of a configuration is not supported, so disable a log delivery configuration that is no + longer needed. + required: + - status + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LogDeliveryAPI + /api/2.0/accounts/{account_id}/dashboard: + post: + operationId: usage_dashboards_create + summary: Create a usage dashboard specified by workspaceId, accountId, and dashboard type. + tags: + - billing + - usage_dashboards + description: "Create a usage dashboard specified by workspaceId, accountId, and dashboard type.\n\n:param dashboard_type:\ + \ :class:`UsageDashboardType` (optional)\n Workspace level usage dashboard shows usage data for the specified workspace\ + \ ID. Global level usage\n dashboard shows usage data for all workspaces in the account.\n:param major_version: :class:`UsageDashboardMajorVersion`\ + \ (optional)\n The major version of the usage dashboard template to use. Defaults to VERSION_1.\n:param workspace_id:\ + \ int (optional)\n The workspace ID of the workspace in which the usage dashboard is created.\n\n:returns: :class:`CreateBillingUsageDashboardResponse`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + dashboard_type: + type: string + description: Workspace level usage dashboard shows usage data for the specified workspace ID. Global level + usage dashboard shows usage data for all workspaces in the account. + major_version: + type: string + description: The major version of the usage dashboard template to use. Defaults to VERSION_1. + workspace_id: + type: string + description: The workspace ID of the workspace in which the usage dashboard is created. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateBillingUsageDashboardResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsageDashboardsAPI + get: + operationId: usage_dashboards_get + summary: Get a usage dashboard specified by workspaceId, accountId, and dashboard type. + tags: + - billing + - usage_dashboards + description: "Get a usage dashboard specified by workspaceId, accountId, and dashboard type.\n\n:param dashboard_type:\ + \ :class:`UsageDashboardType` (optional)\n Workspace level usage dashboard shows usage data for the specified workspace\ + \ ID. Global level usage\n dashboard shows usage data for all workspaces in the account.\n:param workspace_id: int\ + \ (optional)\n The workspace ID of the workspace in which the usage dashboard is created.\n\n:returns: :class:`GetBillingUsageDashboardResponse`" + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: dashboard_type + in: query + required: false + schema: + type: string + description: Workspace level usage dashboard shows usage data for the specified workspace ID. Global level usage dashboard + shows usage data for all workspaces in the account. + - name: workspace_id + in: query + required: false + schema: + type: string + description: The workspace ID of the workspace in which the usage dashboard is created. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetBillingUsageDashboardResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsageDashboardsAPI +components: + schemas: + ActionConfiguration: + type: object + properties: + action_configuration_id: + type: string + action_type: + $ref: '#/components/schemas/ActionConfigurationType' + description: The type of the action. + target: + type: string + description: Target for the action. For example, an email address. + AlertConfiguration: + type: object + properties: + action_configurations: + type: array + items: + $ref: '#/components/schemas/ActionConfiguration' + alert_configuration_id: + type: string + description: Databricks alert configuration ID. + quantity_threshold: + type: string + description: The threshold for the budget alert to determine if it is in a triggered state. The number is evaluated + based on `quantity_type`. + quantity_type: + $ref: '#/components/schemas/AlertConfigurationQuantityType' + description: The way to calculate cost for this budget alert. This is what `quantity_threshold` is measured in. + time_period: + $ref: '#/components/schemas/AlertConfigurationTimePeriod' + description: The time window of usage data for the budget. + trigger_type: + $ref: '#/components/schemas/AlertConfigurationTriggerType' + description: The evaluation method to determine when this budget alert is in a triggered state. + BudgetConfiguration: + type: object + properties: + account_id: + type: string + alert_configurations: + type: array + items: + $ref: '#/components/schemas/AlertConfiguration' + description: Alerts to configure when this budget is in a triggered state. Budgets must have exactly one alert configuration. + budget_configuration_id: + type: string + description: Databricks budget configuration ID. + create_time: + type: integer + description: Creation time of this budget configuration. + display_name: + type: string + description: 'Human-readable name of budget configuration. Max Length: 128' + filter: + $ref: '#/components/schemas/BudgetConfigurationFilter' + description: Configured filters for this budget. These are applied to your account's usage to limit the scope of + what is considered for this budget. Leave empty to include all usage for this account. All provided filters must + be matched for usage to be included. + update_time: + type: integer + description: Update time of this budget configuration. + BudgetConfigurationFilter: + type: object + properties: + tags: + type: array + items: + $ref: '#/components/schemas/BudgetConfigurationFilterTagClause' + workspace_id: + $ref: '#/components/schemas/BudgetConfigurationFilterWorkspaceIdClause' + description: If provided, usage must match with the provided Databricks workspace IDs. + BudgetConfigurationFilterClause: + type: object + properties: + operator: + $ref: '#/components/schemas/BudgetConfigurationFilterOperator' + values: + type: array + items: + type: string + BudgetConfigurationFilterTagClause: + type: object + properties: + key: + type: string + value: + $ref: '#/components/schemas/BudgetConfigurationFilterClause' + BudgetConfigurationFilterWorkspaceIdClause: + type: object + properties: + operator: + $ref: '#/components/schemas/BudgetConfigurationFilterOperator' + values: + type: array + items: + type: integer + BudgetPolicy: + type: object + properties: + binding_workspace_ids: + type: array + items: + type: integer + description: List of workspaces that this budget policy will be exclusively bound to. An empty binding implies that + this budget policy is open to any workspace in the account. + custom_tags: + type: string + description: A list of tags defined by the customer. At most 20 entries are allowed per policy. + policy_id: + type: string + description: The Id of the policy. This field is generated by Databricks and globally unique. + policy_name: + type: string + description: The name of the policy. - Must be unique among active policies. - Can contain only characters from + the ISO 8859-1 (latin1) set. - Can't start with reserved keywords such as `databricks:default-policy`. + description: Contains the BudgetPolicy details. + CreateBillingUsageDashboardResponse: + type: object + properties: + dashboard_id: + type: string + CreateBudgetConfigurationBudget: + type: object + properties: + account_id: + type: string + alert_configurations: + type: array + items: + $ref: '#/components/schemas/CreateBudgetConfigurationBudgetAlertConfigurations' + description: Alerts to configure when this budget is in a triggered state. Budgets must have exactly one alert configuration. + display_name: + type: string + description: 'Human-readable name of budget configuration. Max Length: 128' + filter: + $ref: '#/components/schemas/BudgetConfigurationFilter' + description: Configured filters for this budget. These are applied to your account's usage to limit the scope of + what is considered for this budget. Leave empty to include all usage for this account. All provided filters must + be matched for usage to be included. + CreateBudgetConfigurationBudgetActionConfigurations: + type: object + properties: + action_type: + $ref: '#/components/schemas/ActionConfigurationType' + target: + type: string + description: Target for the action. For example, an email address. + CreateBudgetConfigurationBudgetAlertConfigurations: + type: object + properties: + action_configurations: + type: array + items: + $ref: '#/components/schemas/CreateBudgetConfigurationBudgetActionConfigurations' + quantity_threshold: + type: string + description: The threshold for the budget alert to determine if it is in a triggered state. The number is evaluated + based on `quantity_type`. + quantity_type: + $ref: '#/components/schemas/AlertConfigurationQuantityType' + description: The way to calculate cost for this budget alert. This is what `quantity_threshold` is measured in. + time_period: + $ref: '#/components/schemas/AlertConfigurationTimePeriod' + description: The time window of usage data for the budget. + trigger_type: + $ref: '#/components/schemas/AlertConfigurationTriggerType' + description: The evaluation method to determine when this budget alert is in a triggered state. + CreateBudgetConfigurationResponse: + type: object + properties: + budget: + $ref: '#/components/schemas/BudgetConfiguration' + CreateLogDeliveryConfigurationParams: + type: object + properties: + log_type: + $ref: '#/components/schemas/LogType' + description: 'Log delivery type. Supported values are: * `BILLABLE_USAGE` — Configure [billable usage log delivery]. + For the CSV schema, see the [View billable usage]. * `AUDIT_LOGS` — Configure [audit log delivery]. For the JSON + schema, see [Configure audit logging] [Configure audit logging]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html + [View billable usage]: https://docs.databricks.com/administration-guide/account-settings/usage.html [audit log + delivery]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html [billable usage log + delivery]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html' + output_format: + $ref: '#/components/schemas/OutputFormat' + description: 'The file type of log delivery. * If `log_type` is `BILLABLE_USAGE`, this value must be `CSV`. Only + the CSV (comma-separated values) format is supported. For the schema, see the [View billable usage] * If `log_type` + is `AUDIT_LOGS`, this value must be `JSON`. Only the JSON (JavaScript Object Notation) format is supported. For + the schema, see the [Configuring audit logs]. [Configuring audit logs]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html + [View billable usage]: https://docs.databricks.com/administration-guide/account-settings/usage.html' + credentials_id: + type: string + description: 'The ID for a method:credentials/create that represents the AWS IAM role with policy and trust relationship + as described in the main billable usage documentation page. See [Configure billable usage delivery]. [Configure + billable usage delivery]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html' + storage_configuration_id: + type: string + description: 'The ID for a method:storage/create that represents the S3 bucket with bucket policy as described in + the main billable usage documentation page. See [Configure billable usage delivery]. [Configure billable usage + delivery]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html' + config_name: + type: string + description: The optional human-readable name of the log delivery configuration. Defaults to empty. + delivery_path_prefix: + type: string + description: The optional delivery path prefix within Amazon S3 storage. Defaults to empty, which means that logs + are delivered to the root of the bucket. This must be a valid S3 object key. This must not start or end with a + slash character. + delivery_start_time: + type: string + description: This field applies only if log_type is BILLABLE_USAGE. This is the optional start month and year for + delivery, specified in YYYY-MM format. Defaults to current year and month. BILLABLE_USAGE logs are not available + for usage before March 2019 (2019-03). + status: + $ref: '#/components/schemas/LogDeliveryConfigStatus' + description: Status of log delivery configuration. Set to `ENABLED` (enabled) or `DISABLED` (disabled). Defaults + to `ENABLED`. You can [enable or disable the configuration](#operation/patch-log-delivery-config-status) later. + Deletion of a configuration is not supported, so disable a log delivery configuration that is no longer needed. + workspace_ids_filter: + type: array + items: + type: integer + description: Optional filter that specifies workspace IDs to deliver logs for. By default the workspace filter is + empty and log delivery applies at the account level, delivering workspace-level logs for all workspaces in your + account, plus account level logs. You can optionally set this field to an array of workspace IDs (each one is + an `int64`) to which log delivery should apply, in which case only workspace-level logs relating to the specified + workspaces are delivered. If you plan to use different log delivery configurations for different workspaces, set + this field explicitly. Be aware that delivery configurations mentioning specific workspaces won't apply to new + workspaces created in the future, and delivery won't include account level logs. For some types of Databricks + deployments there is only one workspace per account ID, so this field is unnecessary. + required: + - log_type + - output_format + - credentials_id + - storage_configuration_id + description: '* Log Delivery Configuration' + DeleteBudgetConfigurationResponse: + type: object + properties: {} + DownloadResponse: + type: object + properties: + contents: + type: string + format: binary + Filter: + type: object + properties: + creator_user_id: + type: integer + description: The policy creator user id to be filtered on. If unspecified, all policies will be returned. + creator_user_name: + type: string + description: The policy creator user name to be filtered on. If unspecified, all policies will be returned. + policy_name: + type: string + description: The partial name of policies to be filtered on. If unspecified, all policies will be returned. + description: "Structured representation of a filter to be applied to a list of policies. All specified filters\n \ + \ will be applied in conjunction." + GetBillingUsageDashboardResponse: + type: object + properties: + dashboard_id: + type: string + dashboard_url: + type: string + description: The URL of the usage dashboard. + GetBudgetConfigurationResponse: + type: object + properties: + budget: + $ref: '#/components/schemas/BudgetConfiguration' + GetLogDeliveryConfigurationResponse: + type: object + properties: + log_delivery_configuration: + $ref: '#/components/schemas/LogDeliveryConfiguration' + LimitConfig: + type: object + properties: {} + description: "The limit configuration of the policy. Limit configuration provide a budget policy level cost\n control\ + \ by enforcing the limit." + ListBudgetConfigurationsResponse: + type: object + properties: + budgets: + type: array + items: + $ref: '#/components/schemas/BudgetConfiguration' + next_page_token: + type: string + description: Token which can be sent as `page_token` to retrieve the next page of results. If this field is omitted, + there are no subsequent budgets. + ListBudgetPoliciesResponse: + type: object + properties: + next_page_token: + type: string + description: A token that can be sent as `page_token` to retrieve the next page. If this field is omitted, there + are no subsequent pages. + policies: + type: array + items: + $ref: '#/components/schemas/BudgetPolicy' + previous_page_token: + type: string + description: A token that can be sent as `page_token` to retrieve the previous page. In this field is omitted, there + are no previous pages. + description: A list of policies. + LogDeliveryConfiguration: + type: object + properties: + log_type: + $ref: '#/components/schemas/LogType' + description: 'Log delivery type. Supported values are: * `BILLABLE_USAGE` — Configure [billable usage log delivery]. + For the CSV schema, see the [View billable usage]. * `AUDIT_LOGS` — Configure [audit log delivery]. For the JSON + schema, see [Configure audit logging] [Configure audit logging]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html + [View billable usage]: https://docs.databricks.com/administration-guide/account-settings/usage.html [audit log + delivery]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html [billable usage log + delivery]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html' + output_format: + $ref: '#/components/schemas/OutputFormat' + description: 'The file type of log delivery. * If `log_type` is `BILLABLE_USAGE`, this value must be `CSV`. Only + the CSV (comma-separated values) format is supported. For the schema, see the [View billable usage] * If `log_type` + is `AUDIT_LOGS`, this value must be `JSON`. Only the JSON (JavaScript Object Notation) format is supported. For + the schema, see the [Configuring audit logs]. [Configuring audit logs]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html + [View billable usage]: https://docs.databricks.com/administration-guide/account-settings/usage.html' + credentials_id: + type: string + description: 'The ID for a method:credentials/create that represents the AWS IAM role with policy and trust relationship + as described in the main billable usage documentation page. See [Configure billable usage delivery]. [Configure + billable usage delivery]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html' + storage_configuration_id: + type: string + description: 'The ID for a method:storage/create that represents the S3 bucket with bucket policy as described in + the main billable usage documentation page. See [Configure billable usage delivery]. [Configure billable usage + delivery]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html' + account_id: + type: string + description: Databricks account ID. + config_id: + type: string + description: The unique UUID of log delivery configuration + config_name: + type: string + description: The optional human-readable name of the log delivery configuration. Defaults to empty. + creation_time: + type: integer + description: Time in epoch milliseconds when the log delivery configuration was created. + delivery_path_prefix: + type: string + description: The optional delivery path prefix within Amazon S3 storage. Defaults to empty, which means that logs + are delivered to the root of the bucket. This must be a valid S3 object key. This must not start or end with a + slash character. + delivery_start_time: + type: string + description: This field applies only if log_type is BILLABLE_USAGE. This is the optional start month and year for + delivery, specified in YYYY-MM format. Defaults to current year and month. BILLABLE_USAGE logs are not available + for usage before March 2019 (2019-03). + log_delivery_status: + $ref: '#/components/schemas/LogDeliveryStatus' + description: The LogDeliveryStatus of this log delivery configuration + status: + $ref: '#/components/schemas/LogDeliveryConfigStatus' + description: Status of log delivery configuration. Set to `ENABLED` (enabled) or `DISABLED` (disabled). Defaults + to `ENABLED`. You can [enable or disable the configuration](#operation/patch-log-delivery-config-status) later. + Deletion of a configuration is not supported, so disable a log delivery configuration that is no longer needed. + update_time: + type: integer + description: Time in epoch milliseconds when the log delivery configuration was updated. + workspace_ids_filter: + type: array + items: + type: integer + description: Optional filter that specifies workspace IDs to deliver logs for. By default the workspace filter is + empty and log delivery applies at the account level, delivering workspace-level logs for all workspaces in your + account, plus account level logs. You can optionally set this field to an array of workspace IDs (each one is + an `int64`) to which log delivery should apply, in which case only workspace-level logs relating to the specified + workspaces are delivered. If you plan to use different log delivery configurations for different workspaces, set + this field explicitly. Be aware that delivery configurations mentioning specific workspaces won't apply to new + workspaces created in the future, and delivery won't include account level logs. For some types of Databricks + deployments there is only one workspace per account ID, so this field is unnecessary. + required: + - log_type + - output_format + - credentials_id + - storage_configuration_id + description: '* Log Delivery Configuration' + LogDeliveryStatus: + type: object + properties: + status: + $ref: '#/components/schemas/DeliveryStatus' + message: + type: string + description: Informative message about the latest log delivery attempt. If the log delivery fails with USER_FAILURE, + error details will be provided for fixing misconfigurations in cloud permissions. + last_attempt_time: + type: string + description: The UTC time for the latest log delivery attempt. + last_successful_attempt_time: + type: string + description: The UTC time for the latest successful log delivery. + required: + - status + - message + PatchStatusResponse: + type: object + properties: {} + SortSpec: + type: object + properties: + descending: + type: boolean + field: + $ref: '#/components/schemas/SortSpecField' + description: The filed to sort by + UpdateBudgetConfigurationBudget: + type: object + properties: + account_id: + type: string + alert_configurations: + type: array + items: + $ref: '#/components/schemas/AlertConfiguration' + description: Alerts to configure when this budget is in a triggered state. Budgets must have exactly one alert configuration. + budget_configuration_id: + type: string + description: Databricks budget configuration ID. + display_name: + type: string + description: 'Human-readable name of budget configuration. Max Length: 128' + filter: + $ref: '#/components/schemas/BudgetConfigurationFilter' + description: Configured filters for this budget. These are applied to your account's usage to limit the scope of + what is considered for this budget. Leave empty to include all usage for this account. All provided filters must + be matched for usage to be included. + UpdateBudgetConfigurationResponse: + type: object + properties: + budget: + $ref: '#/components/schemas/BudgetConfiguration' + WrappedLogDeliveryConfiguration: + type: object + properties: + log_delivery_configuration: + $ref: '#/components/schemas/LogDeliveryConfiguration' + WrappedLogDeliveryConfigurations: + type: object + properties: + log_delivery_configurations: + type: array + items: + $ref: '#/components/schemas/LogDeliveryConfiguration' + next_page_token: + type: string + description: Token which can be sent as `page_token` to retrieve the next page of results. If this field is omitted, + there are no subsequent budgets. + ActionConfigurationType: + type: string + enum: + - EMAIL_NOTIFICATION + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + AlertConfigurationQuantityType: + type: string + enum: + - LIST_PRICE_DOLLARS_USD + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + AlertConfigurationTimePeriod: + type: string + enum: + - MONTH + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + AlertConfigurationTriggerType: + type: string + enum: + - CUMULATIVE_SPENDING_EXCEEDED + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + BudgetConfigurationFilterOperator: + type: string + enum: + - IN + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + DeliveryStatus: + type: string + enum: + - CREATED + - NOT_FOUND + - SUCCEEDED + - SYSTEM_FAILURE + - USER_FAILURE + description: '* The status string for log delivery. Possible values are: `CREATED`: There were no log delivery + + attempts since the config was created. `SUCCEEDED`: The latest attempt of log delivery has + + succeeded completely. `USER_FAILURE`: The latest attempt of log delivery failed because of + + misconfiguration of customer provided permissions on role or storage. `SYSTEM_FAILURE`: The + + latest attempt of log delivery failed because of an Databricks internal error. Contact support + + if it doesn''t go away soon. `NOT_FOUND`: The log delivery status as the configuration has been + + disabled since the release of this feature or there are no workspaces in the account.' + LogDeliveryConfigStatus: + type: string + enum: + - DISABLED + - ENABLED + description: '* Log Delivery Status + + + `ENABLED`: All dependencies have executed and succeeded `DISABLED`: At least one dependency has + + succeeded' + LogType: + type: string + enum: + - AUDIT_LOGS + - BILLABLE_USAGE + description: '* Log Delivery Type' + OutputFormat: + type: string + enum: + - CSV + - JSON + description: '* Log Delivery Output Format' + SortSpecField: + type: string + enum: + - POLICY_NAME + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + UsageDashboardMajorVersion: + type: string + enum: + - USAGE_DASHBOARD_MAJOR_VERSION_1 + - USAGE_DASHBOARD_MAJOR_VERSION_2 + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + UsageDashboardType: + type: string + enum: + - USAGE_DASHBOARD_TYPE_GLOBAL + - USAGE_DASHBOARD_TYPE_WORKSPACE + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + x-stackQL-resources: + billable_usage: + id: databricks_account.billing.billable_usage + name: billable_usage + title: Billable Usage + methods: + billable_usage_download: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1usage~1download/get' + response: + mediaType: text/plain + openAPIDocKey: '200' + overrideMediaType: application/json + transform: + body: '[{"contents": {{ toJson . }}}]' + type: golang_template_text_v0.3.0 + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/billable_usage/methods/billable_usage_download' + insert: [] + update: [] + delete: [] + replace: [] + budget_policy: + id: databricks_account.billing.budget_policy + name: budget_policy + title: Budget Policy + methods: + budget_policy_create: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budget-policies/post' + response: + mediaType: application/json + openAPIDocKey: '200' + budget_policy_list: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budget-policies/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.policies + budget_policy_delete: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budget-policies~1{policy_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + budget_policy_get: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budget-policies~1{policy_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + budget_policy_update: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budget-policies~1{policy_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/budget_policy/methods/budget_policy_get' + - $ref: '#/components/x-stackQL-resources/budget_policy/methods/budget_policy_list' + insert: + - $ref: '#/components/x-stackQL-resources/budget_policy/methods/budget_policy_create' + update: + - $ref: '#/components/x-stackQL-resources/budget_policy/methods/budget_policy_update' + delete: + - $ref: '#/components/x-stackQL-resources/budget_policy/methods/budget_policy_delete' + replace: [] + budgets: + id: databricks_account.billing.budgets + name: budgets + title: Budgets + methods: + budgets_create: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets/post' + response: + mediaType: application/json + openAPIDocKey: '200' + budgets_list: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.budgets + budgets_delete: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets~1{budget_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + budgets_get: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets~1{budget_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.budget + budgets_update: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1budgets~1{budget_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/budgets/methods/budgets_get' + - $ref: '#/components/x-stackQL-resources/budgets/methods/budgets_list' + insert: + - $ref: '#/components/x-stackQL-resources/budgets/methods/budgets_create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/budgets/methods/budgets_delete' + replace: + - $ref: '#/components/x-stackQL-resources/budgets/methods/budgets_update' + log_delivery: + id: databricks_account.billing.log_delivery + name: log_delivery + title: Log Delivery + methods: + log_delivery_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery/post' + response: + mediaType: application/json + openAPIDocKey: '200' + log_delivery_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.log_delivery_configurations + log_delivery_get: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery~1{log_delivery_configuration_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.log_delivery_configuration + log_delivery_patch_status: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery~1{log_delivery_configuration_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/log_delivery/methods/log_delivery_get' + - $ref: '#/components/x-stackQL-resources/log_delivery/methods/log_delivery_list' + insert: + - $ref: '#/components/x-stackQL-resources/log_delivery/methods/log_delivery_create' + update: + - $ref: '#/components/x-stackQL-resources/log_delivery/methods/log_delivery_patch_status' + delete: [] + replace: [] + usage_dashboards: + id: databricks_account.billing.usage_dashboards + name: usage_dashboards + title: Usage Dashboards + methods: + usage_dashboards_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1dashboard/post' + response: + mediaType: application/json + openAPIDocKey: '200' + usage_dashboards_get: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1dashboard/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/usage_dashboards/methods/usage_dashboards_get' + insert: + - $ref: '#/components/x-stackQL-resources/usage_dashboards/methods/usage_dashboards_create' + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/catalog.yaml b/providers/src/databricks_account/v00.00.00000/services/catalog.yaml new file mode 100644 index 00000000..1bb8ca53 --- /dev/null +++ b/providers/src/databricks_account/v00.00.00000/services/catalog.yaml @@ -0,0 +1,7303 @@ +openapi: 3.0.0 +info: + title: Databricks Catalog API (account) + description: >- + OpenAPI specification for the Databricks catalog service (account-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.catalog +servers: + - url: https://accounts.cloud.databricks.com +paths: + /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/metastores/{metastore_id}: + post: + operationId: account_metastore_assignments_create + summary: Creates an assignment to a metastore for a workspace + tags: + - catalog + - account_metastore_assignments + description: >- + Creates an assignment to a metastore for a workspace + + + :param workspace_id: int + Workspace ID. + :param metastore_id: str + Unity Catalog metastore ID + :param metastore_assignment: :class:`CreateMetastoreAssignment` + (optional) + + + :returns: :class:`AccountsCreateMetastoreAssignmentResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: Workspace ID. + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + requestBody: + content: + application/json: + schema: + type: object + properties: + metastore_assignment: + type: string + description: ':returns: :class:`AccountsCreateMetastoreAssignmentResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsCreateMetastoreAssignmentResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoreAssignmentsAPI + delete: + operationId: account_metastore_assignments_delete + summary: >- + Deletes a metastore assignment to a workspace, leaving the workspace + with no metastore. + tags: + - catalog + - account_metastore_assignments + description: >- + Deletes a metastore assignment to a workspace, leaving the workspace + with no metastore. + + + :param workspace_id: int + Workspace ID. + :param metastore_id: str + Unity Catalog metastore ID + + :returns: :class:`AccountsDeleteMetastoreAssignmentResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: Workspace ID. + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsDeleteMetastoreAssignmentResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoreAssignmentsAPI + put: + operationId: account_metastore_assignments_update + summary: >- + Updates an assignment to a metastore for a workspace. Currently, only + the default catalog may be + tags: + - catalog + - account_metastore_assignments + description: >- + Updates an assignment to a metastore for a workspace. Currently, only + the default catalog may be + + updated. + + + :param workspace_id: int + Workspace ID. + :param metastore_id: str + Unity Catalog metastore ID + :param metastore_assignment: :class:`UpdateMetastoreAssignment` + (optional) + + + :returns: :class:`AccountsUpdateMetastoreAssignmentResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: Workspace ID. + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + requestBody: + content: + application/json: + schema: + type: object + properties: + metastore_assignment: + type: string + description: ':returns: :class:`AccountsUpdateMetastoreAssignmentResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsUpdateMetastoreAssignmentResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoreAssignmentsAPI + /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/metastore: + get: + operationId: account_metastore_assignments_get + summary: >- + Gets the metastore assignment, if any, for the workspace specified by + ID. If the workspace is assigned + tags: + - catalog + - account_metastore_assignments + description: >- + Gets the metastore assignment, if any, for the workspace specified by + ID. If the workspace is assigned + + a metastore, the mapping will be returned. If no metastore is assigned + to the workspace, the + + assignment will not be found and a 404 returned. + + + :param workspace_id: int + Workspace ID. + + :returns: :class:`AccountsMetastoreAssignment` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: Workspace ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsMetastoreAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoreAssignmentsAPI + /api/2.0/accounts/{account_id}/metastores/{metastore_id}/workspaces: + get: + operationId: account_metastore_assignments_list + summary: >- + Gets a list of all Databricks workspace IDs that have been assigned to + given metastore. + tags: + - catalog + - account_metastore_assignments + description: >- + Gets a list of all Databricks workspace IDs that have been assigned to + given metastore. + + + :param metastore_id: str + Unity Catalog metastore ID + + :returns: Iterator over int + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetCatalogWorkspaceBindingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoreAssignmentsAPI + /api/2.0/accounts/{account_id}/metastores: + post: + operationId: account_metastores_create + summary: Creates a Unity Catalog metastore. + tags: + - catalog + - account_metastores + description: |- + Creates a Unity Catalog metastore. + + :param metastore_info: :class:`CreateAccountsMetastore` (optional) + + :returns: :class:`AccountsCreateMetastoreResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + metastore_info: + type: string + description: ':returns: :class:`AccountsCreateMetastoreResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsCreateMetastoreResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoresAPI + get: + operationId: account_metastores_list + summary: >- + Gets all Unity Catalog metastores associated with an account specified + by ID. + tags: + - catalog + - account_metastores + description: >- + Gets all Unity Catalog metastores associated with an account specified + by ID. + + + + :returns: Iterator over :class:`MetastoreInfo` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsListMetastoresResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoresAPI + /api/2.0/accounts/{account_id}/metastores/{metastore_id}: + delete: + operationId: account_metastores_delete + summary: Deletes a Unity Catalog metastore for an account, both specified by ID. + tags: + - catalog + - account_metastores + description: |- + Deletes a Unity Catalog metastore for an account, both specified by ID. + + :param metastore_id: str + Unity Catalog metastore ID + :param force: bool (optional) + Force deletion even if the metastore is not empty. Default is false. + + :returns: :class:`AccountsDeleteMetastoreResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + - name: force + in: query + required: false + schema: + type: string + description: Force deletion even if the metastore is not empty. Default is false. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsDeleteMetastoreResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoresAPI + get: + operationId: account_metastores_get + summary: Gets a Unity Catalog metastore from an account, both specified by ID. + tags: + - catalog + - account_metastores + description: |- + Gets a Unity Catalog metastore from an account, both specified by ID. + + :param metastore_id: str + Unity Catalog metastore ID + + :returns: :class:`AccountsGetMetastoreResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsGetMetastoreResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoresAPI + put: + operationId: account_metastores_update + summary: Updates an existing Unity Catalog metastore. + tags: + - catalog + - account_metastores + description: |- + Updates an existing Unity Catalog metastore. + + :param metastore_id: str + Unity Catalog metastore ID + :param metastore_info: :class:`UpdateAccountsMetastore` (optional) + Properties of the metastore to change. + + :returns: :class:`AccountsUpdateMetastoreResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + requestBody: + content: + application/json: + schema: + type: object + properties: + metastore_info: + type: string + description: Properties of the metastore to change. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsUpdateMetastoreResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountMetastoresAPI + /api/2.0/accounts/{account_id}/metastores/{metastore_id}/storage-credentials: + post: + operationId: account_storage_credentials_create + summary: >- + Creates a new storage credential. The request object is specific to the + cloud: - **AwsIamRole** for + tags: + - catalog + - account_storage_credentials + description: >- + Creates a new storage credential. The request object is specific to the + cloud: - **AwsIamRole** for + + AWS credentials - **AzureServicePrincipal** for Azure credentials - + **GcpServiceAccountKey** for GCP + + credentials + + + The caller must be a metastore admin and have the + `CREATE_STORAGE_CREDENTIAL` privilege on the + + metastore. + + + :param metastore_id: str + Unity Catalog metastore ID + :param credential_info: :class:`CreateAccountsStorageCredential` + (optional) + + :param skip_validation: bool (optional) + Optional, default false. Supplying true to this argument skips validation of the created set of + credentials. + + :returns: :class:`AccountsCreateStorageCredentialInfo` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + requestBody: + content: + application/json: + schema: + type: object + properties: + credential_info: + type: string + description: >- + :param skip_validation: bool (optional) Optional, default + false. Supplying true to this argument skips validation of + the created set of credentials. + skip_validation: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsCreateStorageCredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountStorageCredentialsAPI + get: + operationId: account_storage_credentials_list + summary: >- + Gets a list of all storage credentials that have been assigned to given + metastore. + tags: + - catalog + - account_storage_credentials + description: >- + Gets a list of all storage credentials that have been assigned to given + metastore. + + + :param metastore_id: str + Unity Catalog metastore ID + + :returns: Iterator over :class:`StorageCredentialInfo` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountStorageCredentialsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountStorageCredentialsAPI + /api/2.0/accounts/{account_id}/metastores/{metastore_id}/storage-credentials/{storage_credential_name}: + delete: + operationId: account_storage_credentials_delete + summary: >- + Deletes a storage credential from the metastore. The caller must be an + owner of the storage + tags: + - catalog + - account_storage_credentials + description: >- + Deletes a storage credential from the metastore. The caller must be an + owner of the storage + + credential. + + + :param metastore_id: str + Unity Catalog metastore ID + :param storage_credential_name: str + Name of the storage credential. + :param force: bool (optional) + Force deletion even if the Storage Credential is not empty. Default is false. + + :returns: :class:`AccountsDeleteStorageCredentialResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + - name: storage_credential_name + in: path + required: true + schema: + type: string + description: Name of the storage credential. + - name: force + in: query + required: false + schema: + type: string + description: >- + Force deletion even if the Storage Credential is not empty. Default + is false. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsDeleteStorageCredentialResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountStorageCredentialsAPI + get: + operationId: account_storage_credentials_get + summary: >- + Gets a storage credential from the metastore. The caller must be a + metastore admin, the owner of the + tags: + - catalog + - account_storage_credentials + description: >- + Gets a storage credential from the metastore. The caller must be a + metastore admin, the owner of the + + storage credential, or have a level of privilege on the storage + credential. + + + :param metastore_id: str + Unity Catalog metastore ID + :param storage_credential_name: str + Required. Name of the storage credential. + + :returns: :class:`AccountsStorageCredentialInfo` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + - name: storage_credential_name + in: path + required: true + schema: + type: string + description: Required. Name of the storage credential. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsStorageCredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountStorageCredentialsAPI + put: + operationId: account_storage_credentials_update + summary: >- + Updates a storage credential on the metastore. The caller must be the + owner of the storage credential. + tags: + - catalog + - account_storage_credentials + description: >- + Updates a storage credential on the metastore. The caller must be the + owner of the storage credential. + + If the caller is a metastore admin, only the **owner** credential can be + changed. + + + :param metastore_id: str + Unity Catalog metastore ID + :param storage_credential_name: str + Name of the storage credential. + :param credential_info: :class:`UpdateAccountsStorageCredential` + (optional) + + :param skip_validation: bool (optional) + Optional. Supplying true to this argument skips validation of the updated set of credentials. + + :returns: :class:`AccountsUpdateStorageCredentialResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: metastore_id + in: path + required: true + schema: + type: string + description: Unity Catalog metastore ID + - name: storage_credential_name + in: path + required: true + schema: + type: string + description: Name of the storage credential. + requestBody: + content: + application/json: + schema: + type: object + properties: + credential_info: + type: string + description: >- + :param skip_validation: bool (optional) Optional. Supplying + true to this argument skips validation of the updated set of + credentials. + skip_validation: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsUpdateStorageCredentialResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountStorageCredentialsAPI + /api/2.1/unity-catalog/credentials: + post: + operationId: credentials_create_credential + summary: >- + Creates a new credential. The type of credential to be created is + determined by the **purpose** field, + tags: + - catalog + - credentials + description: >- + Creates a new credential. The type of credential to be created is + determined by the **purpose** field, + + which should be either **SERVICE** or **STORAGE**. + + + The caller must be a metastore admin or have the metastore privilege + **CREATE_STORAGE_CREDENTIAL** for + + storage credentials, or **CREATE_SERVICE_CREDENTIAL** for service + credentials. + + + :param name: str + The credential name. The name must be unique among storage and service credentials within the + metastore. + :param aws_iam_role: :class:`AwsIamRole` (optional) + The AWS IAM role configuration. + :param azure_managed_identity: :class:`AzureManagedIdentity` (optional) + The Azure managed identity configuration. + :param azure_service_principal: :class:`AzureServicePrincipal` + (optional) + The Azure service principal configuration. + :param comment: str (optional) + Comment associated with the credential. + :param databricks_gcp_service_account: + :class:`DatabricksGcpServiceAccount` (optional) + The Databricks managed GCP service account configuration. + :param purpose: :class:`CredentialPurpose` (optional) + Indicates the purpose of the credential. + :param read_only: bool (optional) + Whether the credential is usable only for read operations. Only applicable when purpose is + **STORAGE**. + :param skip_validation: bool (optional) + Optional. Supplying true to this argument skips validation of the created set of credentials. + + :returns: :class:`CredentialInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: >- + The credential name. The name must be unique among storage + and service credentials within the metastore. + aws_iam_role: + type: string + description: The AWS IAM role configuration. + azure_managed_identity: + type: string + description: The Azure managed identity configuration. + azure_service_principal: + type: string + description: The Azure service principal configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + type: string + description: The Databricks managed GCP service account configuration. + purpose: + type: string + description: Indicates the purpose of the credential. + read_only: + type: string + description: >- + Whether the credential is usable only for read operations. + Only applicable when purpose is **STORAGE**. + skip_validation: + type: string + description: >- + Optional. Supplying true to this argument skips validation + of the created set of credentials. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + get: + operationId: credentials_list_credentials + summary: Gets an array of credentials (as __CredentialInfo__ objects). + tags: + - catalog + - credentials + description: >- + Gets an array of credentials (as __CredentialInfo__ objects). + + + The array is limited to only the credentials that the caller has + permission to access. If the caller + + is a metastore admin, retrieval of credentials is unrestricted. There is + no guarantee of a specific + + ordering of the elements in the array. + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param include_unbound: bool (optional) + Whether to include credentials not bound to the workspace. Effective only if the user has permission + to update the credential–workspace binding. + :param max_results: int (optional) + Maximum number of credentials to return. - If not set, the default max page size is used. - When set + to a value greater than 0, the page length is the minimum of this value and a server-configured + value. - When set to 0, the page length is set to a server-configured value (recommended). - When + set to a value less than 0, an invalid parameter error is returned. + :param page_token: str (optional) + Opaque token to retrieve the next page of results. + :param purpose: :class:`CredentialPurpose` (optional) + Return only credentials for the specified purpose. + + :returns: Iterator over :class:`CredentialInfo` + parameters: + - name: include_unbound + in: query + required: false + schema: + type: string + description: >- + Whether to include credentials not bound to the workspace. Effective + only if the user has permission to update the credential–workspace + binding. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of credentials to return. - If not set, the default + max page size is used. - When set to a value greater than 0, the + page length is the minimum of this value and a server-configured + value. - When set to 0, the page length is set to a + server-configured value (recommended). - When set to a value less + than 0, an invalid parameter error is returned. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque token to retrieve the next page of results. + - name: purpose + in: query + required: false + schema: + type: string + description: Return only credentials for the specified purpose. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCredentialsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + /api/2.1/unity-catalog/credentials/{name_arg}: + delete: + operationId: credentials_delete_credential + summary: >- + Deletes a service or storage credential from the metastore. The caller + must be an owner of the + tags: + - catalog + - credentials + description: >- + Deletes a service or storage credential from the metastore. The caller + must be an owner of the + + credential. + + + :param name_arg: str + Name of the credential. + :param force: bool (optional) + Force an update even if there are dependent services (when purpose is **SERVICE**) or dependent + external locations and external tables (when purpose is **STORAGE**). + parameters: + - name: name_arg + in: path + required: true + schema: + type: string + description: Name of the credential. + - name: force + in: query + required: false + schema: + type: string + description: >- + Force an update even if there are dependent services (when purpose + is **SERVICE**) or dependent external locations and external tables + (when purpose is **STORAGE**). + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + get: + operationId: credentials_get_credential + summary: >- + Gets a service or storage credential from the metastore. The caller must + be a metastore admin, the + tags: + - catalog + - credentials + description: >- + Gets a service or storage credential from the metastore. The caller must + be a metastore admin, the + + owner of the credential, or have any permission on the credential. + + + :param name_arg: str + Name of the credential. + + :returns: :class:`CredentialInfo` + parameters: + - name: name_arg + in: path + required: true + schema: + type: string + description: Name of the credential. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + patch: + operationId: credentials_update_credential + summary: Updates a service or storage credential on the metastore. + tags: + - catalog + - credentials + description: >- + Updates a service or storage credential on the metastore. + + + The caller must be the owner of the credential or a metastore admin or + have the `MANAGE` permission. + + If the caller is a metastore admin, only the __owner__ field can be + changed. + + + :param name_arg: str + Name of the credential. + :param aws_iam_role: :class:`AwsIamRole` (optional) + The AWS IAM role configuration. + :param azure_managed_identity: :class:`AzureManagedIdentity` (optional) + The Azure managed identity configuration. + :param azure_service_principal: :class:`AzureServicePrincipal` + (optional) + The Azure service principal configuration. + :param comment: str (optional) + Comment associated with the credential. + :param databricks_gcp_service_account: + :class:`DatabricksGcpServiceAccount` (optional) + The Databricks managed GCP service account configuration. + :param force: bool (optional) + Force an update even if there are dependent services (when purpose is **SERVICE**) or dependent + external locations and external tables (when purpose is **STORAGE**). + :param isolation_mode: :class:`IsolationMode` (optional) + Whether the current securable is accessible from all workspaces or a specific set of workspaces. + :param new_name: str (optional) + New name of credential. + :param owner: str (optional) + Username of current owner of credential. + :param read_only: bool (optional) + Whether the credential is usable only for read operations. Only applicable when purpose is + **STORAGE**. + :param skip_validation: bool (optional) + Supply true to this argument to skip validation of the updated credential. + + :returns: :class:`CredentialInfo` + parameters: + - name: name_arg + in: path + required: true + schema: + type: string + description: Name of the credential. + requestBody: + content: + application/json: + schema: + type: object + properties: + aws_iam_role: + type: string + description: The AWS IAM role configuration. + azure_managed_identity: + type: string + description: The Azure managed identity configuration. + azure_service_principal: + type: string + description: The Azure service principal configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + type: string + description: The Databricks managed GCP service account configuration. + force: + type: string + description: >- + Force an update even if there are dependent services (when + purpose is **SERVICE**) or dependent external locations and + external tables (when purpose is **STORAGE**). + isolation_mode: + type: string + description: >- + Whether the current securable is accessible from all + workspaces or a specific set of workspaces. + new_name: + type: string + description: New name of credential. + owner: + type: string + description: Username of current owner of credential. + read_only: + type: string + description: >- + Whether the credential is usable only for read operations. + Only applicable when purpose is **STORAGE**. + skip_validation: + type: string + description: >- + Supply true to this argument to skip validation of the + updated credential. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + /api/2.1/unity-catalog/temporary-service-credentials: + post: + operationId: credentials_generate_temporary_service_credential + summary: >- + Returns a set of temporary credentials generated using the specified + service credential. The caller + tags: + - catalog + - credentials + description: >- + Returns a set of temporary credentials generated using the specified + service credential. The caller + + must be a metastore admin or have the metastore privilege **ACCESS** on + the service credential. + + + :param credential_name: str + The name of the service credential used to generate a temporary credential + :param azure_options: + :class:`GenerateTemporaryServiceCredentialAzureOptions` (optional) + + :param gcp_options: + :class:`GenerateTemporaryServiceCredentialGcpOptions` (optional) + + + :returns: :class:`TemporaryCredentials` + requestBody: + content: + application/json: + schema: + type: object + properties: + credential_name: + type: string + description: >- + The name of the service credential used to generate a + temporary credential + azure_options: + type: string + description: >- + :param gcp_options: + :class:`GenerateTemporaryServiceCredentialGcpOptions` + (optional) + gcp_options: + type: string + required: + - credential_name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TemporaryCredentials' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + /api/2.1/unity-catalog/validate-credentials: + post: + operationId: credentials_validate_credential + summary: Validates a credential. + tags: + - catalog + - credentials + description: >- + Validates a credential. + + + For service credentials (purpose is **SERVICE**), either the + __credential_name__ or the cloud-specific + + credential must be provided. + + + For storage credentials (purpose is **STORAGE**), at least one of + __external_location_name__ and + + __url__ need to be provided. If only one of them is provided, it will be + used for validation. And if + + both are provided, the __url__ will be used for validation, and + __external_location_name__ will be + + ignored when checking overlapping urls. Either the __credential_name__ + or the cloud-specific + + credential must be provided. + + + The caller must be a metastore admin or the credential owner or have the + required permission on the + + metastore and the credential (e.g., **CREATE_EXTERNAL_LOCATION** when + purpose is **STORAGE**). + + + :param aws_iam_role: :class:`AwsIamRole` (optional) + + :param azure_managed_identity: :class:`AzureManagedIdentity` (optional) + + :param credential_name: str (optional) + Required. The name of an existing credential or long-lived cloud credential to validate. + :param databricks_gcp_service_account: + :class:`DatabricksGcpServiceAccount` (optional) + + :param external_location_name: str (optional) + The name of an existing external location to validate. Only applicable for storage credentials + (purpose is **STORAGE**.) + :param purpose: :class:`CredentialPurpose` (optional) + The purpose of the credential. This should only be used when the credential is specified. + :param read_only: bool (optional) + Whether the credential is only usable for read operations. Only applicable for storage credentials + (purpose is **STORAGE**.) + :param url: str (optional) + The external location url to validate. Only applicable when purpose is **STORAGE**. + + :returns: :class:`ValidateCredentialResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + aws_iam_role: + type: string + description: >- + :param azure_managed_identity: :class:`AzureManagedIdentity` + (optional) + azure_managed_identity: + type: string + credential_name: + type: string + description: >- + Required. The name of an existing credential or long-lived + cloud credential to validate. + databricks_gcp_service_account: + type: string + description: >- + :param external_location_name: str (optional) The name of an + existing external location to validate. Only applicable for + storage credentials (purpose is **STORAGE**.) + external_location_name: + type: string + purpose: + type: string + description: >- + The purpose of the credential. This should only be used when + the credential is specified. + read_only: + type: string + description: >- + Whether the credential is only usable for read operations. + Only applicable for storage credentials (purpose is + **STORAGE**.) + url: + type: string + description: >- + The external location url to validate. Only applicable when + purpose is **STORAGE**. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ValidateCredentialResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI +components: + schemas: + AccessRequestDestinations: + type: object + properties: + securable: + $ref: '#/components/schemas/Securable' + are_any_destinations_hidden: + type: boolean + description: >- + Indicates whether any destinations are hidden from the caller due to + a lack of permissions. This value is true if the caller does not + have permission to see all destinations. + destination_source_securable: + $ref: '#/components/schemas/Securable' + description: >- + The source securable from which the destinations are inherited. + Either the same value as securable (if destination is set directly + on the securable) or the nearest parent securable with destinations + set. + destinations: + type: array + items: + $ref: '#/components/schemas/NotificationDestination' + description: The access request destinations for the securable. + full_name: + type: string + description: >- + The full name of the securable. Redundant with the name in the + securable object, but necessary for Terraform integration + securable_type: + type: string + description: >- + The type of the securable. Redundant with the type in the securable + object, but necessary for Terraform integration + required: + - securable + AccountsCreateMetastoreAssignmentResponse: + type: object + properties: {} + description: The metastore assignment was successfully created. + AccountsCreateMetastoreResponse: + type: object + properties: + metastore_info: + $ref: '#/components/schemas/MetastoreInfo' + AccountsCreateStorageCredentialInfo: + type: object + properties: + credential_info: + $ref: '#/components/schemas/StorageCredentialInfo' + AccountsDeleteMetastoreAssignmentResponse: + type: object + properties: {} + description: The metastore assignment was successfully deleted. + AccountsDeleteMetastoreResponse: + type: object + properties: {} + description: The metastore was successfully deleted. + AccountsDeleteStorageCredentialResponse: + type: object + properties: {} + description: The storage credential was successfully deleted. + AccountsGetMetastoreResponse: + type: object + properties: + metastore_info: + $ref: '#/components/schemas/MetastoreInfo' + description: The metastore was successfully returned. + AccountsListMetastoresResponse: + type: object + properties: + metastores: + type: array + items: + $ref: '#/components/schemas/MetastoreInfo' + description: An array of metastore information objects. + description: Metastores were returned successfully. + AccountsMetastoreAssignment: + type: object + properties: + metastore_assignment: + $ref: '#/components/schemas/MetastoreAssignment' + description: The workspace metastore assignment was successfully returned. + AccountsStorageCredentialInfo: + type: object + properties: + credential_info: + $ref: '#/components/schemas/StorageCredentialInfo' + description: The storage credential was successfully retrieved. + AccountsUpdateMetastoreAssignmentResponse: + type: object + properties: {} + description: The metastore assignment was successfully updated. + AccountsUpdateMetastoreResponse: + type: object + properties: + metastore_info: + $ref: '#/components/schemas/MetastoreInfo' + description: The metastore update request succeeded. + AccountsUpdateStorageCredentialResponse: + type: object + properties: + credential_info: + $ref: '#/components/schemas/StorageCredentialInfo' + description: The storage credential was successfully updated. + ArtifactAllowlistInfo: + type: object + properties: + artifact_matchers: + type: array + items: + $ref: '#/components/schemas/ArtifactMatcher' + created_at: + type: integer + description: >- + Time at which this artifact allowlist was set, in epoch + milliseconds. + created_by: + type: string + description: Username of the user who set the artifact allowlist. + metastore_id: + type: string + description: Unique identifier of parent metastore. + ArtifactMatcher: + type: object + properties: + artifact: + type: string + match_type: + $ref: '#/components/schemas/MatchType' + description: The pattern matching type of the artifact + required: + - artifact + - match_type + AssignResponse: + type: object + properties: {} + AwsCredentials: + type: object + properties: + access_key_id: + type: string + description: The access key ID that identifies the temporary credentials. + access_point: + type: string + description: >- + The Amazon Resource Name (ARN) of the S3 access point for temporary + credentials related the external location. + secret_access_key: + type: string + description: The secret access key that can be used to sign AWS API requests. + session_token: + type: string + description: >- + The token that users must pass to AWS API to use the temporary + credentials. + description: |- + AWS temporary credentials for API authentication. Read more at + https://docs.aws.amazon.com/STS/latest/APIReference/API_Credentials.html. + AwsIamRole: + type: object + properties: + external_id: + type: string + description: >- + The external ID used in role assumption to prevent the confused + deputy problem. + role_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM role used to vend + temporary credentials. + unity_catalog_iam_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM user managed by + Databricks. This is the identity that is going to assume the AWS IAM + role. + description: The AWS IAM role configuration + AwsIamRoleRequest: + type: object + properties: + role_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM role used to vend + temporary credentials. + required: + - role_arn + description: The AWS IAM role configuration + AwsIamRoleResponse: + type: object + properties: + role_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM role used to vend + temporary credentials. + external_id: + type: string + description: >- + The external ID used in role assumption to prevent the confused + deputy problem. + unity_catalog_iam_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM user managed by + Databricks. This is the identity that is going to assume the AWS IAM + role. + required: + - role_arn + description: The AWS IAM role configuration + AwsSqsQueue: + type: object + properties: + managed_resource_id: + type: string + queue_url: + type: string + description: >- + The AQS queue url in the format + https://sqs.{region}.amazonaws.com/{account id}/{queue name}. Only + required for provided_sqs. + AzureActiveDirectoryToken: + type: object + properties: + aad_token: + type: string + description: >- + Opaque token that contains claims that you can use in Azure Active + Directory to access cloud services. + description: >- + Azure Active Directory token, essentially the Oauth token for Azure + Service Principal or Managed + Identity. Read more at + https://learn.microsoft.com/en-us/azure/databricks/dev-tools/api/latest/aad/service-prin-aad-token + AzureManagedIdentity: + type: object + properties: + access_connector_id: + type: string + description: >- + The Azure resource ID of the Azure Databricks Access Connector. Use + the format + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.Databricks/accessConnectors/{connector-name}`. + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + managed_identity_id: + type: string + description: >- + The Azure resource ID of the managed identity. Use the format, + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identity-name}` + This is only available for user-assgined identities. For + system-assigned identities, the access_connector_id is used to + identify the identity. If this field is not provided, then we assume + the AzureManagedIdentity is using the system-assigned identity. + required: + - access_connector_id + description: The Azure managed identity configuration. + AzureManagedIdentityRequest: + type: object + properties: + access_connector_id: + type: string + description: >- + The Azure resource ID of the Azure Databricks Access Connector. Use + the format + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.Databricks/accessConnectors/{connector-name}`. + managed_identity_id: + type: string + description: >- + The Azure resource ID of the managed identity. Use the format, + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identity-name}` + This is only available for user-assgined identities. For + system-assigned identities, the access_connector_id is used to + identify the identity. If this field is not provided, then we assume + the AzureManagedIdentity is using the system-assigned identity. + required: + - access_connector_id + description: The Azure managed identity configuration. + AzureManagedIdentityResponse: + type: object + properties: + access_connector_id: + type: string + description: >- + The Azure resource ID of the Azure Databricks Access Connector. Use + the format + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.Databricks/accessConnectors/{connector-name}`. + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + managed_identity_id: + type: string + description: >- + The Azure resource ID of the managed identity. Use the format, + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identity-name}` + This is only available for user-assgined identities. For + system-assigned identities, the access_connector_id is used to + identify the identity. If this field is not provided, then we assume + the AzureManagedIdentity is using the system-assigned identity. + required: + - access_connector_id + description: The Azure managed identity configuration. + AzureQueueStorage: + type: object + properties: + managed_resource_id: + type: string + queue_url: + type: string + description: >- + The AQS queue url in the format https://{storage + account}.queue.core.windows.net/{queue name} Only required for + provided_aqs. + resource_group: + type: string + description: >- + Optional resource group for the queue, event grid subscription, and + external location storage account. Only required for locations with + a service principal storage credential + subscription_id: + type: string + description: >- + Optional subscription id for the queue, event grid subscription, and + external location storage account. Required for locations with a + service principal storage credential + AzureServicePrincipal: + type: object + properties: + directory_id: + type: string + description: >- + The directory ID corresponding to the Azure Active Directory (AAD) + tenant of the application. + application_id: + type: string + description: >- + The application ID of the application registration within the + referenced AAD tenant. + client_secret: + type: string + description: The client secret generated for the above app ID in AAD. + required: + - directory_id + - application_id + - client_secret + description: >- + The Azure service principal configuration. Only applicable when purpose + is **STORAGE**. + AzureUserDelegationSas: + type: object + properties: + sas_token: + type: string + description: >- + The signed URI (SAS Token) used to access blob services for a given + path + description: |- + Azure temporary credentials for API authentication. Read more at + https://docs.microsoft.com/en-us/rest/api/storageservices/create-user-delegation-sas + BatchCreateAccessRequestsResponse: + type: object + properties: + responses: + type: array + items: + $ref: '#/components/schemas/CreateAccessRequestResponse' + CancelRefreshResponse: + type: object + properties: {} + CatalogInfo: + type: object + properties: + browse_only: + type: boolean + catalog_type: + $ref: '#/components/schemas/CatalogType' + comment: + type: string + description: User-provided free-form text description. + connection_name: + type: string + description: The name of the connection to an external data source. + created_at: + type: integer + description: Time at which this catalog was created, in epoch milliseconds. + created_by: + type: string + description: Username of catalog creator. + effective_predictive_optimization_flag: + $ref: '#/components/schemas/EffectivePredictiveOptimizationFlag' + enable_predictive_optimization: + $ref: '#/components/schemas/EnablePredictiveOptimization' + description: >- + Whether predictive optimization should be enabled for this object + and objects under it. + full_name: + type: string + description: The full name of the catalog. Corresponds with the name field. + isolation_mode: + $ref: '#/components/schemas/CatalogIsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of catalog. + options: + type: object + description: A map of key-value properties attached to the securable. + owner: + type: string + description: Username of current owner of catalog. + properties: + type: object + description: A map of key-value properties attached to the securable. + provider_name: + type: string + description: >- + The name of delta sharing provider. A Delta Sharing catalog is a + catalog that is based on a Delta share on a remote sharing server. + provisioning_info: + $ref: '#/components/schemas/ProvisioningInfo' + securable_type: + $ref: '#/components/schemas/SecurableType' + share_name: + type: string + description: The name of the share under the share provider. + storage_location: + type: string + description: Storage Location URL (full path) for managed tables within catalog. + storage_root: + type: string + description: Storage root URL for managed tables within catalog. + updated_at: + type: integer + description: Time at which this catalog was last modified, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified catalog. + CloudflareApiToken: + type: object + properties: + access_key_id: + type: string + description: The access key ID associated with the API token. + secret_access_key: + type: string + description: The secret access token generated for the above access key ID. + account_id: + type: string + description: The ID of the account associated with the API token. + required: + - access_key_id + - secret_access_key + description: |- + The Cloudflare API token configuration. Read more at + https://developers.cloudflare.com/r2/api/s3/tokens/ + ColumnInfo: + type: object + properties: + comment: + type: string + mask: + $ref: '#/components/schemas/ColumnMask' + name: + type: string + description: Name of Column. + nullable: + type: boolean + description: 'Whether field may be Null (default: true).' + partition_index: + type: integer + description: Partition index for column. + position: + type: integer + description: Ordinal position of column (starting at position 0). + type_interval_type: + type: string + description: Format of IntervalType. + type_json: + type: string + description: Full data type specification, JSON-serialized. + type_name: + $ref: '#/components/schemas/ColumnTypeName' + type_precision: + type: integer + description: Digits of precision; required for DecimalTypes. + type_scale: + type: integer + description: Digits to right of decimal; Required for DecimalTypes. + type_text: + type: string + description: Full data type specification as SQL/catalogString text. + ColumnMask: + type: object + properties: + function_name: + type: string + using_column_names: + type: array + items: + type: string + description: >- + The list of additional table columns to be passed as input to the + column mask function. The first arg of the mask function should be + of the type of the column being masked and the types of the rest of + the args should match the types of columns in 'using_column_names'. + ColumnMaskOptions: + type: object + properties: + function_name: + type: string + on_column: + type: string + description: >- + The alias of the column to be masked. The alias must refer to one of + matched columns. The values of the column is passed to the column + mask function as the first argument. Required on create and update. + using: + type: array + items: + $ref: '#/components/schemas/FunctionArgument' + description: >- + Optional list of column aliases or constant literals to be passed as + additional arguments to the column mask function. The type of each + column should match the positional argument of the column mask + function. + required: + - function_name + - on_column + ColumnRelationship: + type: object + properties: + source: + type: string + target: + type: string + ConnectionDependency: + type: object + properties: + connection_name: + type: string + description: >- + Full name of the dependent connection, in the form of + __connection_name__. + description: A connection that is dependent on a SQL object. + ConnectionInfo: + type: object + properties: + comment: + type: string + description: User-provided free-form text description. + connection_id: + type: string + description: Unique identifier of the Connection. + connection_type: + $ref: '#/components/schemas/ConnectionType' + description: The type of connection. + created_at: + type: integer + description: Time at which this connection was created, in epoch milliseconds. + created_by: + type: string + description: Username of connection creator. + credential_type: + $ref: '#/components/schemas/CredentialType' + description: The type of credential. + full_name: + type: string + description: Full name of connection. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of the connection. + options: + type: object + description: A map of key-value properties attached to the securable. + owner: + type: string + description: Username of current owner of the connection. + properties: + type: object + description: A map of key-value properties attached to the securable. + provisioning_info: + $ref: '#/components/schemas/ProvisioningInfo' + read_only: + type: boolean + description: If the connection is read only. + securable_type: + $ref: '#/components/schemas/SecurableType' + updated_at: + type: integer + description: Time at which this connection was updated, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified connection. + url: + type: string + description: URL of the remote data source, extracted from options. + description: 'Next ID: 24' + ContinuousUpdateStatus: + type: object + properties: + initial_pipeline_sync_progress: + $ref: '#/components/schemas/PipelineProgress' + description: Progress of the initial data synchronization. + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was synced to the online + table. Note that this Delta version may not be completely synced to + the online table yet. + timestamp: + type: string + description: >- + The timestamp of the last time any data was synchronized from the + source table to the online table. + description: >- + Detailed status of an online table. Shown if the online table is in the + ONLINE_CONTINUOUS_UPDATE + or the ONLINE_UPDATING_PIPELINE_RESOURCES state. + CreateAccessRequest: + type: object + properties: + behalf_of: + $ref: '#/components/schemas/Principal' + comment: + type: string + description: >- + Optional. Comment associated with the request. At most 200 + characters, can only contain lowercase/uppercase letters (a-z, A-Z), + numbers (0-9), punctuation, and spaces. + securable_permissions: + type: array + items: + $ref: '#/components/schemas/SecurablePermissions' + description: >- + List of securables and their corresponding requested UC privileges. + At most 30 securables can be requested for a principal per batched + call. Each securable can only be requested once per principal. + CreateAccessRequestResponse: + type: object + properties: + behalf_of: + $ref: '#/components/schemas/Principal' + request_destinations: + type: array + items: + $ref: '#/components/schemas/AccessRequestDestinations' + description: >- + The access request destinations for all the securables the principal + requested. + CreateAccountsMetastore: + type: object + properties: + name: + type: string + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + region: + type: string + description: >- + Cloud region which the metastore serves (e.g., `us-west-2`, + `westus`). + storage_root: + type: string + description: The storage root URL for metastore + required: + - name + CreateAccountsStorageCredential: + type: object + properties: + name: + type: string + aws_iam_role: + $ref: '#/components/schemas/AwsIamRoleRequest' + description: The AWS IAM role configuration. + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentityRequest' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + cloudflare_api_token: + $ref: '#/components/schemas/CloudflareApiToken' + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccountRequest' + description: The Databricks managed GCP service account configuration. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + required: + - name + CreateFunction: + type: object + properties: + name: + type: string + catalog_name: + type: string + description: Name of parent Catalog. + schema_name: + type: string + description: Name of parent Schema relative to its parent Catalog. + input_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Function input parameters. + data_type: + $ref: '#/components/schemas/ColumnTypeName' + description: Scalar function return data type. + full_data_type: + type: string + description: Pretty printed function data type. + routine_body: + $ref: '#/components/schemas/CreateFunctionRoutineBody' + description: >- + Function language. When **EXTERNAL** is used, the language of the + routine function should be specified in the **external_language** + field, and the **return_params** of the function cannot be used (as + **TABLE** return type is not supported), and the **sql_data_access** + field must be **NO_SQL**. + routine_definition: + type: string + description: Function body. + parameter_style: + $ref: '#/components/schemas/CreateFunctionParameterStyle' + description: Function parameter style. **S** is the value for SQL. + is_deterministic: + type: boolean + description: Whether the function is deterministic. + sql_data_access: + $ref: '#/components/schemas/CreateFunctionSqlDataAccess' + description: Function SQL data access. + is_null_call: + type: boolean + description: Function null call. + security_type: + $ref: '#/components/schemas/CreateFunctionSecurityType' + description: Function security type. + specific_name: + type: string + description: Specific name of the function; Reserved for future use. + comment: + type: string + description: User-provided free-form text description. + external_language: + type: string + description: External function language. + external_name: + type: string + description: External function name. + properties: + type: string + description: JSON-serialized key-value pair map, encoded (escaped) as a string. + return_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Table function return parameters. + routine_dependencies: + $ref: '#/components/schemas/DependencyList' + description: function dependencies. + sql_path: + type: string + description: >- + List of schemes whose objects can be referenced without + qualification. + required: + - name + - catalog_name + - schema_name + - input_params + - data_type + - full_data_type + - routine_body + - routine_definition + - parameter_style + - is_deterministic + - sql_data_access + - is_null_call + - security_type + - specific_name + CreateMetastoreAssignment: + type: object + properties: + workspace_id: + type: integer + metastore_id: + type: string + description: The unique ID of the metastore. + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field is + deprecated. Please use "Default Namespace API" to configure the + default catalog for a Databricks workspace. + required: + - workspace_id + - metastore_id + - default_catalog_name + CreateRequestExternalLineage: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + CredentialDependency: + type: object + properties: + credential_name: + type: string + description: >- + Full name of the dependent credential, in the form of + __credential_name__. + description: A credential that is dependent on a SQL object. + CredentialInfo: + type: object + properties: + aws_iam_role: + $ref: '#/components/schemas/AwsIamRole' + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentity' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + comment: + type: string + description: Comment associated with the credential. + created_at: + type: integer + description: Time at which this credential was created, in epoch milliseconds. + created_by: + type: string + description: Username of credential creator. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccount' + description: The Databricks managed GCP service account configuration. + full_name: + type: string + description: The full name of the credential. + id: + type: string + description: The unique identifier of the credential. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + metastore_id: + type: string + description: Unique identifier of the parent metastore. + name: + type: string + description: >- + The credential name. The name must be unique among storage and + service credentials within the metastore. + owner: + type: string + description: Username of current owner of credential. + purpose: + $ref: '#/components/schemas/CredentialPurpose' + description: Indicates the purpose of the credential. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + updated_at: + type: integer + description: >- + Time at which this credential was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the credential. + used_for_managed_storage: + type: boolean + description: >- + Whether this credential is the current metastore's root storage + credential. Only applicable when purpose is **STORAGE**. + CredentialValidationResult: + type: object + properties: + message: + type: string + result: + $ref: '#/components/schemas/ValidateCredentialResult' + description: The results of the tested operation. + DatabricksGcpServiceAccount: + type: object + properties: + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + email: + type: string + description: The email of the service account. + private_key_id: + type: string + description: The ID that represents the private key for this Service Account + description: >- + GCP long-lived credential. Databricks-created Google Cloud Storage + service account. + DatabricksGcpServiceAccountRequest: + type: object + properties: {} + description: >- + GCP long-lived credential. Databricks-created Google Cloud Storage + service account. + DatabricksGcpServiceAccountResponse: + type: object + properties: + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + email: + type: string + description: The email of the service account. + description: >- + GCP long-lived credential. Databricks-created Google Cloud Storage + service account. + DeleteCredentialResponse: + type: object + properties: {} + DeleteMonitorResponse: + type: object + properties: {} + DeletePolicyResponse: + type: object + properties: {} + DeleteRequestExternalLineage: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + id: + type: string + description: Unique identifier of the external lineage relationship. + required: + - source + - target + DeleteResponse: + type: object + properties: {} + DeleteTableConstraintResponse: + type: object + properties: {} + DeltaRuntimePropertiesKvPairs: + type: object + properties: + delta_runtime_properties: + type: object + description: A map of key-value properties attached to the securable. + required: + - delta_runtime_properties + description: >- + Properties pertaining to the current state of the delta table as given + by the commit server. + This does not contain **delta.*** (input) properties in __TableInfo.properties__. + Dependency: + type: object + properties: + connection: + $ref: '#/components/schemas/ConnectionDependency' + credential: + $ref: '#/components/schemas/CredentialDependency' + function: + $ref: '#/components/schemas/FunctionDependency' + table: + $ref: '#/components/schemas/TableDependency' + description: >- + A dependency of a SQL object. One of the following fields must be + defined: __table__, + __function__, __connection__, or __credential__. + DependencyList: + type: object + properties: + dependencies: + type: array + items: + $ref: '#/components/schemas/Dependency' + description: Array of dependencies. + description: A list of dependencies. + DisableResponse: + type: object + properties: {} + EffectivePermissionsList: + type: object + properties: + next_page_token: + type: string + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/EffectivePrivilegeAssignment' + description: >- + The privileges conveyed to each principal (either directly or via + inheritance) + EffectivePredictiveOptimizationFlag: + type: object + properties: + value: + $ref: '#/components/schemas/EnablePredictiveOptimization' + inherited_from_name: + type: string + description: >- + The name of the object from which the flag was inherited. If there + was no inheritance, this field is left blank. + inherited_from_type: + $ref: >- + #/components/schemas/EffectivePredictiveOptimizationFlagInheritedFromType + description: >- + The type of the object from which the flag was inherited. If there + was no inheritance, this field is left blank. + required: + - value + EffectivePrivilege: + type: object + properties: + inherited_from_name: + type: string + inherited_from_type: + $ref: '#/components/schemas/SecurableType' + description: >- + The type of the object that conveys this privilege via inheritance. + This field is omitted when privilege is not inherited (it's assigned + to the securable itself). + privilege: + $ref: '#/components/schemas/Privilege' + description: The privilege assigned to the principal. + EffectivePrivilegeAssignment: + type: object + properties: + principal: + type: string + privileges: + type: array + items: + $ref: '#/components/schemas/EffectivePrivilege' + description: >- + The privileges conveyed to the principal (either directly or via + inheritance). + EnableResponse: + type: object + properties: {} + EncryptionDetails: + type: object + properties: + sse_encryption_details: + $ref: '#/components/schemas/SseEncryptionDetails' + description: >- + Server-Side Encryption properties for clients communicating with AWS + s3. + description: Encryption options that apply to clients connecting to cloud storage. + EntityTagAssignment: + type: object + properties: + entity_name: + type: string + description: The fully qualified name of the entity to which the tag is assigned + tag_key: + type: string + description: The key of the tag + entity_type: + type: string + description: >- + The type of the entity to which the tag is assigned. Allowed values + are: catalogs, schemas, tables, columns, volumes. + tag_value: + type: string + description: The value of the tag + required: + - entity_name + - tag_key + - entity_type + description: Represents a tag assignment to an entity + ExternalLineageExternalMetadata: + type: object + properties: + name: + type: string + ExternalLineageExternalMetadataInfo: + type: object + properties: + entity_type: + type: string + description: Type of entity represented by the external metadata object. + event_time: + type: string + description: Timestamp of the lineage event. + name: + type: string + description: Name of the external metadata object. + system_type: + $ref: '#/components/schemas/SystemType' + description: Type of external system. + description: Represents the external metadata object in the lineage event. + ExternalLineageFileInfo: + type: object + properties: + event_time: + type: string + description: Timestamp of the lineage event. + path: + type: string + description: URL of the path. + securable_name: + type: string + description: The full name of the securable on the path. + securable_type: + type: string + description: The securable type of the securable on the path. + storage_location: + type: string + description: The storage location associated with securable on the path. + description: Represents the path information in the lineage event. + ExternalLineageInfo: + type: object + properties: + external_lineage_info: + $ref: '#/components/schemas/ExternalLineageRelationshipInfo' + description: >- + Information about the edge metadata of the external lineage + relationship. + external_metadata_info: + $ref: '#/components/schemas/ExternalLineageExternalMetadataInfo' + description: >- + Information about external metadata involved in the lineage + relationship. + file_info: + $ref: '#/components/schemas/ExternalLineageFileInfo' + description: Information about the file involved in the lineage relationship. + model_info: + $ref: '#/components/schemas/ExternalLineageModelVersionInfo' + description: >- + Information about the model version involved in the lineage + relationship. + table_info: + $ref: '#/components/schemas/ExternalLineageTableInfo' + description: Information about the table involved in the lineage relationship. + description: Lineage response containing lineage information of a data asset. + ExternalLineageModelVersion: + type: object + properties: + name: + type: string + version: + type: string + ExternalLineageModelVersionInfo: + type: object + properties: + event_time: + type: string + description: Timestamp of the lineage event. + model_name: + type: string + description: Name of the model. + version: + type: integer + description: Version number of the model. + description: Represents the model version information in the lineage event. + ExternalLineageObject: + type: object + properties: + external_metadata: + $ref: '#/components/schemas/ExternalLineageExternalMetadata' + model_version: + $ref: '#/components/schemas/ExternalLineageModelVersion' + path: + $ref: '#/components/schemas/ExternalLineagePath' + table: + $ref: '#/components/schemas/ExternalLineageTable' + ExternalLineagePath: + type: object + properties: + url: + type: string + ExternalLineageRelationship: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + ExternalLineageRelationshipInfo: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + ExternalLineageTable: + type: object + properties: + name: + type: string + ExternalLineageTableInfo: + type: object + properties: + catalog_name: + type: string + description: Name of Catalog. + event_time: + type: string + description: Timestamp of the lineage event. + name: + type: string + description: Name of Table. + schema_name: + type: string + description: Name of Schema. + description: Represents the table information in the lineage event. + ExternalLocationInfo: + type: object + properties: + browse_only: + type: boolean + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: >- + Time at which this external location was created, in epoch + milliseconds. + created_by: + type: string + description: Username of external location creator. + credential_id: + type: string + description: Unique ID of the location's storage credential. + credential_name: + type: string + description: Name of the storage credential used with this location. + enable_file_events: + type: boolean + description: >- + Whether to enable file events on this external location. Default to + `true`. Set to `false` to disable file events. + encryption_details: + $ref: '#/components/schemas/EncryptionDetails' + fallback: + type: boolean + description: >- + Indicates whether fallback mode is enabled for this external + location. When fallback mode is enabled, the access to the location + falls back to cluster credentials if UC credentials are not + sufficient. + file_event_queue: + $ref: '#/components/schemas/FileEventQueue' + description: >- + File event queue settings. If `enable_file_events` is not `false`, + must be defined and have exactly one of the documented properties. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + metastore_id: + type: string + description: Unique identifier of metastore hosting the external location. + name: + type: string + description: Name of the external location. + owner: + type: string + description: The owner of the external location. + read_only: + type: boolean + description: Indicates whether the external location is read-only. + updated_at: + type: integer + description: >- + Time at which external location this was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the external location. + url: + type: string + description: Path URL of the external location. + ExternalMetadata: + type: object + properties: + name: + type: string + system_type: + $ref: '#/components/schemas/SystemType' + description: Type of external system. + entity_type: + type: string + description: Type of entity within the external system. + columns: + type: array + items: + type: string + description: List of columns associated with the external metadata object. + create_time: + type: string + description: Time at which this external metadata object was created. + created_by: + type: string + description: Username of external metadata object creator. + description: + type: string + description: User-provided free-form text description. + id: + type: string + description: Unique identifier of the external metadata object. + metastore_id: + type: string + description: Unique identifier of parent metastore. + owner: + type: string + description: Owner of the external metadata object. + properties: + type: object + description: >- + A map of key-value properties attached to the external metadata + object. + update_time: + type: string + description: Time at which this external metadata object was last modified. + updated_by: + type: string + description: Username of user who last modified external metadata object. + url: + type: string + description: URL associated with the external metadata object. + required: + - name + - system_type + - entity_type + FailedStatus: + type: object + properties: + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was synced to the online + table. Note that this Delta version may only be partially synced to + the online table. Only populated if the table is still online and + available for serving. + timestamp: + type: string + description: >- + The timestamp of the last time any data was synchronized from the + source table to the online table. Only populated if the table is + still online and available for serving. + description: >- + Detailed status of an online table. Shown if the online table is in the + OFFLINE_FAILED or the + ONLINE_PIPELINE_FAILED state. + FileEventQueue: + type: object + properties: + managed_aqs: + $ref: '#/components/schemas/AzureQueueStorage' + managed_pubsub: + $ref: '#/components/schemas/GcpPubsub' + managed_sqs: + $ref: '#/components/schemas/AwsSqsQueue' + provided_aqs: + $ref: '#/components/schemas/AzureQueueStorage' + provided_pubsub: + $ref: '#/components/schemas/GcpPubsub' + provided_sqs: + $ref: '#/components/schemas/AwsSqsQueue' + ForeignKeyConstraint: + type: object + properties: + name: + type: string + child_columns: + type: array + items: + type: string + description: Column names for this constraint. + parent_table: + type: string + description: The full name of the parent constraint. + parent_columns: + type: array + items: + type: string + description: Column names for this constraint. + rely: + type: boolean + description: True if the constraint is RELY, false or unset if NORELY. + required: + - name + - child_columns + - parent_table + - parent_columns + FunctionArgument: + type: object + properties: + alias: + type: string + constant: + type: string + description: A constant literal. + FunctionDependency: + type: object + properties: + function_full_name: + type: string + description: >- + Full name of the dependent function, in the form of + __catalog_name__.__schema_name__.__function_name__. + required: + - function_full_name + description: A function that is dependent on a SQL object. + FunctionInfo: + type: object + properties: + browse_only: + type: boolean + catalog_name: + type: string + description: Name of parent Catalog. + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: Time at which this function was created, in epoch milliseconds. + created_by: + type: string + description: Username of function creator. + data_type: + $ref: '#/components/schemas/ColumnTypeName' + description: Scalar function return data type. + external_language: + type: string + description: External function language. + external_name: + type: string + description: External function name. + full_data_type: + type: string + description: Pretty printed function data type. + full_name: + type: string + description: >- + Full name of Function, in form of + **catalog_name**.**schema_name**.**function_name** + function_id: + type: string + description: Id of Function, relative to parent schema. + input_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Function input parameters. + is_deterministic: + type: boolean + description: Whether the function is deterministic. + is_null_call: + type: boolean + description: Function null call. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of function, relative to parent schema. + owner: + type: string + description: Username of current owner of the function. + parameter_style: + $ref: '#/components/schemas/FunctionInfoParameterStyle' + description: Function parameter style. **S** is the value for SQL. + properties: + type: string + description: JSON-serialized key-value pair map, encoded (escaped) as a string. + return_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Table function return parameters. + routine_body: + $ref: '#/components/schemas/FunctionInfoRoutineBody' + description: >- + Function language. When **EXTERNAL** is used, the language of the + routine function should be specified in the **external_language** + field, and the **return_params** of the function cannot be used (as + **TABLE** return type is not supported), and the **sql_data_access** + field must be **NO_SQL**. + routine_definition: + type: string + description: Function body. + routine_dependencies: + $ref: '#/components/schemas/DependencyList' + description: function dependencies. + schema_name: + type: string + description: Name of parent Schema relative to its parent Catalog. + security_type: + $ref: '#/components/schemas/FunctionInfoSecurityType' + description: Function security type. + specific_name: + type: string + description: Specific name of the function; Reserved for future use. + sql_data_access: + $ref: '#/components/schemas/FunctionInfoSqlDataAccess' + description: Function SQL data access. + sql_path: + type: string + description: >- + List of schemes whose objects can be referenced without + qualification. + updated_at: + type: integer + description: >- + Time at which this function was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the function. + FunctionParameterInfo: + type: object + properties: + name: + type: string + type_text: + type: string + description: Full data type spec, SQL/catalogString text. + type_name: + $ref: '#/components/schemas/ColumnTypeName' + description: Name of type (INT, STRUCT, MAP, etc.) + position: + type: integer + description: Ordinal position of column (starting at position 0). + comment: + type: string + description: User-provided free-form text description. + parameter_default: + type: string + description: Default value of the parameter. + parameter_mode: + $ref: '#/components/schemas/FunctionParameterMode' + description: Function parameter mode. + parameter_type: + $ref: '#/components/schemas/FunctionParameterType' + description: Function parameter type. + type_interval_type: + type: string + description: Format of IntervalType. + type_json: + type: string + description: Full data type spec, JSON-serialized. + type_precision: + type: integer + description: Digits of precision; required on Create for DecimalTypes. + type_scale: + type: integer + description: Digits to right of decimal; Required on Create for DecimalTypes. + required: + - name + - type_text + - type_name + - position + FunctionParameterInfos: + type: object + properties: + parameters: + type: array + items: + $ref: '#/components/schemas/FunctionParameterInfo' + GcpOauthToken: + type: object + properties: + oauth_token: + type: string + description: |- + GCP temporary credentials for API authentication. Read more at + https://developers.google.com/identity/protocols/oauth2/service-account + GcpPubsub: + type: object + properties: + managed_resource_id: + type: string + subscription_name: + type: string + description: >- + The Pub/Sub subscription name in the format + projects/{project}/subscriptions/{subscription name}. Only required + for provided_pubsub. + GenerateTemporaryPathCredentialResponse: + type: object + properties: + aws_temp_credentials: + $ref: '#/components/schemas/AwsCredentials' + azure_aad: + $ref: '#/components/schemas/AzureActiveDirectoryToken' + azure_user_delegation_sas: + $ref: '#/components/schemas/AzureUserDelegationSas' + expiration_time: + type: integer + description: >- + Server time when the credential will expire, in epoch milliseconds. + The API client is advised to cache the credential given this + expiration time. + gcp_oauth_token: + $ref: '#/components/schemas/GcpOauthToken' + r2_temp_credentials: + $ref: '#/components/schemas/R2Credentials' + url: + type: string + description: The URL of the storage path accessible by the temporary credential. + GenerateTemporaryServiceCredentialAzureOptions: + type: object + properties: + resources: + type: array + items: + type: string + description: >- + The resources to which the temporary Azure credential should apply. + These resources are the scopes that are passed to the token provider + (see + https://learn.microsoft.com/python/api/azure-core/azure.core.credentials.tokencredential?view=azure-python) + description: The Azure cloud options to customize the requested temporary credential + GenerateTemporaryServiceCredentialGcpOptions: + type: object + properties: + scopes: + type: array + items: + type: string + description: >- + The scopes to which the temporary GCP credential should apply. These + resources are the scopes that are passed to the token provider (see + https://google-auth.readthedocs.io/en/latest/reference/google.auth.html#google.auth.credentials.Credentials) + description: The GCP cloud options to customize the requested temporary credential + GenerateTemporaryTableCredentialResponse: + type: object + properties: + aws_temp_credentials: + $ref: '#/components/schemas/AwsCredentials' + azure_aad: + $ref: '#/components/schemas/AzureActiveDirectoryToken' + azure_user_delegation_sas: + $ref: '#/components/schemas/AzureUserDelegationSas' + expiration_time: + type: integer + description: >- + Server time when the credential will expire, in epoch milliseconds. + The API client is advised to cache the credential given this + expiration time. + gcp_oauth_token: + $ref: '#/components/schemas/GcpOauthToken' + r2_temp_credentials: + $ref: '#/components/schemas/R2Credentials' + url: + type: string + description: The URL of the storage path accessible by the temporary credential. + GetCatalogWorkspaceBindingsResponse: + type: object + properties: + workspaces: + type: array + items: + type: integer + GetMetastoreSummaryResponse: + type: object + properties: + cloud: + type: string + created_at: + type: integer + description: Time at which this metastore was created, in epoch milliseconds. + created_by: + type: string + description: Username of metastore creator. + default_data_access_config_id: + type: string + description: >- + Unique identifier of the metastore's (Default) Data Access + Configuration. + delta_sharing_organization_name: + type: string + description: >- + The organization name of a Delta Sharing entity, to be used in + Databricks-to-Databricks Delta Sharing as the official name. + delta_sharing_recipient_token_lifetime_in_seconds: + type: integer + description: The lifetime of delta sharing recipient token in seconds. + delta_sharing_scope: + $ref: '#/components/schemas/DeltaSharingScopeEnum' + description: The scope of Delta Sharing enabled for the metastore. + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + global_metastore_id: + type: string + description: >- + Globally unique metastore ID across clouds and regions, of the form + `cloud:region:metastore_id`. + metastore_id: + type: string + description: Unique identifier of metastore. + name: + type: string + description: The user-specified name of the metastore. + owner: + type: string + description: The owner of the metastore. + privilege_model_version: + type: string + description: >- + Privilege model version of the metastore, of the form `major.minor` + (e.g., `1.0`). + region: + type: string + description: >- + Cloud region which the metastore serves (e.g., `us-west-2`, + `westus`). + storage_root: + type: string + description: The storage root URL for metastore + storage_root_credential_id: + type: string + description: UUID of storage credential to access the metastore storage_root. + storage_root_credential_name: + type: string + description: Name of the storage credential to access the metastore storage_root. + updated_at: + type: integer + description: >- + Time at which the metastore was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the metastore. + GetPermissionsResponse: + type: object + properties: + next_page_token: + type: string + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/PrivilegeAssignment' + description: The privileges assigned to each principal + GetQuotaResponse: + type: object + properties: + quota_info: + $ref: '#/components/schemas/QuotaInfo' + GetWorkspaceBindingsResponse: + type: object + properties: + bindings: + type: array + items: + $ref: '#/components/schemas/WorkspaceBinding' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListAccountMetastoreAssignmentsResponse: + type: object + properties: + workspace_ids: + type: array + items: + type: integer + description: The metastore assignments were successfully returned. + ListAccountStorageCredentialsResponse: + type: object + properties: + storage_credentials: + type: array + items: + $ref: '#/components/schemas/StorageCredentialInfo' + description: An array of metastore storage credentials. + description: The metastore storage credentials were successfully returned. + ListCatalogsResponse: + type: object + properties: + catalogs: + type: array + items: + $ref: '#/components/schemas/CatalogInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListConnectionsResponse: + type: object + properties: + connections: + type: array + items: + $ref: '#/components/schemas/ConnectionInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListCredentialsResponse: + type: object + properties: + credentials: + type: array + items: + $ref: '#/components/schemas/CredentialInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListEntityTagAssignmentsResponse: + type: object + properties: + next_page_token: + type: string + tag_assignments: + type: array + items: + $ref: '#/components/schemas/EntityTagAssignment' + description: The list of tag assignments + ListExternalLineageRelationshipsResponse: + type: object + properties: + external_lineage_relationships: + type: array + items: + $ref: '#/components/schemas/ExternalLineageInfo' + next_page_token: + type: string + ListExternalLocationsResponse: + type: object + properties: + external_locations: + type: array + items: + $ref: '#/components/schemas/ExternalLocationInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListExternalMetadataResponse: + type: object + properties: + external_metadata: + type: array + items: + $ref: '#/components/schemas/ExternalMetadata' + next_page_token: + type: string + ListFunctionsResponse: + type: object + properties: + functions: + type: array + items: + $ref: '#/components/schemas/FunctionInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListMetastoresResponse: + type: object + properties: + metastores: + type: array + items: + $ref: '#/components/schemas/MetastoreInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListModelVersionsResponse: + type: object + properties: + model_versions: + type: array + items: + $ref: '#/components/schemas/ModelVersionInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListPoliciesResponse: + type: object + properties: + next_page_token: + type: string + policies: + type: array + items: + $ref: '#/components/schemas/PolicyInfo' + description: The list of retrieved policies. + ListQuotasResponse: + type: object + properties: + next_page_token: + type: string + quotas: + type: array + items: + $ref: '#/components/schemas/QuotaInfo' + description: An array of returned QuotaInfos. + ListRegisteredModelsResponse: + type: object + properties: + next_page_token: + type: string + registered_models: + type: array + items: + $ref: '#/components/schemas/RegisteredModelInfo' + ListSchemasResponse: + type: object + properties: + next_page_token: + type: string + schemas: + type: array + items: + $ref: '#/components/schemas/SchemaInfo' + description: An array of schema information objects. + ListStorageCredentialsResponse: + type: object + properties: + next_page_token: + type: string + storage_credentials: + type: array + items: + $ref: '#/components/schemas/StorageCredentialInfo' + ListSystemSchemasResponse: + type: object + properties: + next_page_token: + type: string + schemas: + type: array + items: + $ref: '#/components/schemas/SystemSchemaInfo' + description: An array of system schema information objects. + ListTableSummariesResponse: + type: object + properties: + next_page_token: + type: string + tables: + type: array + items: + $ref: '#/components/schemas/TableSummary' + description: List of table summaries. + ListTablesResponse: + type: object + properties: + next_page_token: + type: string + tables: + type: array + items: + $ref: '#/components/schemas/TableInfo' + description: An array of table information objects. + ListVolumesResponseContent: + type: object + properties: + next_page_token: + type: string + volumes: + type: array + items: + $ref: '#/components/schemas/VolumeInfo' + MatchColumn: + type: object + properties: + alias: + type: string + condition: + type: string + description: The condition expression used to match a table column. + MetastoreAssignment: + type: object + properties: + workspace_id: + type: integer + metastore_id: + type: string + description: The unique ID of the metastore. + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field is + deprecated. Please use "Default Namespace API" to configure the + default catalog for a Databricks workspace. + required: + - workspace_id + - metastore_id + MetastoreInfo: + type: object + properties: + cloud: + type: string + created_at: + type: integer + description: Time at which this metastore was created, in epoch milliseconds. + created_by: + type: string + description: Username of metastore creator. + default_data_access_config_id: + type: string + description: >- + Unique identifier of the metastore's (Default) Data Access + Configuration. + delta_sharing_organization_name: + type: string + description: >- + The organization name of a Delta Sharing entity, to be used in + Databricks-to-Databricks Delta Sharing as the official name. + delta_sharing_recipient_token_lifetime_in_seconds: + type: integer + description: The lifetime of delta sharing recipient token in seconds. + delta_sharing_scope: + $ref: '#/components/schemas/DeltaSharingScopeEnum' + description: The scope of Delta Sharing enabled for the metastore. + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + global_metastore_id: + type: string + description: >- + Globally unique metastore ID across clouds and regions, of the form + `cloud:region:metastore_id`. + metastore_id: + type: string + description: Unique identifier of metastore. + name: + type: string + description: The user-specified name of the metastore. + owner: + type: string + description: The owner of the metastore. + privilege_model_version: + type: string + description: >- + Privilege model version of the metastore, of the form `major.minor` + (e.g., `1.0`). + region: + type: string + description: >- + Cloud region which the metastore serves (e.g., `us-west-2`, + `westus`). + storage_root: + type: string + description: The storage root URL for metastore + storage_root_credential_id: + type: string + description: UUID of storage credential to access the metastore storage_root. + storage_root_credential_name: + type: string + description: Name of the storage credential to access the metastore storage_root. + updated_at: + type: integer + description: >- + Time at which the metastore was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the metastore. + ModelVersionInfo: + type: object + properties: + aliases: + type: array + items: + $ref: '#/components/schemas/RegisteredModelAlias' + catalog_name: + type: string + description: The name of the catalog containing the model version + comment: + type: string + description: The comment attached to the model version + created_at: + type: integer + created_by: + type: string + description: The identifier of the user who created the model version + id: + type: string + description: The unique identifier of the model version + metastore_id: + type: string + description: The unique identifier of the metastore containing the model version + model_name: + type: string + description: >- + The name of the parent registered model of the model version, + relative to parent schema + model_version_dependencies: + $ref: '#/components/schemas/DependencyList' + description: Model version dependencies, for feature-store packaged models + run_id: + type: string + description: >- + MLflow run ID used when creating the model version, if ``source`` + was generated by an experiment run stored in an MLflow tracking + server + run_workspace_id: + type: integer + description: >- + ID of the Databricks workspace containing the MLflow run that + generated this model version, if applicable + schema_name: + type: string + description: >- + The name of the schema containing the model version, relative to + parent catalog + source: + type: string + description: >- + URI indicating the location of the source artifacts (files) for the + model version + status: + $ref: '#/components/schemas/ModelVersionInfoStatus' + description: >- + Current status of the model version. Newly created model versions + start in PENDING_REGISTRATION status, then move to READY status once + the model version files are uploaded and the model version is + finalized. Only model versions in READY status can be loaded for + inference or served. + storage_location: + type: string + description: >- + The storage location on the cloud under which model version data + files are stored + updated_at: + type: integer + updated_by: + type: string + description: The identifier of the user who updated the model version last time + version: + type: integer + description: >- + Integer model version number, used to reference the model version in + API requests. + MonitorCronSchedule: + type: object + properties: + quartz_cron_expression: + type: string + timezone_id: + type: string + description: >- + The timezone id (e.g., ``PST``) in which to evaluate the quartz + expression. + pause_status: + $ref: '#/components/schemas/MonitorCronSchedulePauseStatus' + description: Read only field that indicates whether a schedule is paused or not. + required: + - quartz_cron_expression + - timezone_id + MonitorDataClassificationConfig: + type: object + properties: + enabled: + type: boolean + description: Whether to enable data classification. + description: Data classification related configuration. + MonitorDestination: + type: object + properties: + email_addresses: + type: array + items: + type: string + MonitorInferenceLog: + type: object + properties: + problem_type: + $ref: '#/components/schemas/MonitorInferenceLogProblemType' + timestamp_col: + type: string + description: Column for the timestamp. + granularities: + type: array + items: + type: string + description: >- + List of granularities to use when aggregating data into time windows + based on their timestamp. + prediction_col: + type: string + description: Column for the prediction. + model_id_col: + type: string + description: Column for the model identifier. + label_col: + type: string + description: Column for the label. + prediction_proba_col: + type: string + description: Column for prediction probabilities + required: + - problem_type + - timestamp_col + - granularities + - prediction_col + - model_id_col + MonitorInfo: + type: object + properties: + output_schema_name: + type: string + table_name: + type: string + description: >- + [Create:ERR Update:IGN] UC table to monitor. Format: + `catalog.schema.table_name` + status: + $ref: '#/components/schemas/MonitorInfoStatus' + description: '[Create:ERR Update:IGN] The monitor status.' + profile_metrics_table_name: + type: string + description: >- + [Create:ERR Update:IGN] Table that stores profile metrics data. + Format: `catalog.schema.table_name`. + drift_metrics_table_name: + type: string + description: >- + [Create:ERR Update:IGN] Table that stores drift metrics data. + Format: `catalog.schema.table_name`. + monitor_version: + type: integer + description: >- + [Create:ERR Update:IGN] Represents the current monitor configuration + version in use. The version will be represented in a numeric fashion + (1,2,3...). The field has flexibility to take on negative values, + which can indicate corrupted monitor_version numbers. + assets_dir: + type: string + description: >- + [Create:REQ Update:IGN] Field for specifying the absolute path to a + custom directory to store data-monitoring assets. Normally + prepopulated to a default user location via UI and Python APIs. + baseline_table_name: + type: string + description: >- + [Create:OPT Update:OPT] Baseline table name. Baseline data is used + to compute drift from the data in the monitored `table_name`. The + baseline table and the monitored table shall have the same schema. + custom_metrics: + type: array + items: + $ref: '#/components/schemas/MonitorMetric' + description: '[Create:OPT Update:OPT] Custom metrics.' + dashboard_id: + type: string + description: >- + [Create:ERR Update:OPT] Id of dashboard that visualizes the computed + metrics. This can be empty if the monitor is in PENDING state. + data_classification_config: + $ref: '#/components/schemas/MonitorDataClassificationConfig' + description: '[Create:OPT Update:OPT] Data classification related config.' + inference_log: + $ref: '#/components/schemas/MonitorInferenceLog' + latest_monitor_failure_msg: + type: string + description: >- + [Create:ERR Update:IGN] The latest error message for a monitor + failure. + notifications: + $ref: '#/components/schemas/MonitorNotifications' + description: '[Create:OPT Update:OPT] Field for specifying notification settings.' + schedule: + $ref: '#/components/schemas/MonitorCronSchedule' + description: '[Create:OPT Update:OPT] The monitor schedule.' + slicing_exprs: + type: array + items: + type: string + description: >- + [Create:OPT Update:OPT] List of column expressions to slice data + with for targeted analysis. The data is grouped by each expression + independently, resulting in a separate slice for each predicate and + its complements. For example `slicing_exprs=[“col_1”, “col_2 > 10”]` + will generate the following slices: two slices for `col_2 > 10` + (True and False), and one slice per unique value in `col1`. For + high-cardinality columns, only the top 100 unique values by + frequency will generate slices. + snapshot: + $ref: '#/components/schemas/MonitorSnapshot' + description: Configuration for monitoring snapshot tables. + time_series: + $ref: '#/components/schemas/MonitorTimeSeries' + description: Configuration for monitoring time series tables. + required: + - output_schema_name + - table_name + - status + - profile_metrics_table_name + - drift_metrics_table_name + - monitor_version + MonitorMetric: + type: object + properties: + name: + type: string + description: Name of the metric in the output tables. + definition: + type: string + description: >- + Jinja template for a SQL expression that specifies how to compute + the metric. See [create metric definition]. [create metric + definition]: + https://docs.databricks.com/en/lakehouse-monitoring/custom-metrics.html#create-definition + input_columns: + type: array + items: + type: string + description: >- + A list of column names in the input table the metric should be + computed for. Can use ``":table"`` to indicate that the metric needs + information from multiple columns. + output_data_type: + type: string + description: The output type of the custom metric. + type: + $ref: '#/components/schemas/MonitorMetricType' + description: >- + Can only be one of ``"CUSTOM_METRIC_TYPE_AGGREGATE"``, + ``"CUSTOM_METRIC_TYPE_DERIVED"``, or ``"CUSTOM_METRIC_TYPE_DRIFT"``. + The ``"CUSTOM_METRIC_TYPE_AGGREGATE"`` and + ``"CUSTOM_METRIC_TYPE_DERIVED"`` metrics are computed on a single + table, whereas the ``"CUSTOM_METRIC_TYPE_DRIFT"`` compare metrics + across baseline and input table, or across the two consecutive time + windows. - CUSTOM_METRIC_TYPE_AGGREGATE: only depend on the existing + columns in your table - CUSTOM_METRIC_TYPE_DERIVED: depend on + previously computed aggregate metrics - CUSTOM_METRIC_TYPE_DRIFT: + depend on previously computed aggregate or derived metrics + required: + - name + - definition + - input_columns + - output_data_type + - type + description: Custom metric definition. + MonitorNotifications: + type: object + properties: + on_failure: + $ref: '#/components/schemas/MonitorDestination' + on_new_classification_tag_detected: + $ref: '#/components/schemas/MonitorDestination' + description: >- + Destinations to send notifications on new classification tag + detected. + MonitorRefreshInfo: + type: object + properties: + refresh_id: + type: integer + state: + $ref: '#/components/schemas/MonitorRefreshInfoState' + description: The current state of the refresh. + start_time_ms: + type: integer + description: >- + Time at which refresh operation was initiated (milliseconds since + 1/1/1970 UTC). + end_time_ms: + type: integer + description: >- + Time at which refresh operation completed (milliseconds since + 1/1/1970 UTC). + message: + type: string + description: >- + An optional message to give insight into the current state of the + job (e.g. FAILURE messages). + trigger: + $ref: '#/components/schemas/MonitorRefreshInfoTrigger' + description: The method by which the refresh was triggered. + required: + - refresh_id + - state + - start_time_ms + MonitorRefreshListResponse: + type: object + properties: + refreshes: + type: array + items: + $ref: '#/components/schemas/MonitorRefreshInfo' + MonitorSnapshot: + type: object + properties: {} + description: Snapshot analysis configuration + MonitorTimeSeries: + type: object + properties: + timestamp_col: + type: string + description: Column for the timestamp. + granularities: + type: array + items: + type: string + description: >- + Granularities for aggregating data into time windows based on their + timestamp. Currently the following static granularities are + supported: {``\"5 minutes\"``, ``\"30 minutes\"``, ``\"1 hour\"``, + ``\"1 day\"``, ``\"\u003cn\u003e week(s)\"``, ``\"1 month\"``, ``\"1 + year\"``}. + required: + - timestamp_col + - granularities + description: Time series analysis configuration. + NamedTableConstraint: + type: object + properties: + name: + type: string + required: + - name + NotificationDestination: + type: object + properties: + destination_id: + type: string + destination_type: + $ref: '#/components/schemas/DestinationType' + description: The type of the destination. + special_destination: + $ref: '#/components/schemas/SpecialDestination' + description: >- + This field is used to denote whether the destination is the email of + the owner of the securable object. The special destination cannot be + assigned to a securable and only represents the default destination + of the securable. The securable types that support default special + destinations are: "catalog", "external_location", "connection", + "credential", and "metastore". The **destination_type** of a + **special_destination** is always EMAIL. + OnlineTable: + type: object + properties: + name: + type: string + description: Full three-part (catalog, schema, table) name of the table. + spec: + $ref: '#/components/schemas/OnlineTableSpec' + description: Specification of the online table. + status: + $ref: '#/components/schemas/OnlineTableStatus' + description: Online Table data synchronization status + table_serving_url: + type: string + description: Data serving REST API URL for this table + unity_catalog_provisioning_state: + $ref: '#/components/schemas/ProvisioningInfoState' + description: >- + The provisioning state of the online table entity in Unity Catalog. + This is distinct from the state of the data synchronization pipeline + (i.e. the table may be in "ACTIVE" but the pipeline may be in + "PROVISIONING" as it runs asynchronously). + description: Online Table information. + OnlineTableSpec: + type: object + properties: + perform_full_copy: + type: boolean + description: >- + Whether to create a full-copy pipeline -- a pipeline that stops + after creates a full copy of the source table upon initialization + and does not process any change data feeds (CDFs) afterwards. The + pipeline can still be manually triggered afterwards, but it always + perform a full copy of the source table and there are no incremental + updates. This mode is useful for syncing views or tables without + CDFs to online tables. Note that the full-copy pipeline only + supports "triggered" scheduling policy. + pipeline_id: + type: string + description: >- + ID of the associated pipeline. Generated by the server - cannot be + set by the caller. + primary_key_columns: + type: array + items: + type: string + description: >- + Primary Key columns to be used for data insert/update in the + destination. + run_continuously: + $ref: '#/components/schemas/OnlineTableSpecContinuousSchedulingPolicy' + description: Pipeline runs continuously after generating the initial data. + run_triggered: + $ref: '#/components/schemas/OnlineTableSpecTriggeredSchedulingPolicy' + description: >- + Pipeline stops after generating the initial data and can be + triggered later (manually, through a cron job or through data + triggers) + source_table_full_name: + type: string + description: Three-part (catalog, schema, table) name of the source Delta table. + timeseries_key: + type: string + description: >- + Time series key to deduplicate (tie-break) rows with the same + primary key. + description: Specification of an online table. + OnlineTableSpecContinuousSchedulingPolicy: + type: object + properties: {} + OnlineTableSpecTriggeredSchedulingPolicy: + type: object + properties: {} + OnlineTableStatus: + type: object + properties: + continuous_update_status: + $ref: '#/components/schemas/ContinuousUpdateStatus' + detailed_state: + $ref: '#/components/schemas/OnlineTableState' + description: The state of the online table. + failed_status: + $ref: '#/components/schemas/FailedStatus' + message: + type: string + description: A text description of the current state of the online table. + provisioning_status: + $ref: '#/components/schemas/ProvisioningStatus' + triggered_update_status: + $ref: '#/components/schemas/TriggeredUpdateStatus' + description: Status of an online table. + OptionSpec: + type: object + properties: + allowed_values: + type: array + items: + type: string + description: >- + For drop down / radio button selections, UI will want to know the + possible input values, it can also be used by other option types to + limit input selections. + default_value: + type: string + description: >- + The default value of the option, for example, value '443' for 'port' + option. + description: + type: string + description: >- + A concise user facing description of what the input value of this + option should look like. + hint: + type: string + description: >- + The hint is used on the UI to suggest what the input value can + possibly be like, for example: example.com for 'host' option. Unlike + default value, it will not be applied automatically without user + input. + is_copiable: + type: boolean + description: >- + Indicates whether an option should be displayed with copy button on + the UI. + is_creatable: + type: boolean + description: >- + Indicates whether an option can be provided by users in the + create/update path of an entity. + is_hidden: + type: boolean + description: >- + Is the option value not user settable and is thus not shown on the + UI. + is_loggable: + type: boolean + description: >- + Specifies whether this option is safe to log, i.e. no sensitive + information. + is_required: + type: boolean + description: Is the option required. + is_secret: + type: boolean + description: Is the option value considered secret and thus redacted on the UI. + is_updatable: + type: boolean + description: Is the option updatable by users. + name: + type: string + description: The unique name of the option. + oauth_stage: + $ref: '#/components/schemas/OptionSpecOauthStage' + description: >- + Specifies when the option value is displayed on the UI within the + OAuth flow. + type: + $ref: '#/components/schemas/OptionSpecOptionType' + description: The type of the option. + description: >- + Spec of an allowed option on a securable kind and its attributes. This + is mostly used by UI to + provide user friendly hints and descriptions in order to facilitate the securable creation + process. + PermissionsChange: + type: object + properties: + add: + type: array + items: + $ref: '#/components/schemas/Privilege' + principal: + type: string + description: >- + The principal whose privileges we are changing. Only one of + principal or principal_id should be specified, never both at the + same time. + remove: + type: array + items: + $ref: '#/components/schemas/Privilege' + description: The set of privileges to remove. + PipelineProgress: + type: object + properties: + estimated_completion_time_seconds: + type: number + description: The estimated time remaining to complete this update in seconds. + latest_version_currently_processing: + type: integer + description: >- + The source table Delta version that was last processed by the + pipeline. The pipeline may not have completely processed this + version yet. + sync_progress_completion: + type: number + description: >- + The completion ratio of this update. This is a number between 0 and + 1. + synced_row_count: + type: integer + description: The number of rows that have been synced in this update. + total_row_count: + type: integer + description: >- + The total number of rows that need to be synced in this update. This + number may be an estimate. + description: Progress information of the Online Table data synchronization pipeline. + PolicyInfo: + type: object + properties: + to_principals: + type: array + items: + type: string + for_securable_type: + $ref: '#/components/schemas/SecurableType' + description: >- + Type of securables that the policy should take effect on. Only + `TABLE` is supported at this moment. Required on create and optional + on update. + policy_type: + $ref: '#/components/schemas/PolicyType' + description: Type of the policy. Required on create. + column_mask: + $ref: '#/components/schemas/ColumnMaskOptions' + description: >- + Options for column mask policies. Valid only if `policy_type` is + `POLICY_TYPE_COLUMN_MASK`. Required on create and optional on + update. When specified on update, the new options will replace the + existing options as a whole. + comment: + type: string + description: Optional description of the policy. + created_at: + type: integer + description: >- + Time at which the policy was created, in epoch milliseconds. Output + only. + created_by: + type: string + description: Username of the user who created the policy. Output only. + except_principals: + type: array + items: + type: string + description: >- + Optional list of user or group names that should be excluded from + the policy. + id: + type: string + description: >- + Unique identifier of the policy. This field is output only and is + generated by the system. + match_columns: + type: array + items: + $ref: '#/components/schemas/MatchColumn' + description: >- + Optional list of condition expressions used to match table columns. + Only valid when `for_securable_type` is `TABLE`. When specified, the + policy only applies to tables whose columns satisfy all match + conditions. + name: + type: string + description: >- + Name of the policy. Required on create and optional on update. To + rename the policy, set `name` to a different value on update. + on_securable_fullname: + type: string + description: >- + Full name of the securable on which the policy is defined. Required + on create. + on_securable_type: + $ref: '#/components/schemas/SecurableType' + description: >- + Type of the securable on which the policy is defined. Only + `CATALOG`, `SCHEMA` and `TABLE` are supported at this moment. + Required on create. + row_filter: + $ref: '#/components/schemas/RowFilterOptions' + description: >- + Options for row filter policies. Valid only if `policy_type` is + `POLICY_TYPE_ROW_FILTER`. Required on create and optional on update. + When specified on update, the new options will replace the existing + options as a whole. + updated_at: + type: integer + description: >- + Time at which the policy was last modified, in epoch milliseconds. + Output only. + updated_by: + type: string + description: Username of the user who last modified the policy. Output only. + when_condition: + type: string + description: Optional condition when the policy should take effect. + required: + - to_principals + - for_securable_type + - policy_type + PrimaryKeyConstraint: + type: object + properties: + name: + type: string + child_columns: + type: array + items: + type: string + description: Column names for this constraint. + rely: + type: boolean + description: True if the constraint is RELY, false or unset if NORELY. + timeseries_columns: + type: array + items: + type: string + description: Column names that represent a timeseries. + required: + - name + - child_columns + Principal: + type: object + properties: + id: + type: string + principal_type: + $ref: '#/components/schemas/PrincipalType' + PrivilegeAssignment: + type: object + properties: + principal: + type: string + privileges: + type: array + items: + $ref: '#/components/schemas/Privilege' + description: The privileges assigned to the principal. + ProvisioningInfo: + type: object + properties: + state: + $ref: '#/components/schemas/ProvisioningInfoState' + description: The provisioning state of the resource. + description: Status of an asynchronously provisioned resource. + ProvisioningStatus: + type: object + properties: + initial_pipeline_sync_progress: + $ref: '#/components/schemas/PipelineProgress' + description: >- + Details about initial data synchronization. Only populated when in + the PROVISIONING_INITIAL_SNAPSHOT state. + description: |- + Detailed status of an online table. Shown if the online table is in the + PROVISIONING_PIPELINE_RESOURCES or the PROVISIONING_INITIAL_SNAPSHOT state. + QuotaInfo: + type: object + properties: + last_refreshed_at: + type: integer + parent_full_name: + type: string + description: >- + Name of the parent resource. Returns metastore ID if the parent is a + metastore. + parent_securable_type: + $ref: '#/components/schemas/SecurableType' + description: The quota parent securable type. + quota_count: + type: integer + description: The current usage of the resource quota. + quota_limit: + type: integer + description: The current limit of the resource quota. + quota_name: + type: string + description: The name of the quota. + R2Credentials: + type: object + properties: + access_key_id: + type: string + description: The access key ID that identifies the temporary credentials. + secret_access_key: + type: string + description: The secret access key associated with the access key. + session_token: + type: string + description: >- + The generated JWT that users must pass to use the temporary + credentials. + description: |- + R2 temporary credentials for API authentication. Read more at + https://developers.cloudflare.com/r2/api/s3/tokens/. + RegenerateDashboardResponse: + type: object + properties: + dashboard_id: + type: string + parent_folder: + type: string + description: Parent folder is equivalent to {assets_dir}/{tableName} + RegisteredModelAlias: + type: object + properties: + alias_name: + type: string + catalog_name: + type: string + description: The name of the catalog containing the model version + id: + type: string + description: The unique identifier of the alias + model_name: + type: string + description: >- + The name of the parent registered model of the model version, + relative to parent schema + schema_name: + type: string + description: >- + The name of the schema containing the model version, relative to + parent catalog + version_num: + type: integer + description: >- + Integer version number of the model version to which this alias + points. + RegisteredModelInfo: + type: object + properties: + aliases: + type: array + items: + $ref: '#/components/schemas/RegisteredModelAlias' + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: >- + The name of the catalog where the schema and the registered model + reside + comment: + type: string + description: The comment attached to the registered model + created_at: + type: integer + description: >- + Creation timestamp of the registered model in milliseconds since the + Unix epoch + created_by: + type: string + description: The identifier of the user who created the registered model + full_name: + type: string + description: The three-level (fully qualified) name of the registered model + metastore_id: + type: string + description: The unique identifier of the metastore + name: + type: string + description: The name of the registered model + owner: + type: string + description: The identifier of the user who owns the registered model + schema_name: + type: string + description: The name of the schema where the registered model resides + storage_location: + type: string + description: >- + The storage location on the cloud under which model version data + files are stored + updated_at: + type: integer + description: >- + Last-update timestamp of the registered model in milliseconds since + the Unix epoch + updated_by: + type: string + description: >- + The identifier of the user who updated the registered model last + time + RowFilterOptions: + type: object + properties: + function_name: + type: string + using: + type: array + items: + $ref: '#/components/schemas/FunctionArgument' + description: >- + Optional list of column aliases or constant literals to be passed as + arguments to the row filter function. The type of each column should + match the positional argument of the row filter function. + required: + - function_name + SchemaInfo: + type: object + properties: + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: Name of parent catalog. + catalog_type: + $ref: '#/components/schemas/CatalogType' + description: The type of the parent catalog. + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: Time at which this schema was created, in epoch milliseconds. + created_by: + type: string + description: Username of schema creator. + effective_predictive_optimization_flag: + $ref: '#/components/schemas/EffectivePredictiveOptimizationFlag' + enable_predictive_optimization: + $ref: '#/components/schemas/EnablePredictiveOptimization' + description: >- + Whether predictive optimization should be enabled for this object + and objects under it. + full_name: + type: string + description: Full name of schema, in form of __catalog_name__.__schema_name__. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of schema, relative to parent catalog. + owner: + type: string + description: Username of current owner of schema. + properties: + type: object + description: A map of key-value properties attached to the securable. + schema_id: + type: string + description: The unique identifier of the schema. + storage_location: + type: string + description: Storage location for managed tables within schema. + storage_root: + type: string + description: Storage root URL for managed tables within schema. + updated_at: + type: integer + description: Time at which this schema was created, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified schema. + description: 'Next ID: 45' + Securable: + type: object + properties: + full_name: + type: string + description: >- + Required. The full name of the catalog/schema/table. Optional if + resource_name is present. + provider_share: + type: string + description: >- + Optional. The name of the Share object that contains the securable + when the securable is getting shared in D2D Delta Sharing. + type: + $ref: '#/components/schemas/SecurableType' + description: >- + Required. The type of securable (catalog/schema/table). Optional if + resource_name is present. + description: >- + Generic definition of a securable, which is uniquely defined in a + metastore by its type and full + name. + SecurableKindManifest: + type: object + properties: + assignable_privileges: + type: array + items: + type: string + description: Privileges that can be assigned to the securable. + capabilities: + type: array + items: + type: string + description: A list of capabilities in the securable kind. + options: + type: array + items: + $ref: '#/components/schemas/OptionSpec' + description: Detailed specs of allowed options. + securable_kind: + $ref: '#/components/schemas/SecurableKind' + description: Securable kind to get manifest of. + securable_type: + $ref: '#/components/schemas/SecurableType' + description: Securable Type of the kind. + description: Manifest of a specific securable kind. + SecurablePermissions: + type: object + properties: + permissions: + type: array + items: + type: string + securable: + $ref: '#/components/schemas/Securable' + description: >- + The securable for which the access request destinations are being + requested. + SseEncryptionDetails: + type: object + properties: + algorithm: + $ref: '#/components/schemas/SseEncryptionDetailsAlgorithm' + description: >- + Sets the value of the 'x-amz-server-side-encryption' header in S3 + request. + aws_kms_key_arn: + type: string + description: >- + Optional. The ARN of the SSE-KMS key used with the S3 location, when + algorithm = "SSE-KMS". Sets the value of the + 'x-amz-server-side-encryption-aws-kms-key-id' header. + description: Server-Side Encryption properties for clients communicating with AWS s3. + StorageCredentialInfo: + type: object + properties: + aws_iam_role: + $ref: '#/components/schemas/AwsIamRoleResponse' + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentityResponse' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + cloudflare_api_token: + $ref: '#/components/schemas/CloudflareApiToken' + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + created_at: + type: integer + description: Time at which this credential was created, in epoch milliseconds. + created_by: + type: string + description: Username of credential creator. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccountResponse' + description: The Databricks managed GCP service account configuration. + full_name: + type: string + description: The full name of the credential. + id: + type: string + description: The unique identifier of the credential. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + metastore_id: + type: string + description: Unique identifier of the parent metastore. + name: + type: string + description: >- + The credential name. The name must be unique among storage and + service credentials within the metastore. + owner: + type: string + description: Username of current owner of credential. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + updated_at: + type: integer + description: >- + Time at which this credential was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the credential. + used_for_managed_storage: + type: boolean + description: >- + Whether this credential is the current metastore's root storage + credential. Only applicable when purpose is **STORAGE**. + SystemSchemaInfo: + type: object + properties: + schema: + type: string + state: + type: string + description: >- + The current state of enablement for the system schema. An empty + string means the system schema is available and ready for opt-in. + Possible values: AVAILABLE | ENABLE_INITIALIZED | ENABLE_COMPLETED | + DISABLE_INITIALIZED | UNAVAILABLE | MANAGED + required: + - schema + - state + TableConstraint: + type: object + properties: + foreign_key_constraint: + $ref: '#/components/schemas/ForeignKeyConstraint' + named_table_constraint: + $ref: '#/components/schemas/NamedTableConstraint' + primary_key_constraint: + $ref: '#/components/schemas/PrimaryKeyConstraint' + description: >- + A table constraint, as defined by *one* of the following fields being + set: + __primary_key_constraint__, __foreign_key_constraint__, __named_table_constraint__. + TableDependency: + type: object + properties: + table_full_name: + type: string + description: >- + Full name of the dependent table, in the form of + __catalog_name__.__schema_name__.__table_name__. + required: + - table_full_name + description: A table that is dependent on a SQL object. + TableExistsResponse: + type: object + properties: + table_exists: + type: boolean + TableInfo: + type: object + properties: + access_point: + type: string + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: Name of parent catalog. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnInfo' + description: The array of __ColumnInfo__ definitions of the table's columns. + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: Time at which this table was created, in epoch milliseconds. + created_by: + type: string + description: Username of table creator. + data_access_configuration_id: + type: string + description: >- + Unique ID of the Data Access Configuration to use with the table + data. + data_source_format: + $ref: '#/components/schemas/DataSourceFormat' + deleted_at: + type: integer + description: >- + Time at which this table was deleted, in epoch milliseconds. Field + is omitted if table is not deleted. + delta_runtime_properties_kvpairs: + $ref: '#/components/schemas/DeltaRuntimePropertiesKvPairs' + description: Information pertaining to current state of the delta table. + effective_predictive_optimization_flag: + $ref: '#/components/schemas/EffectivePredictiveOptimizationFlag' + enable_predictive_optimization: + $ref: '#/components/schemas/EnablePredictiveOptimization' + encryption_details: + $ref: '#/components/schemas/EncryptionDetails' + full_name: + type: string + description: >- + Full name of table, in form of + __catalog_name__.__schema_name__.__table_name__ + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of table, relative to parent schema. + owner: + type: string + description: Username of current owner of table. + pipeline_id: + type: string + description: >- + The pipeline ID of the table. Applicable for tables created by + pipelines (Materialized View, Streaming Table, etc.). + properties: + type: object + description: A map of key-value properties attached to the securable. + row_filter: + $ref: '#/components/schemas/TableRowFilter' + schema_name: + type: string + description: Name of parent schema relative to its parent catalog. + securable_kind_manifest: + $ref: '#/components/schemas/SecurableKindManifest' + description: >- + SecurableKindManifest of table, including capabilities the table + has. + sql_path: + type: string + description: >- + List of schemes whose objects can be referenced without + qualification. + storage_credential_name: + type: string + description: >- + Name of the storage credential, when a storage credential is + configured for use with this table. + storage_location: + type: string + description: Storage root URL for table (for **MANAGED**, **EXTERNAL** tables). + table_constraints: + type: array + items: + $ref: '#/components/schemas/TableConstraint' + description: >- + List of table constraints. Note: this field is not set in the output + of the __listTables__ API. + table_id: + type: string + description: The unique identifier of the table. + table_type: + $ref: '#/components/schemas/TableType' + updated_at: + type: integer + description: Time at which this table was last modified, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified the table. + view_definition: + type: string + description: >- + View definition SQL (when __table_type__ is **VIEW**, + **MATERIALIZED_VIEW**, or **STREAMING_TABLE**) + view_dependencies: + $ref: '#/components/schemas/DependencyList' + description: >- + View dependencies (when table_type == **VIEW** or + **MATERIALIZED_VIEW**, **STREAMING_TABLE**) - when DependencyList is + None, the dependency is not provided; - when DependencyList is an + empty list, the dependency is provided but is empty; - when + DependencyList is not an empty list, dependencies are provided and + recorded. Note: this field is not set in the output of the + __listTables__ API. + TableRowFilter: + type: object + properties: + function_name: + type: string + input_column_names: + type: array + items: + type: string + description: >- + The list of table columns to be passed as input to the row filter + function. The column types should match the types of the filter + function arguments. + required: + - function_name + - input_column_names + TableSummary: + type: object + properties: + full_name: + type: string + securable_kind_manifest: + $ref: '#/components/schemas/SecurableKindManifest' + description: >- + SecurableKindManifest of table, including capabilities the table + has. + table_type: + $ref: '#/components/schemas/TableType' + TagKeyValue: + type: object + properties: + key: + type: string + value: + type: string + description: value of the tag associated with the key, could be optional + TemporaryCredentials: + type: object + properties: + aws_temp_credentials: + $ref: '#/components/schemas/AwsCredentials' + azure_aad: + $ref: '#/components/schemas/AzureActiveDirectoryToken' + expiration_time: + type: integer + description: >- + Server time when the credential will expire, in epoch milliseconds. + The API client is advised to cache the credential given this + expiration time. + gcp_oauth_token: + $ref: '#/components/schemas/GcpOauthToken' + TriggeredUpdateStatus: + type: object + properties: + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was synced to the online + table. Note that this Delta version may not be completely synced to + the online table yet. + timestamp: + type: string + description: >- + The timestamp of the last time any data was synchronized from the + source table to the online table. + triggered_update_progress: + $ref: '#/components/schemas/PipelineProgress' + description: Progress of the active data synchronization pipeline. + description: >- + Detailed status of an online table. Shown if the online table is in the + ONLINE_TRIGGERED_UPDATE + or the ONLINE_NO_PENDING_UPDATE state. + UnassignResponse: + type: object + properties: {} + UpdateAccountsMetastore: + type: object + properties: + delta_sharing_organization_name: + type: string + delta_sharing_recipient_token_lifetime_in_seconds: + type: integer + description: The lifetime of delta sharing recipient token in seconds. + delta_sharing_scope: + $ref: '#/components/schemas/DeltaSharingScopeEnum' + description: The scope of Delta Sharing enabled for the metastore. + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + owner: + type: string + description: The owner of the metastore. + privilege_model_version: + type: string + description: >- + Privilege model version of the metastore, of the form `major.minor` + (e.g., `1.0`). + storage_root_credential_id: + type: string + description: UUID of storage credential to access the metastore storage_root. + UpdateAccountsStorageCredential: + type: object + properties: + aws_iam_role: + $ref: '#/components/schemas/AwsIamRoleRequest' + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentityResponse' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + cloudflare_api_token: + $ref: '#/components/schemas/CloudflareApiToken' + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccountRequest' + description: The Databricks managed GCP service account configuration. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + owner: + type: string + description: Username of current owner of credential. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + UpdateAssignmentResponse: + type: object + properties: {} + UpdateCatalogWorkspaceBindingsResponse: + type: object + properties: + workspaces: + type: array + items: + type: integer + UpdateMetastoreAssignment: + type: object + properties: + workspace_id: + type: integer + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field is + deprecated. Please use "Default Namespace API" to configure the + default catalog for a Databricks workspace. + metastore_id: + type: string + description: The unique ID of the metastore. + required: + - workspace_id + UpdatePermissionsResponse: + type: object + properties: + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/PrivilegeAssignment' + UpdateRequestExternalLineage: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + UpdateResponse: + type: object + properties: {} + UpdateWorkspaceBindingsResponse: + type: object + properties: + bindings: + type: array + items: + $ref: '#/components/schemas/WorkspaceBinding' + description: List of workspace bindings. + description: A list of workspace IDs that are bound to the securable + ValidateCredentialResponse: + type: object + properties: + isDir: + type: boolean + results: + type: array + items: + $ref: '#/components/schemas/CredentialValidationResult' + description: The results of the validation check. + ValidateStorageCredentialResponse: + type: object + properties: + isDir: + type: boolean + results: + type: array + items: + $ref: '#/components/schemas/ValidationResult' + description: The results of the validation check. + ValidationResult: + type: object + properties: + message: + type: string + operation: + $ref: '#/components/schemas/ValidationResultOperation' + description: The operation tested. + result: + $ref: '#/components/schemas/ValidationResultResult' + description: The results of the tested operation. + VolumeInfo: + type: object + properties: + access_point: + type: string + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: The name of the catalog where the schema and the volume are + comment: + type: string + description: The comment attached to the volume + created_at: + type: integer + created_by: + type: string + description: The identifier of the user who created the volume + encryption_details: + $ref: '#/components/schemas/EncryptionDetails' + full_name: + type: string + description: The three-level (fully qualified) name of the volume + metastore_id: + type: string + description: The unique identifier of the metastore + name: + type: string + description: The name of the volume + owner: + type: string + description: The identifier of the user who owns the volume + schema_name: + type: string + description: The name of the schema where the volume is + storage_location: + type: string + description: The storage location on the cloud + updated_at: + type: integer + updated_by: + type: string + description: The identifier of the user who updated the volume last time + volume_id: + type: string + description: The unique identifier of the volume + volume_type: + $ref: '#/components/schemas/VolumeType' + description: >- + The type of the volume. An external volume is located in the + specified external location. A managed volume is located in the + default location which is specified by the parent schema, or the + parent catalog, or the Metastore. [Learn more] [Learn more]: + https://docs.databricks.com/aws/en/volumes/managed-vs-external + WorkspaceBinding: + type: object + properties: + workspace_id: + type: integer + binding_type: + $ref: '#/components/schemas/WorkspaceBindingBindingType' + description: One of READ_WRITE/READ_ONLY. Default is READ_WRITE. + required: + - workspace_id + ArtifactType: + type: string + enum: + - INIT_SCRIPT + - LIBRARY_JAR + - LIBRARY_MAVEN + description: The artifact type + CatalogIsolationMode: + type: string + enum: + - ISOLATED + - OPEN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CatalogType: + type: string + enum: + - DELTASHARING_CATALOG + - FOREIGN_CATALOG + - INTERNAL_CATALOG + - MANAGED_CATALOG + - MANAGED_ONLINE_CATALOG + - SYSTEM_CATALOG + description: The type of the catalog. + ColumnTypeName: + type: string + enum: + - ARRAY + - BINARY + - BOOLEAN + - BYTE + - CHAR + - DATE + - DECIMAL + - DOUBLE + - FLOAT + - GEOGRAPHY + - GEOMETRY + - INT + - INTERVAL + - LONG + - MAP + - 'NULL' + - SHORT + - STRING + - STRUCT + - TABLE_TYPE + - TIMESTAMP + - TIMESTAMP_NTZ + - USER_DEFINED_TYPE + - VARIANT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ConnectionType: + type: string + enum: + - BIGQUERY + - DATABRICKS + - GA4_RAW_DATA + - GLUE + - HIVE_METASTORE + - HTTP + - MYSQL + - ORACLE + - POSTGRESQL + - POWER_BI + - REDSHIFT + - SALESFORCE + - SALESFORCE_DATA_CLOUD + - SERVICENOW + - SNOWFLAKE + - SQLDW + - SQLSERVER + - TERADATA + - UNKNOWN_CONNECTION_TYPE + - WORKDAY_RAAS + description: 'Next Id: 72' + CreateFunctionParameterStyle: + type: string + enum: + - S + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CreateFunctionRoutineBody: + type: string + enum: + - EXTERNAL + - SQL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CreateFunctionSecurityType: + type: string + enum: + - DEFINER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CreateFunctionSqlDataAccess: + type: string + enum: + - CONTAINS_SQL + - NO_SQL + - READS_SQL_DATA + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CredentialPurpose: + type: string + enum: + - SERVICE + - STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CredentialType: + type: string + enum: + - ANY_STATIC_CREDENTIAL + - BEARER_TOKEN + - OAUTH_ACCESS_TOKEN + - OAUTH_M2M + - OAUTH_MTLS + - OAUTH_REFRESH_TOKEN + - OAUTH_RESOURCE_OWNER_PASSWORD + - OAUTH_U2M + - OAUTH_U2M_MAPPING + - OIDC_TOKEN + - PEM_PRIVATE_KEY + - SERVICE_CREDENTIAL + - SSWS_TOKEN + - UNKNOWN_CREDENTIAL_TYPE + - USERNAME_PASSWORD + description: 'Next Id: 15' + DataSourceFormat: + type: string + enum: + - AVRO + - BIGQUERY_FORMAT + - CSV + - DATABRICKS_FORMAT + - DATABRICKS_ROW_STORE_FORMAT + - DELTA + - DELTASHARING + - DELTA_UNIFORM_HUDI + - DELTA_UNIFORM_ICEBERG + - HIVE + - ICEBERG + - JSON + - MONGODB_FORMAT + - MYSQL_FORMAT + - NETSUITE_FORMAT + - ORACLE_FORMAT + - ORC + - PARQUET + - POSTGRESQL_FORMAT + - REDSHIFT_FORMAT + - SALESFORCE_DATA_CLOUD_FORMAT + - SALESFORCE_FORMAT + - SNOWFLAKE_FORMAT + - SQLDW_FORMAT + - SQLSERVER_FORMAT + - TERADATA_FORMAT + - TEXT + - UNITY_CATALOG + - VECTOR_INDEX_FORMAT + - WORKDAY_RAAS_FORMAT + description: Data source format + DeltaSharingScopeEnum: + type: string + enum: + - INTERNAL + - INTERNAL_AND_EXTERNAL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DestinationType: + type: string + enum: + - EMAIL + - GENERIC_WEBHOOK + - MICROSOFT_TEAMS + - SLACK + - URL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EffectivePredictiveOptimizationFlagInheritedFromType: + type: string + enum: + - CATALOG + - SCHEMA + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EnablePredictiveOptimization: + type: string + enum: + - DISABLE + - ENABLE + - INHERIT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoParameterStyle: + type: string + enum: + - S + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoRoutineBody: + type: string + enum: + - EXTERNAL + - SQL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoSecurityType: + type: string + enum: + - DEFINER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoSqlDataAccess: + type: string + enum: + - CONTAINS_SQL + - NO_SQL + - READS_SQL_DATA + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionParameterMode: + type: string + enum: + - IN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionParameterType: + type: string + enum: + - COLUMN + - PARAM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + IsolationMode: + type: string + enum: + - ISOLATION_MODE_ISOLATED + - ISOLATION_MODE_OPEN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + LineageDirection: + type: string + enum: + - DOWNSTREAM + - UPSTREAM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MatchType: + type: string + enum: + - PREFIX_MATCH + description: The artifact pattern matching type + ModelVersionInfoStatus: + type: string + enum: + - FAILED_REGISTRATION + - MODEL_VERSION_STATUS_UNKNOWN + - PENDING_REGISTRATION + - READY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MonitorCronSchedulePauseStatus: + type: string + enum: + - PAUSED + - UNPAUSED + - UNSPECIFIED + description: >- + Source link: + + https://src.dev.databricks.com/databricks/universe/-/blob/elastic-spark-common/api/messages/schedule.proto + + Monitoring workflow schedule pause status. + MonitorInferenceLogProblemType: + type: string + enum: + - PROBLEM_TYPE_CLASSIFICATION + - PROBLEM_TYPE_REGRESSION + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MonitorInfoStatus: + type: string + enum: + - MONITOR_STATUS_ACTIVE + - MONITOR_STATUS_DELETE_PENDING + - MONITOR_STATUS_ERROR + - MONITOR_STATUS_FAILED + - MONITOR_STATUS_PENDING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MonitorMetricType: + type: string + enum: + - CUSTOM_METRIC_TYPE_AGGREGATE + - CUSTOM_METRIC_TYPE_DERIVED + - CUSTOM_METRIC_TYPE_DRIFT + description: >- + Can only be one of ``"CUSTOM_METRIC_TYPE_AGGREGATE"``, + ``"CUSTOM_METRIC_TYPE_DERIVED"``, or + + ``"CUSTOM_METRIC_TYPE_DRIFT"``. The ``"CUSTOM_METRIC_TYPE_AGGREGATE"`` + and + + ``"CUSTOM_METRIC_TYPE_DERIVED"`` metrics are computed on a single table, + whereas the + + ``"CUSTOM_METRIC_TYPE_DRIFT"`` compare metrics across baseline and input + table, or across the + + two consecutive time windows. - CUSTOM_METRIC_TYPE_AGGREGATE: only + depend on the existing + + columns in your table - CUSTOM_METRIC_TYPE_DERIVED: depend on previously + computed aggregate + + metrics - CUSTOM_METRIC_TYPE_DRIFT: depend on previously computed + aggregate or derived metrics + MonitorRefreshInfoState: + type: string + enum: + - CANCELED + - FAILED + - PENDING + - RUNNING + - SUCCESS + - UNKNOWN + description: The current state of the refresh. + MonitorRefreshInfoTrigger: + type: string + enum: + - MANUAL + - SCHEDULE + - UNKNOWN_TRIGGER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + OnlineTableState: + type: string + enum: + - OFFLINE + - OFFLINE_FAILED + - ONLINE + - ONLINE_CONTINUOUS_UPDATE + - ONLINE_NO_PENDING_UPDATE + - ONLINE_PIPELINE_FAILED + - ONLINE_TRIGGERED_UPDATE + - ONLINE_UPDATING_PIPELINE_RESOURCES + - PROVISIONING + - PROVISIONING_INITIAL_SNAPSHOT + - PROVISIONING_PIPELINE_RESOURCES + description: The state of an online table. + OptionSpecOauthStage: + type: string + enum: + - BEFORE_ACCESS_TOKEN + - BEFORE_AUTHORIZATION_CODE + description: >- + During the OAuth flow, specifies which stage the option should be + displayed in the UI. + + OAUTH_STAGE_UNSPECIFIED is the default value for options unrelated to + the OAuth flow. + + BEFORE_AUTHORIZATION_CODE corresponds to options necessary to initiate + the OAuth process. + + BEFORE_ACCESS_TOKEN corresponds to options that are necessary to create + a foreign connection, + + but that should be displayed after the authorization code has already + been received. + OptionSpecOptionType: + type: string + enum: + - OPTION_BIGINT + - OPTION_BOOLEAN + - OPTION_ENUM + - OPTION_MULTILINE_STRING + - OPTION_NUMBER + - OPTION_SERVICE_CREDENTIAL + - OPTION_STRING + description: >- + Type of the option, we purposely follow JavaScript types so that the UI + can map the options to + + JS types. https://www.w3schools.com/js/js_datatypes.asp Enum is a + special case that it's just + + string with selections. + PathOperation: + type: string + enum: + - PATH_CREATE_TABLE + - PATH_READ + - PATH_READ_WRITE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PolicyType: + type: string + enum: + - POLICY_TYPE_COLUMN_MASK + - POLICY_TYPE_ROW_FILTER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PrincipalType: + type: string + enum: + - GROUP_PRINCIPAL + - SERVICE_PRINCIPAL + - USER_PRINCIPAL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + Privilege: + type: string + enum: + - ACCESS + - ALL_PRIVILEGES + - APPLY_TAG + - BROWSE + - CREATE + - CREATE_CATALOG + - CREATE_CLEAN_ROOM + - CREATE_CONNECTION + - CREATE_EXTERNAL_LOCATION + - CREATE_EXTERNAL_TABLE + - CREATE_EXTERNAL_VOLUME + - CREATE_FOREIGN_CATALOG + - CREATE_FOREIGN_SECURABLE + - CREATE_FUNCTION + - CREATE_MANAGED_STORAGE + - CREATE_MATERIALIZED_VIEW + - CREATE_MODEL + - CREATE_PROVIDER + - CREATE_RECIPIENT + - CREATE_SCHEMA + - CREATE_SERVICE_CREDENTIAL + - CREATE_SHARE + - CREATE_STORAGE_CREDENTIAL + - CREATE_TABLE + - CREATE_VIEW + - CREATE_VOLUME + - EXECUTE + - EXECUTE_CLEAN_ROOM_TASK + - EXTERNAL_USE_SCHEMA + - MANAGE + - MANAGE_ALLOWLIST + - MODIFY + - MODIFY_CLEAN_ROOM + - READ_FILES + - READ_PRIVATE_FILES + - READ_VOLUME + - REFRESH + - SELECT + - SET_SHARE_PERMISSION + - USAGE + - USE_CATALOG + - USE_CONNECTION + - USE_MARKETPLACE_ASSETS + - USE_PROVIDER + - USE_RECIPIENT + - USE_SCHEMA + - USE_SHARE + - WRITE_FILES + - WRITE_PRIVATE_FILES + - WRITE_VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ProvisioningInfoState: + type: string + enum: + - ACTIVE + - DEGRADED + - DELETING + - FAILED + - PROVISIONING + - UPDATING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SecurableKind: + type: string + enum: + - TABLE_DB_STORAGE + - TABLE_DELTA + - TABLE_DELTASHARING + - TABLE_DELTASHARING_MUTABLE + - TABLE_DELTASHARING_OPEN_DIR_BASED + - TABLE_DELTA_EXTERNAL + - TABLE_DELTA_ICEBERG_DELTASHARING + - TABLE_DELTA_ICEBERG_MANAGED + - TABLE_DELTA_UNIFORM_HUDI_EXTERNAL + - TABLE_DELTA_UNIFORM_ICEBERG_EXTERNAL + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_DELTASHARING + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_HIVE_METASTORE_EXTERNAL + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_HIVE_METASTORE_MANAGED + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_SNOWFLAKE + - TABLE_EXTERNAL + - TABLE_FEATURE_STORE + - TABLE_FEATURE_STORE_EXTERNAL + - TABLE_FOREIGN_BIGQUERY + - TABLE_FOREIGN_DATABRICKS + - TABLE_FOREIGN_DELTASHARING + - TABLE_FOREIGN_HIVE_METASTORE + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_SHALLOW_CLONE_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_SHALLOW_CLONE_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_VIEW + - TABLE_FOREIGN_HIVE_METASTORE_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_SHALLOW_CLONE_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_SHALLOW_CLONE_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_VIEW + - TABLE_FOREIGN_MONGODB + - TABLE_FOREIGN_MYSQL + - TABLE_FOREIGN_NETSUITE + - TABLE_FOREIGN_ORACLE + - TABLE_FOREIGN_POSTGRESQL + - TABLE_FOREIGN_REDSHIFT + - TABLE_FOREIGN_SALESFORCE + - TABLE_FOREIGN_SALESFORCE_DATA_CLOUD + - TABLE_FOREIGN_SALESFORCE_DATA_CLOUD_FILE_SHARING + - TABLE_FOREIGN_SALESFORCE_DATA_CLOUD_FILE_SHARING_VIEW + - TABLE_FOREIGN_SNOWFLAKE + - TABLE_FOREIGN_SQLDW + - TABLE_FOREIGN_SQLSERVER + - TABLE_FOREIGN_TERADATA + - TABLE_FOREIGN_WORKDAY_RAAS + - TABLE_ICEBERG_UNIFORM_MANAGED + - TABLE_INTERNAL + - TABLE_MANAGED_POSTGRESQL + - TABLE_MATERIALIZED_VIEW + - TABLE_MATERIALIZED_VIEW_DELTASHARING + - TABLE_METRIC_VIEW + - TABLE_METRIC_VIEW_DELTASHARING + - TABLE_ONLINE_VECTOR_INDEX_DIRECT + - TABLE_ONLINE_VECTOR_INDEX_REPLICA + - TABLE_ONLINE_VIEW + - TABLE_STANDARD + - TABLE_STREAMING_LIVE_TABLE + - TABLE_STREAMING_LIVE_TABLE_DELTASHARING + - TABLE_SYSTEM + - TABLE_SYSTEM_DELTASHARING + - TABLE_VIEW + - TABLE_VIEW_DELTASHARING + description: 'Latest kind: CONNECTION_OKTA_SYSTEM_LOGS_SSWS_TOKEN = 295; Next id: 296' + SecurableType: + type: string + enum: + - CATALOG + - CLEAN_ROOM + - CONNECTION + - CREDENTIAL + - EXTERNAL_LOCATION + - EXTERNAL_METADATA + - FUNCTION + - METASTORE + - PIPELINE + - PROVIDER + - RECIPIENT + - SCHEMA + - SHARE + - STAGING_TABLE + - STORAGE_CREDENTIAL + - TABLE + - VOLUME + description: The type of Unity Catalog securable. + SpecialDestination: + type: string + enum: + - SPECIAL_DESTINATION_CATALOG_OWNER + - SPECIAL_DESTINATION_CONNECTION_OWNER + - SPECIAL_DESTINATION_CREDENTIAL_OWNER + - SPECIAL_DESTINATION_EXTERNAL_LOCATION_OWNER + - SPECIAL_DESTINATION_METASTORE_OWNER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SseEncryptionDetailsAlgorithm: + type: string + enum: + - AWS_SSE_KMS + - AWS_SSE_S3 + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SystemType: + type: string + enum: + - AMAZON_REDSHIFT + - AZURE_SYNAPSE + - CONFLUENT + - DATABRICKS + - GOOGLE_BIGQUERY + - KAFKA + - LOOKER + - MICROSOFT_FABRIC + - MICROSOFT_SQL_SERVER + - MONGODB + - MYSQL + - ORACLE + - OTHER + - POSTGRESQL + - POWER_BI + - SALESFORCE + - SAP + - SERVICENOW + - SNOWFLAKE + - STREAM_NATIVE + - TABLEAU + - TERADATA + - WORKDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TableOperation: + type: string + enum: + - READ + - READ_WRITE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TableType: + type: string + enum: + - EXTERNAL + - EXTERNAL_SHALLOW_CLONE + - FOREIGN + - MANAGED + - MANAGED_SHALLOW_CLONE + - MATERIALIZED_VIEW + - METRIC_VIEW + - STREAMING_TABLE + - VIEW + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ValidateCredentialResult: + type: string + enum: + - FAIL + - PASS + - SKIP + description: A enum represents the result of the file operation + ValidationResultOperation: + type: string + enum: + - DELETE + - LIST + - PATH_EXISTS + - READ + - WRITE + description: >- + A enum represents the file operation performed on the external location + with the storage + + credential + ValidationResultResult: + type: string + enum: + - FAIL + - PASS + - SKIP + description: A enum represents the result of the file operation + VolumeType: + type: string + enum: + - EXTERNAL + - MANAGED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + WorkspaceBindingBindingType: + type: string + enum: + - BINDING_TYPE_READ_ONLY + - BINDING_TYPE_READ_WRITE + description: >- + Using `BINDING_TYPE_` prefix here to avoid conflict with + `TableOperation` enum in + + `credentials_common.proto`. + x-stackQL-resources: + account_metastore_assignments: + id: databricks_account.catalog.account_metastore_assignments + name: account_metastore_assignments + title: Account Metastore Assignments + methods: + account_metastore_assignments_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastores~1{metastore_id}/post + response: + mediaType: application/json + openAPIDocKey: '200' + account_metastore_assignments_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastores~1{metastore_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_metastore_assignments_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastores~1{metastore_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + account_metastore_assignments_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastore/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.metastore_assignment + account_metastore_assignments_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1workspaces/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_metastore_assignments/methods/account_metastore_assignments_get + - $ref: >- + #/components/x-stackQL-resources/account_metastore_assignments/methods/account_metastore_assignments_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_metastore_assignments/methods/account_metastore_assignments_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/account_metastore_assignments/methods/account_metastore_assignments_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/account_metastore_assignments/methods/account_metastore_assignments_update + account_metastores: + id: databricks_account.catalog.account_metastores + name: account_metastores + title: Account Metastores + methods: + account_metastores_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores/post' + response: + mediaType: application/json + openAPIDocKey: '200' + account_metastores_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.metastores + account_metastores_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_metastores_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.metastore_info + account_metastores_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_metastores/methods/account_metastores_get + - $ref: >- + #/components/x-stackQL-resources/account_metastores/methods/account_metastores_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_metastores/methods/account_metastores_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/account_metastores/methods/account_metastores_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/account_metastores/methods/account_metastores_update + account_storage_credentials: + id: databricks_account.catalog.account_storage_credentials + name: account_storage_credentials + title: Account Storage Credentials + methods: + account_storage_credentials_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials/post + response: + mediaType: application/json + openAPIDocKey: '200' + account_storage_credentials_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.storage_credentials + account_storage_credentials_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials~1{storage_credential_name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_storage_credentials_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials~1{storage_credential_name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.credential_info + account_storage_credentials_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials~1{storage_credential_name}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_storage_credentials/methods/account_storage_credentials_get + - $ref: >- + #/components/x-stackQL-resources/account_storage_credentials/methods/account_storage_credentials_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_storage_credentials/methods/account_storage_credentials_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/account_storage_credentials/methods/account_storage_credentials_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/account_storage_credentials/methods/account_storage_credentials_update + credentials: + id: databricks_account.catalog.credentials + name: credentials + title: Credentials + methods: + create_credential: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list_credentials: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.credentials + delete_credential: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials~1{name_arg}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get_credential: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials~1{name_arg}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update_credential: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials~1{name_arg}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + credentials_generate_temporary_service_credential: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1temporary-service-credentials/post + response: + mediaType: application/json + openAPIDocKey: '200' + credentials_validate_credential: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1validate-credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/get_credential + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/list_credentials + insert: + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/create_credential + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/credentials_generate_temporary_service_credential + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/credentials_validate_credential + update: + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/update_credential + delete: + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/delete_credential + replace: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/iam.yaml b/providers/src/databricks_account/v00.00.00000/services/iam.yaml index 9986bd7d..03deda59 100644 --- a/providers/src/databricks_account/v00.00.00000/services/iam.yaml +++ b/providers/src/databricks_account/v00.00.00000/services/iam.yaml @@ -1,2037 +1,3159 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Iam API - description: Identity and Access Management -servers: -- url: https://accounts.cloud.databricks.com -paths: - /api/2.0/preview/accounts/{account_id}/access-control/assignable-roles: - get: - operationId: assignable-roles-getassignablerolesforresource - externalDocs: - url: https://docs.databricks.com/api/account/accountaccesscontrol/getassignablerolesforresource - x-stackQL-resource: assignable_roles - x-stackQL-method: getassignablerolesforresource - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: resource - required: true - description: string - in: query - responses: - '200': - description: Assignable roles were returned successfully. - content: - application/json: - schema: - type: object - properties: - roles: - type: array - items: - type: object - properties: - name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.roles - description: Gets all the roles that can be granted on an account level resource. - A role is grantable if the rule set on the resource can contain an access - rule of the role. - /api/2.0/accounts/{account_id}/scim/v2/Groups: - post: - operationId: groups-create - externalDocs: - url: https://docs.databricks.com/api/account/accountgroups/create - x-stackQL-resource: groups - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The group creation request succeeded. - content: - application/json: - schema: - type: object - properties: - id: - type: string - displayName: - type: string - members: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: integer - format: int64 - displayName: - type: string - externalId: - type: string - members: - type: Array of object - roles: - type: Array of object - example: - id: string - displayName: string - members: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - description: Properties of the new group. - description: Creates a group in the Databricks account with a unique name, using - the supplied group details. - get: - operationId: groups-list - externalDocs: - url: https://docs.databricks.com/api/account/accountgroups/list - x-stackQL-resource: groups - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: List groups operation was succesful. - content: - application/json: - schema: - type: object - properties: - totalResults: - type: integer - startIndex: - type: integer - itemsPerPage: - type: integer - Resources: - type: array - items: - type: object - properties: - id: - type: string - displayName: - type: string - members: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.Resources - description: Gets all details of the groups associated with the Databricks account. - /api/2.0/accounts/{account_id}/scim/v2/Groups/{id}: - delete: - operationId: groups-delete - externalDocs: - url: https://docs.databricks.com/api/account/accountgroups/delete - x-stackQL-resource: groups - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '204': - description: Group was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a group from the Databricks account. - get: - operationId: groups-get - externalDocs: - url: https://docs.databricks.com/api/account/accountgroups/get - x-stackQL-resource: groups - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Group information was returned successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - displayName: - type: string - members: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the information for a specific group in the Databricks account. - patch: - operationId: groups-patch - externalDocs: - url: https://docs.databricks.com/api/account/accountgroups/patch - x-stackQL-resource: groups - x-stackQL-method: patch - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Group information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - Operations: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:api:messages:2.0:PatchOp - Operations: - - op: add - path: string - value: null - description: Operations to be applied on group information. - description: Partially updates the details of a group. - put: - operationId: groups-update - externalDocs: - url: https://docs.databricks.com/api/account/accountgroups/update - x-stackQL-resource: groups - x-stackQL-method: update - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Group information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: integer - format: int64 - displayName: - type: string - externalId: - type: string - members: - type: Array of object - roles: - type: Array of object - example: - id: string - displayName: string - members: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - description: Operations to be applied on group information. - description: Updates the details of a group by replacing the entire group entity. - /api/2.0/preview/accounts/{account_id}/access-control/rule-sets: - get: - operationId: rule-sets-getruleset - externalDocs: - url: https://docs.databricks.com/api/account/accountaccesscontrol/getruleset - x-stackQL-resource: rule_sets - x-stackQL-method: getruleset - x-stackQL-verb: select - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: name - required: true - description: string - in: query - - name: etag - required: true - description: string - in: query - responses: - '200': - description: The rule set was returned successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - grant_rules: - type: array - items: - type: object - properties: - principals: - type: array - items: - type: string - role: - type: string - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - description: Get a rule set by its name. A rule set is always attached to a - resource and contains a list of access rules on the said resource. Currently - only a default rule set for each resource is supported. - put: - operationId: rule-sets-updateruleset - externalDocs: - url: https://docs.databricks.com/api/account/accountaccesscontrol/updateruleset - x-stackQL-resource: rule_sets - x-stackQL-method: updateruleset - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: Rule sets on the resource was updated successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - grant_rules: - type: array - items: - type: object - properties: - role: - type: string - principals: - type: array - items: - type: string - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - rule_set: - type: object - example: - name: string - rule_set: - name: string - grant_rules: - - principals: - - string - role: string - etag: string - description: The rule set to update to. - description: Replace the rules of a rule set. First, use get to read the current - version of the rule set before modifying it. This pattern helps prevent conflicts - between concurrent updates. - /api/2.0/accounts/{account_id}/scim/v2/ServicePrincipals: - post: - operationId: service-principals-create - externalDocs: - url: https://docs.databricks.com/api/account/accountserviceprincipals/create - x-stackQL-resource: service_principals - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The user creation request succeeded. - content: - application/json: - schema: - type: object - properties: - id: - type: string - applicationId: - type: string - displayName: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: integer - format: int64 - applicationId: - type: uuid - displayName: - type: string - externalId: - type: string - active: - type: boolean - roles: - type: Array of object - example: - id: string - applicationId: 97ab27fa-30e2-43e3-92a3-160e80f4c0d5 - displayName: etl-service - roles: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Properties of the new service principal. - description: Creates a new service principal in the Databricks account. - get: - operationId: service-principals-list - externalDocs: - url: https://docs.databricks.com/api/account/accountserviceprincipals/list - x-stackQL-resource: service_principals - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: List service principals operation was succesful. - content: - application/json: - schema: - type: object - properties: - totalResults: - type: integer - startIndex: - type: integer - itemsPerPage: - type: integer - Resources: - type: array - items: - type: object - properties: - id: - type: string - applicationId: - type: string - displayName: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.Resources - description: Gets the set of service principals associated with a Databricks - account. - /api/2.0/accounts/{account_id}/scim/v2/ServicePrincipals/{id}: - delete: - operationId: service-principals-delete - externalDocs: - url: https://docs.databricks.com/api/account/accountserviceprincipals/delete - x-stackQL-resource: service_principals - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '204': - description: Service principal was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Delete a single service principal in the Databricks account. - get: - operationId: service-principals-get - externalDocs: - url: https://docs.databricks.com/api/account/accountserviceprincipals/get - x-stackQL-resource: service_principals - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Service principal information was returned successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - applicationId: - type: string - displayName: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the details for a single service principal define in the Databricks - account. - patch: - operationId: service-principals-patch - externalDocs: - url: https://docs.databricks.com/api/account/accountserviceprincipals/patch - x-stackQL-resource: service_principals - x-stackQL-method: patch - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Service principal information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - Operations: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:api:messages:2.0:PatchOp - Operations: - - op: add - path: string - value: null - description: Operations to be applied on service principal information. - description: Partially updates the details of a single service principal in - the Databricks account. - put: - operationId: service-principals-update - externalDocs: - url: https://docs.databricks.com/api/account/accountserviceprincipals/update - x-stackQL-resource: service_principals - x-stackQL-method: update - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Service principal information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: integer - format: int64 - applicationId: - type: uuid - displayName: - type: string - externalId: - type: string - active: - type: boolean - roles: - type: Array of object - example: - id: string - applicationId: 97ab27fa-30e2-43e3-92a3-160e80f4c0d5 - displayName: etl-service - roles: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Updates the details of a single service principal. - /api/2.0/accounts/{account_id}/scim/v2/Users: - post: - operationId: users-create - externalDocs: - url: https://docs.databricks.com/api/account/accountusers/create - x-stackQL-resource: users - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The user creation request succeeded. - content: - application/json: - schema: - type: object - properties: - id: - type: string - userName: - type: string - emails: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - name: - type: object - properties: - givenName: - type: string - familyName: - type: string - displayName: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: integer - format: int64 - userName: - type: email - displayName: - type: string - externalId: - type: string - active: - type: boolean - emails: - type: Array of object - name: - type: object - roles: - type: Array of object - example: - id: string - userName: user@example.com - emails: - - $ref: string - value: string - display: string - primary: true - type: string - name: - givenName: string - familyName: string - displayName: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Properties of the new user. - description: Creates a new user in the Databricks account. This new user will - also be added to the Databricks account. - get: - operationId: users-list - externalDocs: - url: https://docs.databricks.com/api/account/accountusers/list - x-stackQL-resource: users - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: List users operation was succesful. - content: - application/json: - schema: - type: object - properties: - totalResults: - type: integer - startIndex: - type: integer - itemsPerPage: - type: integer - Resources: - type: array - items: - type: object - properties: - id: - type: string - userName: - type: string - emails: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - name: - type: object - properties: - givenName: - type: string - familyName: - type: string - displayName: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.Resources - description: Gets details for all the users associated with a Databricks account. - /api/2.0/accounts/{account_id}/scim/v2/Users/{id}: - delete: - operationId: users-delete - externalDocs: - url: https://docs.databricks.com/api/account/accountusers/delete - x-stackQL-resource: users - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '204': - description: User was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a user. Deleting a user from a Databricks account also - removes objects associated with the user. - get: - operationId: users-get - externalDocs: - url: https://docs.databricks.com/api/account/accountusers/get - x-stackQL-resource: users - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: User information was returned successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - userName: - type: string - emails: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - name: - type: object - properties: - givenName: - type: string - familyName: - type: string - displayName: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets information for a specific user in Databricks account. - patch: - operationId: users-patch - externalDocs: - url: https://docs.databricks.com/api/account/accountusers/patch - x-stackQL-resource: users - x-stackQL-method: patch - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: User information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - Operations: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:api:messages:2.0:PatchOp - Operations: - - op: add - path: string - value: null - description: Operations to be applied on user information. - description: Partially updates a user resource by applying the supplied operations - on specific user attributes. - put: - operationId: users-update - externalDocs: - url: https://docs.databricks.com/api/account/accountusers/update - x-stackQL-resource: users - x-stackQL-method: update - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: User information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: integer - format: int64 - userName: - type: email - displayName: - type: string - externalId: - type: string - active: - type: boolean - emails: - type: Array of object - name: - type: object - roles: - type: Array of object - example: - id: string - userName: user@example.com - emails: - - $ref: string - value: string - display: string - primary: true - type: string - name: - givenName: string - familyName: string - displayName: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Replaces a user's information with the data supplied in request. - /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/permissionassignments/permissions: - get: - operationId: workspace-permissions-get - externalDocs: - url: https://docs.databricks.com/api/account/workspaceassignment/get - x-stackQL-resource: workspace_permissions - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permissions - description: Get an array of workspace permissions for the specified account - and workspace. - /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/permissionassignments/principals/{principal_id}: - put: - operationId: workspace-permission-assignments-createorupdate - externalDocs: - url: https://docs.databricks.com/api/account/workspaceassignment/update - x-stackQL-resource: workspace_permission_assignments - x-stackQL-method: createorupdate - x-stackQL-verb: insert - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: string - in: path - - name: workspace_id - required: true - description: int64 - in: path - - name: principal_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - principal: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - principal_id: - type: integer - display_name: - type: string - permissions: - type: array - items: - type: string - error: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - permissions: - type: Array of string - example: - permissions: - - UNKNOWN - description: Creates or updates the workspace permissions assignment in a given - account and workspace for the specified principal. - delete: - operationId: workspace-permission-assignments-delete - externalDocs: - url: https://docs.databricks.com/api/account/workspaceassignment/delete - x-stackQL-resource: workspace_permission_assignments - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: string - in: path - - name: workspace_id - required: true - description: int64 - in: path - - name: principal_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - description: Deletes the workspace permissions assignment in a given account - and workspace for the specified principal. - /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/permissionassignments: - get: - operationId: workspace-permission-assignments-list - externalDocs: - url: https://docs.databricks.com/api/account/workspaceassignment/list - x-stackQL-resource: workspace_permission_assignments - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - permission_assignments: - type: array - items: - type: object - properties: - principal: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - principal_id: - type: integer - display_name: - type: string - permissions: - type: array - items: - type: string - error: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_assignments - description: Get the permission assignments for the specified Databricks account - and Databricks workspace. -components: - x-stackQL-resources: - assignable_roles: - id: databricks_account.iam.assignable_roles - name: assignable_roles - title: Assignable_roles - methods: - getassignablerolesforresource: - operation: - $ref: '#/paths/~1api~12.0~1preview~1accounts~1{account_id}~1access-control~1assignable-roles/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.roles - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/assignable_roles/methods/getassignablerolesforresource' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - groups: - id: databricks_account.iam.groups - name: groups - title: Groups - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.Resources - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - patch: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/groups/methods/get' - - $ref: '#/components/x-stackQL-resources/groups/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/groups/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/groups/methods/patch' - replace: - - $ref: '#/components/x-stackQL-resources/groups/methods/update' - delete: - - $ref: '#/components/x-stackQL-resources/groups/methods/delete' - exec: [] - rule_sets: - id: databricks_account.iam.rule_sets - name: rule_sets - title: Rule_sets - methods: - getruleset: - operation: - $ref: '#/paths/~1api~12.0~1preview~1accounts~1{account_id}~1access-control~1rule-sets/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updateruleset: - operation: - $ref: '#/paths/~1api~12.0~1preview~1accounts~1{account_id}~1access-control~1rule-sets/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/rule_sets/methods/updateruleset' - - $ref: '#/components/x-stackQL-resources/rule_sets/methods/getruleset' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - service_principals: - id: databricks_account.iam.service_principals - name: service_principals - title: Service_principals - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.Resources - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - patch: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/get' - - $ref: '#/components/x-stackQL-resources/service_principals/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/patch' - replace: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/update' - delete: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/delete' - exec: [] - users: - id: databricks_account.iam.users - name: users - title: Users - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.Resources - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - patch: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/users/methods/get' - - $ref: '#/components/x-stackQL-resources/users/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/users/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/users/methods/patch' - replace: - - $ref: '#/components/x-stackQL-resources/users/methods/update' - delete: - - $ref: '#/components/x-stackQL-resources/users/methods/delete' - exec: [] - workspace_permissions: - id: databricks_account.iam.workspace_permissions - name: workspace_permissions - title: Workspace_permissions - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments~1permissions/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permissions - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/workspace_permissions/methods/get' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - workspace_permission_assignments: - id: databricks_account.iam.workspace_permission_assignments - name: workspace_permission_assignments - title: Workspace_permission_assignments - methods: - createorupdate: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments~1principals~1{principal_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments~1principals~1{principal_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_assignments - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/workspace_permission_assignments/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/workspace_permission_assignments/methods/createorupdate' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/workspace_permission_assignments/methods/delete' - exec: [] +openapi: 3.0.0 +info: + title: Databricks Iam API (account) + description: >- + OpenAPI specification for the Databricks iam service (account-level APIs), + generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.iam +servers: + - url: https://accounts.cloud.databricks.com +paths: + /api/2.0/preview/accounts/{account_id}/access-control/assignable-roles: + get: + operationId: account_access_control_get_assignable_roles_for_resource + summary: >- + Gets all the roles that can be granted on an account level resource. A + role is grantable if the rule + tags: + - iam + - account_access_control + description: >- + Gets all the roles that can be granted on an account level resource. A + role is grantable if the rule + + set on the resource can contain an access rule of the role. + + + :param resource: str + The resource name for which assignable roles will be listed. + + Examples | Summary :--- | :--- `resource=accounts/` | A resource name for the account. + `resource=accounts//groups/` | A resource name for the group. + `resource=accounts//servicePrincipals/` | A resource name for the service + principal. `resource=accounts//tagPolicies/` | A resource name for the + tag policy. + + :returns: :class:`GetAssignableRolesForResourceResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: resource + in: query + required: true + schema: + type: string + description: >- + The resource name for which assignable roles will be listed. + Examples | Summary :--- | :--- `resource=accounts/` | A + resource name for the account. + `resource=accounts//groups/` | A resource name + for the group. + `resource=accounts//servicePrincipals/` | A + resource name for the service principal. + `resource=accounts//tagPolicies/` | A + resource name for the tag policy. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetAssignableRolesForResourceResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountAccessControlAPI + /api/2.0/preview/accounts/{account_id}/access-control/rule-sets: + get: + operationId: account_access_control_get_rule_set + summary: >- + Get a rule set by its name. A rule set is always attached to a resource + and contains a list of access + tags: + - iam + - account_access_control + description: >- + Get a rule set by its name. A rule set is always attached to a resource + and contains a list of access + + rules on the said resource. Currently only a default rule set for each + resource is supported. + + + :param name: str + The ruleset name associated with the request. + + Examples | Summary :--- | :--- `name=accounts//ruleSets/default` | A name for a rule set + on the account. `name=accounts//groups//ruleSets/default` | A name for a rule + set on the group. + `name=accounts//servicePrincipals//ruleSets/default` | + A name for a rule set on the service principal. + `name=accounts//tagPolicies//ruleSets/default` | A name for a rule set on + the tag policy. + :param etag: str + Etag used for versioning. The response is at least as fresh as the eTag provided. Etag is used for + optimistic concurrency control as a way to help prevent simultaneous updates of a rule set from + overwriting each other. It is strongly suggested that systems make use of the etag in the read -> + modify -> write pattern to perform rule set updates in order to avoid race conditions that is get an + etag from a GET rule set request, and pass it with the PUT update request to identify the rule set + version you are updating. + + Examples | Summary :--- | :--- `etag=` | An empty etag can only be used in GET to indicate no + freshness requirements. `etag=RENUAAABhSweA4NvVmmUYdiU717H3Tgy0UJdor3gE4a+mq/oj9NjAf8ZsQ==` | An + etag encoded a specific version of the rule set to get or to be updated. + + :returns: :class:`RuleSetResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: name + in: query + required: true + schema: + type: string + description: >- + The ruleset name associated with the request. Examples | Summary + :--- | :--- `name=accounts//ruleSets/default` | A name + for a rule set on the account. + `name=accounts//groups//ruleSets/default` | A + name for a rule set on the group. + `name=accounts//servicePrincipals//ruleSets/default` + | A name for a rule set on the service principal. + `name=accounts//tagPolicies//ruleSets/default` + | A name for a rule set on the tag policy. + - name: etag + in: query + required: true + schema: + type: string + description: >- + Etag used for versioning. The response is at least as fresh as the + eTag provided. Etag is used for optimistic concurrency control as a + way to help prevent simultaneous updates of a rule set from + overwriting each other. It is strongly suggested that systems make + use of the etag in the read -> modify -> write pattern to perform + rule set updates in order to avoid race conditions that is get an + etag from a GET rule set request, and pass it with the PUT update + request to identify the rule set version you are updating. Examples + | Summary :--- | :--- `etag=` | An empty etag can only be used in + GET to indicate no freshness requirements. + `etag=RENUAAABhSweA4NvVmmUYdiU717H3Tgy0UJdor3gE4a+mq/oj9NjAf8ZsQ==` + | An etag encoded a specific version of the rule set to get or to be + updated. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RuleSetResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountAccessControlAPI + put: + operationId: account_access_control_update_rule_set + summary: >- + Replace the rules of a rule set. First, use get to read the current + version of the rule set before + tags: + - iam + - account_access_control + description: >- + Replace the rules of a rule set. First, use get to read the current + version of the rule set before + + modifying it. This pattern helps prevent conflicts between concurrent + updates. + + + :param name: str + Name of the rule set. + :param rule_set: :class:`RuleSetUpdateRequest` + + + :returns: :class:`RuleSetResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the rule set. + rule_set: + type: string + description: ':returns: :class:`RuleSetResponse`' + required: + - name + - rule_set + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RuleSetResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountAccessControlAPI + /api/2.0/accounts/{account_id}/scim/v2/Groups: + post: + operationId: account_groups_v2_create + summary: >- + Creates a group in the Databricks account with a unique name, using the + supplied group details. + tags: + - iam + - account_groups_v2 + description: >- + Creates a group in the Databricks account with a unique name, using the + supplied group details. + + + :param display_name: str (optional) + String that represents a human-readable group name + :param external_id: str (optional) + + :param id: str (optional) + Databricks group ID + :param members: List[:class:`ComplexValue`] (optional) + + :param meta: :class:`ResourceMeta` (optional) + Container for the group identifier. Workspace local versus account. + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + + :returns: :class:`AccountGroup` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + display_name: + type: string + description: String that represents a human-readable group name + external_id: + type: string + description: ':param id: str (optional) Databricks group ID' + id: + type: string + members: + type: string + description: >- + :param meta: :class:`ResourceMeta` (optional) Container for + the group identifier. Workspace local versus account. + meta: + type: string + roles: + type: string + description: Indicates if the group has the admin role. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountGroup' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountGroupsV2API + get: + operationId: account_groups_v2_list + summary: >- + Gets all details of the groups associated with the Databricks account. + As of 08/22/2025, this endpoint + tags: + - iam + - account_groups_v2 + description: >- + Gets all details of the groups associated with the Databricks account. + As of 08/22/2025, this endpoint + + will no longer return members. Instead, members should be retrieved by + iterating through `Get group + + details`. Existing accounts that rely on this attribute will not be + impacted and will continue + + receiving member data as before. + + + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. Default is 10000. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. + :param sort_order: :class:`ListSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: Iterator over :class:`AccountGroup` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. Default is 10000. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: Attribute to sort the results. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountGroupsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountGroupsV2API + /api/2.0/accounts/{account_id}/scim/v2/Groups/{id}: + delete: + operationId: account_groups_v2_delete + summary: Deletes a group from the Databricks account. + tags: + - iam + - account_groups_v2 + description: |- + Deletes a group from the Databricks account. + + :param id: str + Unique ID for a group in the Databricks account. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a group in the Databricks account. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountGroupsV2API + get: + operationId: account_groups_v2_get + summary: Gets the information for a specific group in the Databricks account. + tags: + - iam + - account_groups_v2 + description: |- + Gets the information for a specific group in the Databricks account. + + :param id: str + Unique ID for a group in the Databricks account. + + :returns: :class:`AccountGroup` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a group in the Databricks account. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountGroup' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountGroupsV2API + patch: + operationId: account_groups_v2_patch + summary: Partially updates the details of a group. + tags: + - iam + - account_groups_v2 + description: |- + Partially updates the details of a group. + + :param id: str + Unique ID in the Databricks workspace. + :param operations: List[:class:`Patch`] (optional) + :param schemas: List[:class:`PatchSchema`] (optional) + The schema of the patch request. Must be ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID in the Databricks workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + operations: + type: string + description: >- + :param schemas: List[:class:`PatchSchema`] (optional) The + schema of the patch request. Must be + ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + schemas: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountGroupsV2API + put: + operationId: account_groups_v2_update + summary: Updates the details of a group by replacing the entire group entity. + tags: + - iam + - account_groups_v2 + description: |- + Updates the details of a group by replacing the entire group entity. + + :param id: str + Databricks group ID + :param display_name: str (optional) + String that represents a human-readable group name + :param external_id: str (optional) + :param members: List[:class:`ComplexValue`] (optional) + :param meta: :class:`ResourceMeta` (optional) + Container for the group identifier. Workspace local versus account. + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Databricks group ID + requestBody: + content: + application/json: + schema: + type: object + properties: + display_name: + type: string + description: String that represents a human-readable group name + external_id: + type: string + description: ':param members: List[:class:`ComplexValue`] (optional)' + members: + type: string + meta: + type: string + description: >- + Container for the group identifier. Workspace local versus + account. + roles: + type: string + description: Indicates if the group has the admin role. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountGroupsV2API + /api/2.0/accounts/{account_id}/scim/v2/ServicePrincipals: + post: + operationId: account_service_principals_v2_create + summary: Creates a new service principal in the Databricks account. + tags: + - iam + - account_service_principals_v2 + description: |- + Creates a new service principal in the Databricks account. + + :param active: bool (optional) + If this user is active + :param application_id: str (optional) + UUID relating to the service principal + :param display_name: str (optional) + String that represents a concatenation of given and family names. + :param external_id: str (optional) + :param id: str (optional) + Databricks service principal ID. + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + + :returns: :class:`AccountServicePrincipal` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + application_id: + type: string + description: UUID relating to the service principal + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. + external_id: + type: string + description: ':param id: str (optional) Databricks service principal ID.' + id: + type: string + roles: + type: string + description: Indicates if the group has the admin role. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountServicePrincipal' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountServicePrincipalsV2API + get: + operationId: account_service_principals_v2_list + summary: Gets the set of service principals associated with a Databricks account. + tags: + - iam + - account_service_principals_v2 + description: |- + Gets the set of service principals associated with a Databricks account. + + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. Default is 10000. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. + :param sort_order: :class:`ListSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: Iterator over :class:`AccountServicePrincipal` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. Default is 10000. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: Attribute to sort the results. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountServicePrincipalsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountServicePrincipalsV2API + /api/2.0/accounts/{account_id}/scim/v2/ServicePrincipals/{id}: + delete: + operationId: account_service_principals_v2_delete + summary: Delete a single service principal in the Databricks account. + tags: + - iam + - account_service_principals_v2 + description: |- + Delete a single service principal in the Databricks account. + + :param id: str + Unique ID for a service principal in the Databricks account. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a service principal in the Databricks account. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountServicePrincipalsV2API + get: + operationId: account_service_principals_v2_get + summary: >- + Gets the details for a single service principal define in the Databricks + account. + tags: + - iam + - account_service_principals_v2 + description: >- + Gets the details for a single service principal define in the Databricks + account. + + + :param id: str + Unique ID for a service principal in the Databricks account. + + :returns: :class:`AccountServicePrincipal` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a service principal in the Databricks account. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountServicePrincipal' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountServicePrincipalsV2API + patch: + operationId: account_service_principals_v2_patch + summary: >- + Partially updates the details of a single service principal in the + Databricks account. + tags: + - iam + - account_service_principals_v2 + description: >- + Partially updates the details of a single service principal in the + Databricks account. + + + :param id: str + Unique ID in the Databricks workspace. + :param operations: List[:class:`Patch`] (optional) + + :param schemas: List[:class:`PatchSchema`] (optional) + The schema of the patch request. Must be ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID in the Databricks workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + operations: + type: string + description: >- + :param schemas: List[:class:`PatchSchema`] (optional) The + schema of the patch request. Must be + ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + schemas: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountServicePrincipalsV2API + put: + operationId: account_service_principals_v2_update + summary: Updates the details of a single service principal. + tags: + - iam + - account_service_principals_v2 + description: |- + Updates the details of a single service principal. + + This action replaces the existing service principal with the same name. + + :param id: str + Databricks service principal ID. + :param active: bool (optional) + If this user is active + :param application_id: str (optional) + UUID relating to the service principal + :param display_name: str (optional) + String that represents a concatenation of given and family names. + :param external_id: str (optional) + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Databricks service principal ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + application_id: + type: string + description: UUID relating to the service principal + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. + external_id: + type: string + description: >- + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + roles: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountServicePrincipalsV2API + /api/2.0/accounts/{account_id}/scim/v2/Users: + post: + operationId: account_users_v2_create + summary: >- + Creates a new user in the Databricks account. This new user will also be + added to the Databricks + tags: + - iam + - account_users_v2 + description: >- + Creates a new user in the Databricks account. This new user will also be + added to the Databricks + + account. + + + :param active: bool (optional) + If this user is active + :param display_name: str (optional) + String that represents a concatenation of given and family names. For example `John Smith`. + :param emails: List[:class:`ComplexValue`] (optional) + All the emails associated with the Databricks user. + :param external_id: str (optional) + External ID is not currently supported. It is reserved for future use. + :param id: str (optional) + Databricks user ID. + :param name: :class:`Name` (optional) + + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + :param user_name: str (optional) + Email address of the Databricks user. + + :returns: :class:`AccountUser` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. For example `John Smith`. + emails: + type: string + description: All the emails associated with the Databricks user. + external_id: + type: string + description: >- + External ID is not currently supported. It is reserved for + future use. + id: + type: string + description: Databricks user ID. + name: + type: string + description: >- + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + roles: + type: string + user_name: + type: string + description: Email address of the Databricks user. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountUser' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountUsersV2API + get: + operationId: account_users_v2_list + summary: Gets details for all the users associated with a Databricks account. + tags: + - iam + - account_users_v2 + description: |- + Gets details for all the users associated with a Databricks account. + + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. Default is 10000. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. Multi-part paths are supported. For example, `userName`, + `name.givenName`, and `emails`. + :param sort_order: :class:`ListSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: Iterator over :class:`AccountUser` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. Default is 10000. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: >- + Attribute to sort the results. Multi-part paths are supported. For + example, `userName`, `name.givenName`, and `emails`. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountUsersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountUsersV2API + /api/2.0/accounts/{account_id}/scim/v2/Users/{id}: + delete: + operationId: account_users_v2_delete + summary: >- + Deletes a user. Deleting a user from a Databricks account also removes + objects associated with the + tags: + - iam + - account_users_v2 + description: >- + Deletes a user. Deleting a user from a Databricks account also removes + objects associated with the + + user. + + + :param id: str + Unique ID for a user in the Databricks account. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a user in the Databricks account. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountUsersV2API + get: + operationId: account_users_v2_get + summary: Gets information for a specific user in Databricks account. + tags: + - iam + - account_users_v2 + description: |- + Gets information for a specific user in Databricks account. + + :param id: str + Unique ID for a user in the Databricks account. + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. Default is 10000. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. Multi-part paths are supported. For example, `userName`, + `name.givenName`, and `emails`. + :param sort_order: :class:`GetSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: :class:`AccountUser` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a user in the Databricks account. + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. Default is 10000. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: >- + Attribute to sort the results. Multi-part paths are supported. For + example, `userName`, `name.givenName`, and `emails`. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountUser' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountUsersV2API + patch: + operationId: account_users_v2_patch + summary: >- + Partially updates a user resource by applying the supplied operations on + specific user attributes. + tags: + - iam + - account_users_v2 + description: >- + Partially updates a user resource by applying the supplied operations on + specific user attributes. + + + :param id: str + Unique ID in the Databricks workspace. + :param operations: List[:class:`Patch`] (optional) + + :param schemas: List[:class:`PatchSchema`] (optional) + The schema of the patch request. Must be ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Unique ID in the Databricks workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + operations: + type: string + description: >- + :param schemas: List[:class:`PatchSchema`] (optional) The + schema of the patch request. Must be + ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + schemas: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountUsersV2API + put: + operationId: account_users_v2_update + summary: Replaces a user's information with the data supplied in request. + tags: + - iam + - account_users_v2 + description: |- + Replaces a user's information with the data supplied in request. + + :param id: str + Databricks user ID. + :param active: bool (optional) + If this user is active + :param display_name: str (optional) + String that represents a concatenation of given and family names. For example `John Smith`. + :param emails: List[:class:`ComplexValue`] (optional) + All the emails associated with the Databricks user. + :param external_id: str (optional) + External ID is not currently supported. It is reserved for future use. + :param name: :class:`Name` (optional) + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + :param user_name: str (optional) + Email address of the Databricks user. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: id + in: path + required: true + schema: + type: string + description: Databricks user ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. For example `John Smith`. + emails: + type: string + description: All the emails associated with the Databricks user. + external_id: + type: string + description: >- + External ID is not currently supported. It is reserved for + future use. + name: + type: string + description: >- + :param roles: List[:class:`ComplexValue`] (optional) + Indicates if the group has the admin role. + roles: + type: string + user_name: + type: string + description: Email address of the Databricks user. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountUsersV2API + /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/permissionassignments/principals/{principal_id}: + delete: + operationId: workspace_assignment_delete + summary: >- + Deletes the workspace permissions assignment in a given account and + workspace for the specified + tags: + - iam + - workspace_assignment + description: >- + Deletes the workspace permissions assignment in a given account and + workspace for the specified + + principal. + + + :param workspace_id: int + The workspace ID for the account. + :param principal_id: int + The ID of the user, service principal, or group. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: The workspace ID for the account. + - name: principal_id + in: path + required: true + schema: + type: integer + description: The ID of the user, service principal, or group. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAssignmentAPI + put: + operationId: workspace_assignment_update + summary: >- + Creates or updates the workspace permissions assignment in a given + account and workspace for the + tags: + - iam + - workspace_assignment + description: >- + Creates or updates the workspace permissions assignment in a given + account and workspace for the + + specified principal. + + + :param workspace_id: int + The workspace ID. + :param principal_id: int + The ID of the user, service principal, or group. + :param permissions: List[:class:`WorkspacePermission`] (optional) + Array of permissions assignments to update on the workspace. Valid values are "USER" and "ADMIN" + (case-sensitive). If both "USER" and "ADMIN" are provided, "ADMIN" takes precedence. Other values + will be ignored. Note that excluding this field, or providing unsupported values, will have the same + effect as providing an empty list, which will result in the deletion of all permissions for the + principal. + + :returns: :class:`PermissionAssignment` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: The workspace ID. + - name: principal_id + in: path + required: true + schema: + type: integer + description: The ID of the user, service principal, or group. + requestBody: + content: + application/json: + schema: + type: object + properties: + permissions: + type: string + description: >- + Array of permissions assignments to update on the workspace. + Valid values are "USER" and "ADMIN" (case-sensitive). If + both "USER" and "ADMIN" are provided, "ADMIN" takes + precedence. Other values will be ignored. Note that + excluding this field, or providing unsupported values, will + have the same effect as providing an empty list, which will + result in the deletion of all permissions for the principal. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PermissionAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAssignmentAPI + /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/permissionassignments/permissions: + get: + operationId: workspace_assignment_get + summary: >- + Get an array of workspace permissions for the specified account and + workspace. + tags: + - iam + - workspace_assignment + description: >- + Get an array of workspace permissions for the specified account and + workspace. + + + :param workspace_id: int + The workspace ID. + + :returns: :class:`WorkspacePermissions` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: The workspace ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspacePermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAssignmentAPI + /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/permissionassignments: + get: + operationId: workspace_assignment_list + summary: >- + Get the permission assignments for the specified Databricks account and + Databricks workspace. + tags: + - iam + - workspace_assignment + description: >- + Get the permission assignments for the specified Databricks account and + Databricks workspace. + + + :param workspace_id: int + The workspace ID for the account. + + :returns: Iterator over :class:`PermissionAssignment` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: The workspace ID for the account. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PermissionAssignments' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAssignmentAPI +components: + schemas: + AccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/PermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + AccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/Permission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + AccountGroup: + type: object + properties: + account_id: + type: string + displayName: + type: string + description: String that represents a human-readable group name + externalId: + type: string + description: external_id should be unique for identifying groups + id: + type: string + description: Databricks group ID + members: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + meta: + $ref: '#/components/schemas/ResourceMeta' + description: Container for the group identifier. Workspace local versus account. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Indicates if the group has the admin role. + AccountServicePrincipal: + type: object + properties: + account_id: + type: string + active: + type: boolean + description: If this user is active + applicationId: + type: string + description: UUID relating to the service principal + displayName: + type: string + description: String that represents a concatenation of given and family names. + externalId: + type: string + id: + type: string + description: Databricks service principal ID. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Indicates if the group has the admin role. + AccountUser: + type: object + properties: + account_id: + type: string + active: + type: boolean + description: If this user is active + displayName: + type: string + description: >- + String that represents a concatenation of given and family names. + For example `John Smith`. + emails: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: All the emails associated with the Databricks user. + externalId: + type: string + description: >- + External ID is not currently supported. It is reserved for future + use. + id: + type: string + description: Databricks user ID. + name: + $ref: '#/components/schemas/Name' + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Indicates if the group has the admin role. + userName: + type: string + description: Email address of the Databricks user. + Actor: + type: object + properties: + actor_id: + type: integer + description: >- + represents an identity trying to access a resource - user or a service + principal group can be a + principal of a permission set assignment but an actor is always a user or a service principal + CheckPolicyResponse: + type: object + properties: + consistency_token: + $ref: '#/components/schemas/ConsistencyToken' + is_permitted: + type: boolean + required: + - consistency_token + ComplexValue: + type: object + properties: + display: + type: string + primary: + type: boolean + $ref: + type: string + type: + type: string + value: + type: string + ConsistencyToken: + type: object + properties: + value: + type: string + required: + - value + DeleteWorkspacePermissionAssignmentResponse: + type: object + properties: {} + GetAssignableRolesForResourceResponse: + type: object + properties: + roles: + type: array + items: + $ref: '#/components/schemas/Role' + GetPasswordPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/PasswordPermissionsDescription' + GetPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/PermissionsDescription' + GrantRule: + type: object + properties: + role: + type: string + principals: + type: array + items: + type: string + description: >- + Principals this grant rule applies to. A principal can be a user + (for end users), a service principal (for applications and compute + workloads), or an account group. Each principal has its own + identifier format: * users/ * groups/ * + servicePrincipals/ + required: + - role + Group: + type: object + properties: + displayName: + type: string + entitlements: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: >- + Entitlements assigned to the group. See [assigning entitlements] for + a full list of supported values. [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + externalId: + type: string + description: external_id should be unique for identifying groups + groups: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + id: + type: string + description: Databricks group ID + members: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + meta: + $ref: '#/components/schemas/ResourceMeta' + description: Container for the group identifier. Workspace local versus account. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Corresponds to AWS instance profile/arn role. + schemas: + type: array + items: + $ref: '#/components/schemas/GroupSchema' + description: The schema of the group. + ListAccountGroupsResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/AccountGroup' + description: User objects returned in the response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListAccountServicePrincipalsResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/AccountServicePrincipal' + description: User objects returned in the response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListAccountUsersResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/AccountUser' + description: User objects returned in the response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListGroupsResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/Group' + description: User objects returned in the response. + schemas: + type: array + items: + $ref: '#/components/schemas/ListResponseSchema' + description: The schema of the service principal. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListServicePrincipalResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/ServicePrincipal' + description: User objects returned in the response. + schemas: + type: array + items: + $ref: '#/components/schemas/ListResponseSchema' + description: The schema of the List response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListUsersResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/User' + description: User objects returned in the response. + schemas: + type: array + items: + $ref: '#/components/schemas/ListResponseSchema' + description: The schema of the List response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + MigratePermissionsResponse: + type: object + properties: + permissions_migrated: + type: integer + Name: + type: object + properties: + familyName: + type: string + givenName: + type: string + description: Given name of the Databricks user. + ObjectPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/AccessControlResponse' + object_id: + type: string + object_type: + type: string + PasswordAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/PasswordPermissionLevel' + description: Permission level + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + PasswordAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/PasswordPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + PasswordPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/PasswordPermissionLevel' + description: Permission level + PasswordPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/PasswordAccessControlResponse' + object_id: + type: string + object_type: + type: string + PasswordPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/PasswordPermissionLevel' + description: Permission level + Patch: + type: object + properties: + op: + $ref: '#/components/schemas/PatchOp' + path: + type: string + description: Selection of patch operation + value: + type: object + description: Value to modify + Permission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/PermissionLevel' + PermissionAssignment: + type: object + properties: + error: + type: string + description: >- + Error response associated with a workspace permission assignment, if + any. + permissions: + type: array + items: + $ref: '#/components/schemas/WorkspacePermission' + description: The permissions level of the principal. + principal: + $ref: '#/components/schemas/PrincipalOutput' + description: Information about the principal assigned to the workspace. + description: >- + The output format for existing workspace PermissionAssignment records, + which contains some info + for user consumption. + PermissionAssignments: + type: object + properties: + permission_assignments: + type: array + items: + $ref: '#/components/schemas/PermissionAssignment' + PermissionOutput: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/WorkspacePermission' + PermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/PermissionLevel' + PrincipalOutput: + type: object + properties: + display_name: + type: string + description: The display name of the principal. + group_name: + type: string + description: >- + The group name of the group. Present only if the principal is a + group. + principal_id: + type: integer + description: The unique, opaque id of the principal. + service_principal_name: + type: string + description: >- + The name of the service principal. Present only if the principal is + a service principal. + user_name: + type: string + description: The username of the user. Present only if the principal is a user. + description: Information about the principal assigned to the workspace. + ResourceInfo: + type: object + properties: + id: + type: string + legacy_acl_path: + type: string + description: The legacy acl path of the current resource. + parent_resource_info: + $ref: '#/components/schemas/ResourceInfo' + description: >- + Parent resource info for the current resource. The parent may have + another parent. + required: + - id + ResourceMeta: + type: object + properties: + resourceType: + type: string + Role: + type: object + properties: + name: + type: string + required: + - name + RuleSetResponse: + type: object + properties: + name: + type: string + etag: + type: string + description: >- + Identifies the version of the rule set returned. Etag used for + versioning. The response is at least as fresh as the eTag provided. + Etag is used for optimistic concurrency control as a way to help + prevent simultaneous updates of a rule set from overwriting each + other. It is strongly suggested that systems make use of the etag in + the read -> modify -> write pattern to perform rule set updates in + order to avoid race conditions that is get an etag from a GET rule + set request, and pass it with the PUT update request to identify the + rule set version you are updating. + grant_rules: + type: array + items: + $ref: '#/components/schemas/GrantRule' + required: + - name + - etag + RuleSetUpdateRequest: + type: object + properties: + name: + type: string + etag: + type: string + description: >- + Identifies the version of the rule set returned. Etag used for + versioning. The response is at least as fresh as the eTag provided. + Etag is used for optimistic concurrency control as a way to help + prevent simultaneous updates of a rule set from overwriting each + other. It is strongly suggested that systems make use of the etag in + the read -> modify -> write pattern to perform rule set updates in + order to avoid race conditions that is get an etag from a GET rule + set request, and pass it with the PUT update request to identify the + rule set version you are updating. + grant_rules: + type: array + items: + $ref: '#/components/schemas/GrantRule' + required: + - name + - etag + ServicePrincipal: + type: object + properties: + active: + type: boolean + applicationId: + type: string + description: UUID relating to the service principal + displayName: + type: string + description: String that represents a concatenation of given and family names. + entitlements: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: >- + Entitlements assigned to the service principal. See [assigning + entitlements] for a full list of supported values. [assigning + entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + externalId: + type: string + groups: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + id: + type: string + description: Databricks service principal ID. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Corresponds to AWS instance profile/arn role. + schemas: + type: array + items: + $ref: '#/components/schemas/ServicePrincipalSchema' + description: The schema of the List response. + User: + type: object + properties: + active: + type: boolean + displayName: + type: string + description: >- + String that represents a concatenation of given and family names. + For example `John Smith`. This field cannot be updated through the + Workspace SCIM APIs when [identity federation is enabled]. Use + Account SCIM APIs to update `displayName`. [identity federation is + enabled]: + https://docs.databricks.com/administration-guide/users-groups/best-practices.html#enable-identity-federation + emails: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: All the emails associated with the Databricks user. + entitlements: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: >- + Entitlements assigned to the user. See [assigning entitlements] for + a full list of supported values. [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + externalId: + type: string + description: >- + External ID is not currently supported. It is reserved for future + use. + groups: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + id: + type: string + description: Databricks user ID. + name: + $ref: '#/components/schemas/Name' + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Corresponds to AWS instance profile/arn role. + schemas: + type: array + items: + $ref: '#/components/schemas/UserSchema' + description: The schema of the user. + userName: + type: string + description: Email address of the Databricks user. + WorkspacePermissions: + type: object + properties: + permissions: + type: array + items: + $ref: '#/components/schemas/PermissionOutput' + GetSortOrder: + type: string + enum: + - ascending + - descending + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + GroupSchema: + type: string + enum: + - urn:ietf:params:scim:schemas:core:2.0:Group + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListResponseSchema: + type: string + enum: + - urn:ietf:params:scim:api:messages:2.0:ListResponse + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListSortOrder: + type: string + enum: + - ascending + - descending + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PasswordPermissionLevel: + type: string + enum: + - CAN_USE + description: Permission level + PatchOp: + type: string + enum: + - add + - remove + - replace + description: Type of patch operation. + PatchSchema: + type: string + enum: + - urn:ietf:params:scim:api:messages:2.0:PatchOp + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PermissionLevel: + type: string + enum: + - CAN_ATTACH_TO + - CAN_BIND + - CAN_CREATE + - CAN_EDIT + - CAN_EDIT_METADATA + - CAN_MANAGE + - CAN_MANAGE_PRODUCTION_VERSIONS + - CAN_MANAGE_RUN + - CAN_MANAGE_STAGING_VERSIONS + - CAN_MONITOR + - CAN_MONITOR_ONLY + - CAN_QUERY + - CAN_READ + - CAN_RESTART + - CAN_RUN + - CAN_USE + - CAN_VIEW + - CAN_VIEW_METADATA + - IS_OWNER + description: Permission level + RequestAuthzIdentity: + type: string + enum: + - REQUEST_AUTHZ_IDENTITY_SERVICE_IDENTITY + - REQUEST_AUTHZ_IDENTITY_USER_CONTEXT + description: >- + Defines the identity to be used for authZ of the request on the server + side. See one pager for + + for more information: http://go/acl/service-identity + ServicePrincipalSchema: + type: string + enum: + - urn:ietf:params:scim:schemas:core:2.0:ServicePrincipal + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + UserSchema: + type: string + enum: + - urn:ietf:params:scim:schemas:core:2.0:User + - urn:ietf:params:scim:schemas:extension:workspace:2.0:User + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + WorkspacePermission: + type: string + enum: + - ADMIN + - UNKNOWN + - USER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + account_assignable_roles: + id: databricks_account.iam.account_assignable_roles + name: account_assignable_roles + title: Account Assignable Roles + methods: + get_assignable_roles_for_resource: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1accounts~1{account_id}~1access-control~1assignable-roles/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_assignable_roles/methods/get_assignable_roles_for_resource + insert: [] + update: [] + delete: [] + replace: [] + account_rule_sets: + id: databricks_account.iam.account_rule_sets + name: account_rule_sets + title: Account Rule Sets + methods: + get_rule_set: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1accounts~1{account_id}~1access-control~1rule-sets/get + response: + mediaType: application/json + openAPIDocKey: '200' + update_rule_set: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1accounts~1{account_id}~1access-control~1rule-sets/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_rule_sets/methods/get_rule_set + insert: [] + update: [] + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/account_rule_sets/methods/update_rule_set + account_groups: + id: databricks_account.iam.account_groups + name: account_groups + title: Account Groups + methods: + account_groups_v2_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups/post' + response: + mediaType: application/json + openAPIDocKey: '200' + account_groups_v2_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.Resources + account_groups_v2_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_groups_v2_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + account_groups_v2_patch: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + account_groups_v2_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Groups~1{id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_groups/methods/account_groups_v2_get + - $ref: >- + #/components/x-stackQL-resources/account_groups/methods/account_groups_v2_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_groups/methods/account_groups_v2_create + update: + - $ref: >- + #/components/x-stackQL-resources/account_groups/methods/account_groups_v2_patch + delete: + - $ref: >- + #/components/x-stackQL-resources/account_groups/methods/account_groups_v2_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/account_groups/methods/account_groups_v2_update + account_service_principals: + id: databricks_account.iam.account_service_principals + name: account_service_principals + title: Account Service Principals + methods: + account_service_principals_v2_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals/post + response: + mediaType: application/json + openAPIDocKey: '200' + account_service_principals_v2_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.Resources + account_service_principals_v2_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_service_principals_v2_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + account_service_principals_v2_patch: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + account_service_principals_v2_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1ServicePrincipals~1{id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_service_principals/methods/account_service_principals_v2_get + - $ref: >- + #/components/x-stackQL-resources/account_service_principals/methods/account_service_principals_v2_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_service_principals/methods/account_service_principals_v2_create + update: + - $ref: >- + #/components/x-stackQL-resources/account_service_principals/methods/account_service_principals_v2_patch + delete: + - $ref: >- + #/components/x-stackQL-resources/account_service_principals/methods/account_service_principals_v2_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/account_service_principals/methods/account_service_principals_v2_update + account_users: + id: databricks_account.iam.account_users + name: account_users + title: Account Users + methods: + account_users_v2_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users/post' + response: + mediaType: application/json + openAPIDocKey: '200' + account_users_v2_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.Resources + account_users_v2_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_users_v2_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + account_users_v2_patch: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + account_users_v2_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1scim~1v2~1Users~1{id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_users/methods/account_users_v2_get + - $ref: >- + #/components/x-stackQL-resources/account_users/methods/account_users_v2_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_users/methods/account_users_v2_create + update: + - $ref: >- + #/components/x-stackQL-resources/account_users/methods/account_users_v2_patch + delete: + - $ref: >- + #/components/x-stackQL-resources/account_users/methods/account_users_v2_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/account_users/methods/account_users_v2_update + workspace_assignment: + id: databricks_account.iam.workspace_assignment + name: workspace_assignment + title: Workspace Assignment + methods: + workspace_assignment_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments~1principals~1{principal_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + workspace_assignment_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments~1principals~1{principal_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + workspace_assignment_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.permission_assignments + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/workspace_assignment/methods/workspace_assignment_list + insert: [] + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/workspace_assignment/methods/workspace_assignment_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/workspace_assignment/methods/workspace_assignment_update + workspace_permissions: + id: databricks_account.iam.workspace_permissions + name: workspace_permissions + title: Workspace Permissions + methods: + workspace_assignment_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1permissionassignments~1permissions/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/workspace_permissions/methods/workspace_assignment_get + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/iamv2.yaml b/providers/src/databricks_account/v00.00.00000/services/iamv2.yaml new file mode 100644 index 00000000..ec8a80c5 --- /dev/null +++ b/providers/src/databricks_account/v00.00.00000/services/iamv2.yaml @@ -0,0 +1,466 @@ +openapi: 3.0.0 +info: + title: Databricks Iamv2 API (account) + description: >- + OpenAPI specification for the Databricks iamv2 service (account-level APIs), + generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.iamv2 +servers: + - url: https://accounts.cloud.databricks.com +paths: + /api/2.0/identity/accounts/{account_id}/workspaces/{workspace_id}/workspaceAccessDetails/{principal_id}: + get: + operationId: account_iam_v2_get_workspace_access_detail + summary: >- + Returns the access details for a principal in a workspace. Allows for + checking access details for any + tags: + - iamv2 + - account_iam_v2 + description: >- + Returns the access details for a principal in a workspace. Allows for + checking access details for any + + provisioned principal (user, service principal, or group) in a + workspace. * Provisioned principal here + + refers to one that has been synced into Databricks from the customer's + IdP or added explicitly to + + Databricks via SCIM/UI. Allows for passing in a "view" parameter to + control what fields are returned + + (BASIC by default or FULL). + + + :param workspace_id: int + Required. The workspace ID for which the access details are being requested. + :param principal_id: int + Required. The internal ID of the principal (user/sp/group) for which the access details are being + requested. + :param view: :class:`WorkspaceAccessDetailView` (optional) + Controls what fields are returned. + + :returns: :class:`WorkspaceAccessDetail` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: >- + Required. The workspace ID for which the access details are being + requested. + - name: principal_id + in: path + required: true + schema: + type: integer + description: >- + Required. The internal ID of the principal (user/sp/group) for which + the access details are being requested. + - name: view + in: query + required: false + schema: + type: string + description: Controls what fields are returned. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspaceAccessDetail' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIamV2API + /api/2.0/identity/accounts/{account_id}/groups/resolveByExternalId: + post: + operationId: account_iam_v2_resolve_group + summary: >- + Resolves a group with the given external ID from the customer's IdP. If + the group does not exist, it + tags: + - iamv2 + - account_iam_v2 + description: >- + Resolves a group with the given external ID from the customer's IdP. If + the group does not exist, it + + will be created in the account. If the customer is not onboarded onto + Automatic Identity Management + + (AIM), this will return an error. + + + :param external_id: str + Required. The external ID of the group in the customer's IdP. + + :returns: :class:`ResolveGroupResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + external_id: + type: string + description: >- + Required. The external ID of the group in the customer's + IdP. + required: + - external_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ResolveGroupResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIamV2API + /api/2.0/identity/accounts/{account_id}/servicePrincipals/resolveByExternalId: + post: + operationId: account_iam_v2_resolve_service_principal + summary: >- + Resolves an SP with the given external ID from the customer's IdP. If + the SP does not exist, it will + tags: + - iamv2 + - account_iam_v2 + description: >- + Resolves an SP with the given external ID from the customer's IdP. If + the SP does not exist, it will + + be created. If the customer is not onboarded onto Automatic Identity + Management (AIM), this will + + return an error. + + + :param external_id: str + Required. The external ID of the service principal in the customer's IdP. + + :returns: :class:`ResolveServicePrincipalResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + external_id: + type: string + description: >- + Required. The external ID of the service principal in the + customer's IdP. + required: + - external_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ResolveServicePrincipalResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIamV2API + /api/2.0/identity/accounts/{account_id}/users/resolveByExternalId: + post: + operationId: account_iam_v2_resolve_user + summary: >- + Resolves a user with the given external ID from the customer's IdP. If + the user does not exist, it + tags: + - iamv2 + - account_iam_v2 + description: >- + Resolves a user with the given external ID from the customer's IdP. If + the user does not exist, it + + will be created. If the customer is not onboarded onto Automatic + Identity Management (AIM), this will + + return an error. + + + :param external_id: str + Required. The external ID of the user in the customer's IdP. + + :returns: :class:`ResolveUserResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + external_id: + type: string + description: Required. The external ID of the user in the customer's IdP. + required: + - external_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ResolveUserResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIamV2API +components: + schemas: + Group: + type: object + properties: + account_id: + type: string + description: The parent account ID for group in Databricks. + external_id: + type: string + description: ExternalId of the group in the customer's IdP. + group_name: + type: string + description: Display name of the group. + internal_id: + type: integer + description: Internal group ID of the group in Databricks. + description: The details of a Group resource. + ResolveGroupResponse: + type: object + properties: + group: + $ref: '#/components/schemas/Group' + ResolveServicePrincipalResponse: + type: object + properties: + service_principal: + $ref: '#/components/schemas/ServicePrincipal' + ResolveUserResponse: + type: object + properties: + user: + $ref: '#/components/schemas/User' + ServicePrincipal: + type: object + properties: + account_id: + type: string + description: The parent account ID for the service principal in Databricks. + account_sp_status: + $ref: '#/components/schemas/State' + description: The activity status of a service principal in a Databricks account. + application_id: + type: string + description: Application ID of the service principal. + display_name: + type: string + description: Display name of the service principal. + external_id: + type: string + description: ExternalId of the service principal in the customer's IdP. + internal_id: + type: integer + description: >- + Internal service principal ID of the service principal in + Databricks. + description: The details of a ServicePrincipal resource. + User: + type: object + properties: + account_id: + type: string + description: The accountId parent of the user in Databricks. + account_user_status: + $ref: '#/components/schemas/State' + description: The activity status of a user in a Databricks account. + external_id: + type: string + description: ExternalId of the user in the customer's IdP. + internal_id: + type: integer + description: Internal userId of the user in Databricks. + name: + $ref: '#/components/schemas/UserName' + username: + type: string + description: Username/email of the user. + description: The details of a User resource. + UserName: + type: object + properties: + family_name: + type: string + given_name: + type: string + WorkspaceAccessDetail: + type: object + properties: + access_type: + $ref: '#/components/schemas/WorkspaceAccessDetailAccessType' + account_id: + type: string + description: >- + The account ID parent of the workspace where the principal has + access. + permissions: + type: array + items: + $ref: '#/components/schemas/WorkspacePermission' + description: The permissions granted to the principal in the workspace. + principal_id: + type: integer + description: The internal ID of the principal (user/sp/group) in Databricks. + principal_type: + $ref: '#/components/schemas/PrincipalType' + status: + $ref: '#/components/schemas/State' + description: >- + The activity status of the principal in the workspace. Not + applicable for groups at the moment. + workspace_id: + type: integer + description: The workspace ID where the principal has access. + description: The details of a principal's access to a workspace. + PrincipalType: + type: string + enum: + - GROUP + - SERVICE_PRINCIPAL + - USER + description: The type of the principal (user/sp/group). + State: + type: string + enum: + - ACTIVE + - INACTIVE + description: >- + The activity status of a user or service principal in a Databricks + account or workspace. + WorkspaceAccessDetailAccessType: + type: string + enum: + - DIRECT + - INDIRECT + description: The type of access the principal has to the workspace. + WorkspaceAccessDetailView: + type: string + enum: + - BASIC + - FULL + description: >- + Controls what fields are returned in the GetWorkspaceAccessDetail + response. + WorkspacePermission: + type: string + enum: + - ADMIN_PERMISSION + - USER_PERMISSION + description: The type of permission a principal has to a workspace (admin/user). + x-stackQL-resources: + account_iam: + id: databricks_account.iamv2.account_iam + name: account_iam + title: Account Iam + methods: + get_workspace_access_detail: + operation: + $ref: >- + #/paths/~1api~12.0~1identity~1accounts~1{account_id}~1workspaces~1{workspace_id}~1workspaceAccessDetails~1{principal_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + resolve_group: + operation: + $ref: >- + #/paths/~1api~12.0~1identity~1accounts~1{account_id}~1groups~1resolveByExternalId/post + response: + mediaType: application/json + openAPIDocKey: '200' + resolve_service_principal: + operation: + $ref: >- + #/paths/~1api~12.0~1identity~1accounts~1{account_id}~1servicePrincipals~1resolveByExternalId/post + response: + mediaType: application/json + openAPIDocKey: '200' + resolve_user: + operation: + $ref: >- + #/paths/~1api~12.0~1identity~1accounts~1{account_id}~1users~1resolveByExternalId/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_iam/methods/get_workspace_access_detail + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/logging.yaml b/providers/src/databricks_account/v00.00.00000/services/logging.yaml deleted file mode 100644 index e7338fa4..00000000 --- a/providers/src/databricks_account/v00.00.00000/services/logging.yaml +++ /dev/null @@ -1,371 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Logging API - description: Log Delivery -servers: -- url: https://accounts.cloud.databricks.com -paths: - /api/2.0/accounts/{account_id}/log-delivery: - post: - operationId: log-delivery-create - externalDocs: - url: https://docs.databricks.com/api/account/logdelivery/create - x-stackQL-resource: log_delivery - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The log delivery configuration creation request succeeded. - content: - application/json: - schema: - type: object - properties: - log_delivery_configuration: - type: object - properties: - account_id: - type: string - config_id: - type: string - creation_time: - type: integer - update_time: - type: integer - log_delivery_status: - type: object - properties: - status: - type: string - message: - type: string - last_attempt_time: - type: string - last_successful_attempt_time: - type: string - config_name: - type: string - status: - type: string - log_type: - type: string - output_format: - type: string - credentials_id: - type: string - storage_configuration_id: - type: string - workspace_ids_filter: - type: array - items: - type: integer - delivery_path_prefix: - type: string - delivery_start_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - log_delivery_configuration: - type: object - example: - log_delivery_configuration: - config_name: string - status: ENABLED - log_type: BILLABLE_USAGE - output_format: CSV - credentials_id: c7814269-df58-4ca3-85e9-f6672ef43d77 - storage_configuration_id: 04aae505-1b1e-4cb9-997d-e1c49282675d - workspace_ids_filter: - - 0 - delivery_path_prefix: string - delivery_start_time: string - description: Properties of the new log delivery configuration. - description: Creates a new Databricks log delivery configuration to enable delivery - of the specified type of logs to your storage location. This requires that - you already created a - get: - operationId: log-delivery-list - externalDocs: - url: https://docs.databricks.com/api/account/logdelivery/list - x-stackQL-resource: log_delivery - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: status - in: query - - name: credentials_id - in: query - - name: storage_configuration_id - in: query - responses: - '200': - description: Log delivery configurations were returned successfully. - content: - application/json: - schema: - type: object - properties: - log_delivery_configurations: - type: array - items: - type: object - properties: - account_id: - type: string - config_id: - type: string - creation_time: - type: integer - update_time: - type: integer - log_delivery_status: - type: object - properties: - status: - type: string - message: - type: string - last_attempt_time: - type: string - last_successful_attempt_time: - type: string - config_name: - type: string - status: - type: string - log_type: - type: string - output_format: - type: string - credentials_id: - type: string - storage_configuration_id: - type: string - workspace_ids_filter: - type: array - items: - type: integer - delivery_path_prefix: - type: string - delivery_start_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.log_delivery_configurations - description: Gets all Databricks log delivery configurations associated with - an account specified by ID. - /api/2.0/accounts/{account_id}/log-delivery/{log_delivery_configuration_id}: - get: - operationId: log-delivery-get - externalDocs: - url: https://docs.databricks.com/api/account/logdelivery/get - x-stackQL-resource: log_delivery - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: log_delivery_configuration_id - required: true - description: uuid - in: path - responses: - '200': - description: The log delivery configuration was successfully returned. - content: - application/json: - schema: - type: object - properties: - log_delivery_configuration: - type: object - properties: - account_id: - type: string - config_id: - type: string - creation_time: - type: integer - update_time: - type: integer - log_delivery_status: - type: object - properties: - status: - type: string - message: - type: string - last_attempt_time: - type: string - last_successful_attempt_time: - type: string - config_name: - type: string - status: - type: string - log_type: - type: string - output_format: - type: string - credentials_id: - type: string - storage_configuration_id: - type: string - workspace_ids_filter: - type: array - items: - type: integer - delivery_path_prefix: - type: string - delivery_start_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.log_delivery_configuration - description: Gets a Databricks log delivery configuration object for an account, - both specified by ID. - patch: - operationId: log-delivery-patchstatus - externalDocs: - url: https://docs.databricks.com/api/account/logdelivery/patchstatus - x-stackQL-resource: log_delivery - x-stackQL-method: patchstatus - x-stackQL-verb: exec - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: log_delivery_configuration_id - required: true - description: uuid - in: path - responses: - '200': - description: The log delivery configuration was successfully updated. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - status: - type: required - example: - status: ENABLED - description: The new status for this log delivery configuration object. - description: Enables or disables a log delivery configuration. Deletion of delivery - configurations is not supported, so disable log delivery configurations that - are no longer needed. Note that you can't re-enable a delivery configuration - if this would violate the delivery configuration limits described under -components: - x-stackQL-resources: - log_delivery: - id: databricks_account.logging.log_delivery - name: log_delivery - title: Log_delivery - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.log_delivery_configurations - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery~1{log_delivery_configuration_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.log_delivery_configuration - patchstatus: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1log-delivery~1{log_delivery_configuration_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/log_delivery/methods/get' - - $ref: '#/components/x-stackQL-resources/log_delivery/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/log_delivery/methods/create' - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/log_delivery/methods/patchstatus' diff --git a/providers/src/databricks_account/v00.00.00000/services/oauth.yaml b/providers/src/databricks_account/v00.00.00000/services/oauth.yaml deleted file mode 100644 index e8457f4d..00000000 --- a/providers/src/databricks_account/v00.00.00000/services/oauth.yaml +++ /dev/null @@ -1,935 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Oauth API - description: OAuth Integrations -servers: -- url: https://accounts.cloud.databricks.com -paths: - /api/2.0/accounts/{account_id}/oauth2/custom-app-integrations: - post: - operationId: custom-app-integrations-create - externalDocs: - url: https://docs.databricks.com/api/account/customappintegration/create - x-stackQL-resource: custom_app_integrations - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - client_id: - type: string - client_secret: - type: string - integration_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - redirect_urls: - type: Array of string - name: - type: string - confidential: - type: boolean - scopes: - type: Array of string - token_access_policy: - type: object - example: - summary: Create Tableau Cloud OAuth App Integration - value: - name: Example Tableau Server - scopes: - - all-apis - - offline_access - token_access_policy: - access_token_ttl_in_minutes: 120 - refresh_token_ttl_in_minutes: 200 - redirect_urls: - - https://example.online.tableau.com/auth/add_oauth_token - confidential: true - description: Create Custom OAuth App Integration. - get: - operationId: custom-app-integrations-list - externalDocs: - url: https://docs.databricks.com/api/account/customappintegration/list - x-stackQL-resource: custom_app_integrations - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - - name: include_creator_username - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - apps: - type: array - items: - type: object - properties: - integration_id: - type: string - client_id: - type: string - redirect_urls: - type: array - items: - type: string - name: - type: string - confidential: - type: boolean - token_access_policy: - type: object - properties: - access_token_ttl_in_minutes: - type: integer - refresh_token_ttl_in_minutes: - type: integer - scopes: - type: array - items: - type: string - created_by: - type: integer - create_time: - type: string - creator_username: - type: string - next_page_token: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - x-stackQL-objectKey: $.apps - description: Get the list of custom OAuth app integrations for the specified - Databricks account - /api/2.0/accounts/{account_id}/oauth2/custom-app-integrations/{integration_id}: - delete: - operationId: custom-app-integrations-delete - externalDocs: - url: https://docs.databricks.com/api/account/customappintegration/delete - x-stackQL-resource: custom_app_integrations - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: integration_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - description: Delete an existing Custom OAuth App Integration. You can retrieve - the custom OAuth app integration via - get: - operationId: custom-app-integrations-get - externalDocs: - url: https://docs.databricks.com/api/account/customappintegration/get - x-stackQL-resource: custom_app_integrations - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: integration_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - integration_id: - type: string - client_id: - type: string - redirect_urls: - type: array - items: - type: string - name: - type: string - confidential: - type: boolean - token_access_policy: - type: object - properties: - access_token_ttl_in_minutes: - type: integer - refresh_token_ttl_in_minutes: - type: integer - scopes: - type: array - items: - type: string - created_by: - type: integer - create_time: - type: string - creator_username: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - description: Gets the Custom OAuth App Integration for the given integration - id. - patch: - operationId: custom-app-integrations-update - externalDocs: - url: https://docs.databricks.com/api/account/customappintegration/update - x-stackQL-resource: custom_app_integrations - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: integration_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - redirect_urls: - type: Array of string - token_access_policy: - type: object - example: - value: - token_access_policy: - access_token_ttl_in_minutes: 120 - refresh_token_ttl_in_minutes: 200 - description: Updates an existing custom OAuth App Integration. You can retrieve - the custom OAuth app integration via - /api/2.0/accounts/{account_id}/oauth2/published-apps: - get: - operationId: published-apps-list - externalDocs: - url: https://docs.databricks.com/api/account/oauthpublishedapps/list - x-stackQL-resource: published_apps - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - apps: - type: array - items: - type: object - properties: - app_id: - type: string - client_id: - type: string - name: - type: string - description: - type: string - is_confidential_client: - type: boolean - redirect_urls: - type: array - items: - type: string - scopes: - type: array - items: - type: string - next_page_token: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - x-stackQL-objectKey: $.apps - description: Get all the available published OAuth apps in Databricks. - /api/2.0/accounts/{account_id}/oauth2/published-app-integrations: - post: - operationId: published-app-integrations-create - externalDocs: - url: https://docs.databricks.com/api/account/publishedappintegration/create - x-stackQL-resource: published_app_integrations - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - integration_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - app_id: - type: string - token_access_policy: - type: object - example: - PowerBI: - summary: Enable PowerBI OAuth Login - value: - app_id: power-bi - TableauDesktop: - summary: Enable Tableau Desktop OAuth Login - value: - app_id: tableau-desktop - description: Create Published OAuth App Integration. - get: - operationId: published-app-integrations-list - externalDocs: - url: https://docs.databricks.com/api/account/publishedappintegration/list - x-stackQL-resource: published_app_integrations - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - apps: - type: array - items: - type: object - properties: - app_id: - type: string - integration_id: - type: string - name: - type: string - token_access_policy: - type: object - properties: - access_token_ttl_in_minutes: - type: integer - refresh_token_ttl_in_minutes: - type: integer - created_by: - type: integer - create_time: - type: string - next_page_token: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - x-stackQL-objectKey: $.apps - description: Get the list of published OAuth app integrations for the specified - Databricks account - /api/2.0/accounts/{account_id}/oauth2/published-app-integrations/{integration_id}: - delete: - operationId: published-app-integrations-delete - externalDocs: - url: https://docs.databricks.com/api/account/publishedappintegration/delete - x-stackQL-resource: published_app_integrations - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: integration_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - description: Delete an existing Published OAuth App Integration. You can retrieve - the published OAuth app integration via - get: - operationId: published-app-integrations-get - externalDocs: - url: https://docs.databricks.com/api/account/publishedappintegration/get - x-stackQL-resource: published_app_integrations - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: integration_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - app_id: - type: string - integration_id: - type: string - name: - type: string - token_access_policy: - type: object - properties: - access_token_ttl_in_minutes: - type: integer - refresh_token_ttl_in_minutes: - type: integer - created_by: - type: integer - create_time: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - description: Gets the Published OAuth App Integration for the given integration - id. - patch: - operationId: published-app-integrations-update - externalDocs: - url: https://docs.databricks.com/api/account/publishedappintegration/update - x-stackQL-resource: published_app_integrations - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: integration_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - '501': - description: Operation is not implemented or is not supported/enabled in - this service. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - token_access_policy: - type: object - example: - value: - token_access_policy: - access_token_ttl_in_minutes: 120 - refresh_token_ttl_in_minutes: 200 - description: Updates an existing published OAuth App Integration. You can retrieve - the published OAuth app integration via - /api/2.0/accounts/{account_id}/servicePrincipals/{service_principal_id}/credentials/secrets: - post: - operationId: service-principal-secrets-create - externalDocs: - url: https://docs.databricks.com/api/account/serviceprincipalsecrets/create - x-stackQL-resource: service_principal_secrets - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: service_principal_id - required: true - description: int64 - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - id: - type: string - secret: - type: string - secret_hash: - type: string - create_time: - type: string - update_time: - type: string - status: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Create a secret for the given service principal. - get: - operationId: service-principal-secrets-list - externalDocs: - url: https://docs.databricks.com/api/account/serviceprincipalsecrets/list - x-stackQL-resource: service_principal_secrets - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: service_principal_id - required: true - description: int64 - in: path - - name: page_token - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - secrets: - type: array - items: - type: object - properties: - id: - type: string - secret_hash: - type: string - create_time: - type: string - update_time: - type: string - status: - type: string - next_page_token: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.secrets - description: List all secrets associated with the given service principal. This - operation only returns information about the secrets themselves and does not - include the secret values. - /api/2.0/accounts/{account_id}/servicePrincipals/{service_principal_id}/credentials/secrets/{secret_id}: - delete: - operationId: service-principal-secrets-delete - externalDocs: - url: https://docs.databricks.com/api/account/serviceprincipalsecrets/delete - x-stackQL-resource: service_principal_secrets - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: string - in: path - - name: service_principal_id - required: true - description: int64 - in: path - - name: secret_id - required: true - description: string - in: path - responses: - '200': - description: The secret was successfully deleted. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Delete a secret from the given service principal. -components: - x-stackQL-resources: - custom_app_integrations: - id: databricks_account.oauth.custom_app_integrations - name: custom_app_integrations - title: Custom_app_integrations - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.apps - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations~1{integration_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations~1{integration_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations~1{integration_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/custom_app_integrations/methods/get' - - $ref: '#/components/x-stackQL-resources/custom_app_integrations/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/custom_app_integrations/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/custom_app_integrations/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/custom_app_integrations/methods/delete' - exec: [] - published_apps: - id: databricks_account.oauth.published_apps - name: published_apps - title: Published_apps - methods: - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-apps/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.apps - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/published_apps/methods/list' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - published_app_integrations: - id: databricks_account.oauth.published_app_integrations - name: published_app_integrations - title: Published_app_integrations - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.apps - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations~1{integration_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations~1{integration_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations~1{integration_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/published_app_integrations/methods/get' - - $ref: '#/components/x-stackQL-resources/published_app_integrations/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/published_app_integrations/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/published_app_integrations/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/published_app_integrations/methods/delete' - exec: [] - service_principal_secrets: - id: databricks_account.oauth.service_principal_secrets - name: service_principal_secrets - title: Service_principal_secrets - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1credentials~1secrets/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1credentials~1secrets/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.secrets - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1credentials~1secrets~1{secret_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/service_principal_secrets/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/service_principal_secrets/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/service_principal_secrets/methods/delete' - exec: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/oauth2.yaml b/providers/src/databricks_account/v00.00.00000/services/oauth2.yaml new file mode 100644 index 00000000..67a854cd --- /dev/null +++ b/providers/src/databricks_account/v00.00.00000/services/oauth2.yaml @@ -0,0 +1,2029 @@ +openapi: 3.0.0 +info: + title: Databricks Oauth2 API (account) + description: >- + OpenAPI specification for the Databricks oauth2 service (account-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.oauth2 +servers: + - url: https://accounts.cloud.databricks.com +paths: + /api/2.0/accounts/{account_id}/federationPolicies: + post: + operationId: account_federation_policy_create + summary: Create account federation policy. + tags: + - oauth2 + - account_federation_policy + description: |- + Create account federation policy. + + :param policy: :class:`FederationPolicy` + :param policy_id: str (optional) + The identifier for the federation policy. The identifier must contain only lowercase alphanumeric + characters, numbers, hyphens, and slashes. If unspecified, the id will be assigned by Databricks. + + :returns: :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: policy_id + in: query + required: false + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + policy: + type: string + description: >- + :param policy_id: str (optional) The identifier for the + federation policy. The identifier must contain only + lowercase alphanumeric characters, numbers, hyphens, and + slashes. If unspecified, the id will be assigned by + Databricks. + required: + - policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountFederationPolicyAPI + get: + operationId: account_federation_policy_list + summary: List account federation policies. + tags: + - oauth2 + - account_federation_policy + description: |- + List account federation policies. + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFederationPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountFederationPolicyAPI + /api/2.0/accounts/{account_id}/federationPolicies/{policy_id}: + delete: + operationId: account_federation_policy_delete + summary: Delete account federation policy. + tags: + - oauth2 + - account_federation_policy + description: |- + Delete account federation policy. + + :param policy_id: str + The identifier for the federation policy. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: policy_id + in: path + required: true + schema: + type: string + description: The identifier for the federation policy. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountFederationPolicyAPI + get: + operationId: account_federation_policy_get + summary: Get account federation policy. + tags: + - oauth2 + - account_federation_policy + description: |- + Get account federation policy. + + :param policy_id: str + The identifier for the federation policy. + + :returns: :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: policy_id + in: path + required: true + schema: + type: string + description: The identifier for the federation policy. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountFederationPolicyAPI + patch: + operationId: account_federation_policy_update + summary: Update account federation policy. + tags: + - oauth2 + - account_federation_policy + description: |- + Update account federation policy. + + :param policy_id: str + The identifier for the federation policy. + :param policy: :class:`FederationPolicy` + :param update_mask: str (optional) + The field mask specifies which fields of the policy to update. To specify multiple fields in the + field mask, use comma as the separator (no space). The special value '*' indicates that all fields + should be updated (full replacement). If unspecified, all fields that are set in the policy provided + in the update request will overwrite the corresponding fields in the existing policy. Example value: + 'description,oidc_policy.audiences'. + + :returns: :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: policy_id + in: path + required: true + schema: + type: string + description: The identifier for the federation policy. + - name: update_mask + in: query + required: false + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + policy: + type: string + description: >- + :param update_mask: str (optional) The field mask specifies + which fields of the policy to update. To specify multiple + fields in the field mask, use comma as the separator (no + space). The special value '*' indicates that all fields + should be updated (full replacement). If unspecified, all + fields that are set in the policy provided in the update + request will overwrite the corresponding fields in the + existing policy. Example value: + 'description,oidc_policy.audiences'. + required: + - policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountFederationPolicyAPI + /api/2.0/accounts/{account_id}/oauth2/custom-app-integrations: + post: + operationId: custom_app_integration_create + summary: Create Custom OAuth App Integration. + tags: + - oauth2 + - custom_app_integration + description: >- + Create Custom OAuth App Integration. + + + You can retrieve the custom OAuth app integration via + :method:CustomAppIntegration/get. + + + :param confidential: bool (optional) + This field indicates whether an OAuth client secret is required to authenticate this client. + :param name: str (optional) + Name of the custom OAuth app + :param redirect_urls: List[str] (optional) + List of OAuth redirect urls + :param scopes: List[str] (optional) + OAuth scopes granted to the application. Supported scopes: all-apis, sql, offline_access, openid, + profile, email. + :param token_access_policy: :class:`TokenAccessPolicy` (optional) + Token access policy + :param user_authorized_scopes: List[str] (optional) + Scopes that will need to be consented by end user to mint the access token. If the user does not + authorize the access token will not be minted. Must be a subset of scopes. + + :returns: :class:`CreateCustomAppIntegrationOutput` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + confidential: + type: string + description: >- + This field indicates whether an OAuth client secret is + required to authenticate this client. + name: + type: string + description: Name of the custom OAuth app + redirect_urls: + type: string + description: List of OAuth redirect urls + scopes: + type: string + description: >- + OAuth scopes granted to the application. Supported scopes: + all-apis, sql, offline_access, openid, profile, email. + token_access_policy: + type: string + description: Token access policy + user_authorized_scopes: + type: string + description: >- + Scopes that will need to be consented by end user to mint + the access token. If the user does not authorize the access + token will not be minted. Must be a subset of scopes. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateCustomAppIntegrationOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CustomAppIntegrationAPI + get: + operationId: custom_app_integration_list + summary: >- + Get the list of custom OAuth app integrations for the specified + Databricks account + tags: + - oauth2 + - custom_app_integration + description: >- + Get the list of custom OAuth app integrations for the specified + Databricks account + + + :param include_creator_username: bool (optional) + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`GetCustomAppIntegrationOutput` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: include_creator_username + in: query + required: false + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`GetCustomAppIntegrationOutput`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetCustomAppIntegrationsOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CustomAppIntegrationAPI + /api/2.0/accounts/{account_id}/oauth2/custom-app-integrations/{integration_id}: + delete: + operationId: custom_app_integration_delete + summary: >- + Delete an existing Custom OAuth App Integration. You can retrieve the + custom OAuth app integration via + tags: + - oauth2 + - custom_app_integration + description: >- + Delete an existing Custom OAuth App Integration. You can retrieve the + custom OAuth app integration via + + :method:CustomAppIntegration/get. + + + :param integration_id: str + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: integration_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CustomAppIntegrationAPI + get: + operationId: custom_app_integration_get + summary: Gets the Custom OAuth App Integration for the given integration id. + tags: + - oauth2 + - custom_app_integration + description: |- + Gets the Custom OAuth App Integration for the given integration id. + + :param integration_id: str + The OAuth app integration ID. + + :returns: :class:`GetCustomAppIntegrationOutput` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: integration_id + in: path + required: true + schema: + type: string + description: The OAuth app integration ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetCustomAppIntegrationOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CustomAppIntegrationAPI + patch: + operationId: custom_app_integration_update + summary: >- + Updates an existing custom OAuth App Integration. You can retrieve the + custom OAuth app integration + tags: + - oauth2 + - custom_app_integration + description: >- + Updates an existing custom OAuth App Integration. You can retrieve the + custom OAuth app integration + + via :method:CustomAppIntegration/get. + + + :param integration_id: str + + :param redirect_urls: List[str] (optional) + List of OAuth redirect urls to be updated in the custom OAuth app integration + :param scopes: List[str] (optional) + List of OAuth scopes to be updated in the custom OAuth app integration, similar to redirect URIs + this will fully replace the existing values instead of appending + :param token_access_policy: :class:`TokenAccessPolicy` (optional) + Token access policy to be updated in the custom OAuth app integration + :param user_authorized_scopes: List[str] (optional) + Scopes that will need to be consented by end user to mint the access token. If the user does not + authorize the access token will not be minted. Must be a subset of scopes. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: integration_id + in: path + required: true + schema: + type: string + description: >- + :param redirect_urls: List[str] (optional) List of OAuth redirect + urls to be updated in the custom OAuth app integration + requestBody: + content: + application/json: + schema: + type: object + properties: + redirect_urls: + type: string + scopes: + type: string + description: >- + List of OAuth scopes to be updated in the custom OAuth app + integration, similar to redirect URIs this will fully + replace the existing values instead of appending + token_access_policy: + type: string + description: >- + Token access policy to be updated in the custom OAuth app + integration + user_authorized_scopes: + type: string + description: >- + Scopes that will need to be consented by end user to mint + the access token. If the user does not authorize the access + token will not be minted. Must be a subset of scopes. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CustomAppIntegrationAPI + /api/2.0/accounts/{account_id}/oauth2/published-apps: + get: + operationId: o_auth_published_apps_list + summary: Get all the available published OAuth apps in Databricks. + tags: + - oauth2 + - o_auth_published_apps + description: |- + Get all the available published OAuth apps in Databricks. + + :param page_size: int (optional) + The max number of OAuth published apps to return in one page. + :param page_token: str (optional) + A token that can be used to get the next page of results. + + :returns: Iterator over :class:`PublishedAppOutput` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: page_size + in: query + required: false + schema: + type: string + description: The max number of OAuth published apps to return in one page. + - name: page_token + in: query + required: false + schema: + type: string + description: A token that can be used to get the next page of results. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPublishedAppsOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: OAuthPublishedAppsAPI + /api/2.0/accounts/{account_id}/oauth2/published-app-integrations: + post: + operationId: published_app_integration_create + summary: Create Published OAuth App Integration. + tags: + - oauth2 + - published_app_integration + description: >- + Create Published OAuth App Integration. + + + You can retrieve the published OAuth app integration via + :method:PublishedAppIntegration/get. + + + :param app_id: str (optional) + App id of the OAuth published app integration. For example power-bi, tableau-deskop + :param token_access_policy: :class:`TokenAccessPolicy` (optional) + Token access policy + + :returns: :class:`CreatePublishedAppIntegrationOutput` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + app_id: + type: string + description: >- + App id of the OAuth published app integration. For example + power-bi, tableau-deskop + token_access_policy: + type: string + description: Token access policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreatePublishedAppIntegrationOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PublishedAppIntegrationAPI + get: + operationId: published_app_integration_list + summary: >- + Get the list of published OAuth app integrations for the specified + Databricks account + tags: + - oauth2 + - published_app_integration + description: >- + Get the list of published OAuth app integrations for the specified + Databricks account + + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`GetPublishedAppIntegrationOutput` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPublishedAppIntegrationsOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PublishedAppIntegrationAPI + /api/2.0/accounts/{account_id}/oauth2/published-app-integrations/{integration_id}: + delete: + operationId: published_app_integration_delete + summary: >- + Delete an existing Published OAuth App Integration. You can retrieve the + published OAuth app + tags: + - oauth2 + - published_app_integration + description: >- + Delete an existing Published OAuth App Integration. You can retrieve the + published OAuth app + + integration via :method:PublishedAppIntegration/get. + + + :param integration_id: str + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: integration_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PublishedAppIntegrationAPI + get: + operationId: published_app_integration_get + summary: Gets the Published OAuth App Integration for the given integration id. + tags: + - oauth2 + - published_app_integration + description: |- + Gets the Published OAuth App Integration for the given integration id. + + :param integration_id: str + + :returns: :class:`GetPublishedAppIntegrationOutput` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: integration_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetPublishedAppIntegrationOutput`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPublishedAppIntegrationOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PublishedAppIntegrationAPI + patch: + operationId: published_app_integration_update + summary: >- + Updates an existing published OAuth App Integration. You can retrieve + the published OAuth app + tags: + - oauth2 + - published_app_integration + description: >- + Updates an existing published OAuth App Integration. You can retrieve + the published OAuth app + + integration via :method:PublishedAppIntegration/get. + + + :param integration_id: str + + :param token_access_policy: :class:`TokenAccessPolicy` (optional) + Token access policy to be updated in the published OAuth app integration + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: integration_id + in: path + required: true + schema: + type: string + description: >- + :param token_access_policy: :class:`TokenAccessPolicy` (optional) + Token access policy to be updated in the published OAuth app + integration + requestBody: + content: + application/json: + schema: + type: object + properties: + token_access_policy: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PublishedAppIntegrationAPI + /api/2.0/accounts/{account_id}/servicePrincipals/{service_principal_id}/federationPolicies: + post: + operationId: service_principal_federation_policy_create + summary: Create account federation policy. + tags: + - oauth2 + - service_principal_federation_policy + description: |- + Create account federation policy. + + :param service_principal_id: int + The service principal id for the federation policy. + :param policy: :class:`FederationPolicy` + :param policy_id: str (optional) + The identifier for the federation policy. The identifier must contain only lowercase alphanumeric + characters, numbers, hyphens, and slashes. If unspecified, the id will be assigned by Databricks. + + :returns: :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: integer + description: The service principal id for the federation policy. + - name: policy_id + in: query + required: false + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + policy: + type: string + description: >- + :param policy_id: str (optional) The identifier for the + federation policy. The identifier must contain only + lowercase alphanumeric characters, numbers, hyphens, and + slashes. If unspecified, the id will be assigned by + Databricks. + required: + - policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalFederationPolicyAPI + get: + operationId: service_principal_federation_policy_list + summary: List account federation policies. + tags: + - oauth2 + - service_principal_federation_policy + description: |- + List account federation policies. + + :param service_principal_id: int + The service principal id for the federation policy. + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: integer + description: The service principal id for the federation policy. + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFederationPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalFederationPolicyAPI + /api/2.0/accounts/{account_id}/servicePrincipals/{service_principal_id}/federationPolicies/{policy_id}: + delete: + operationId: service_principal_federation_policy_delete + summary: Delete account federation policy. + tags: + - oauth2 + - service_principal_federation_policy + description: |- + Delete account federation policy. + + :param service_principal_id: int + The service principal id for the federation policy. + :param policy_id: str + The identifier for the federation policy. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: integer + description: The service principal id for the federation policy. + - name: policy_id + in: path + required: true + schema: + type: string + description: The identifier for the federation policy. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalFederationPolicyAPI + get: + operationId: service_principal_federation_policy_get + summary: Get account federation policy. + tags: + - oauth2 + - service_principal_federation_policy + description: |- + Get account federation policy. + + :param service_principal_id: int + The service principal id for the federation policy. + :param policy_id: str + The identifier for the federation policy. + + :returns: :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: integer + description: The service principal id for the federation policy. + - name: policy_id + in: path + required: true + schema: + type: string + description: The identifier for the federation policy. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalFederationPolicyAPI + patch: + operationId: service_principal_federation_policy_update + summary: Update account federation policy. + tags: + - oauth2 + - service_principal_federation_policy + description: |- + Update account federation policy. + + :param service_principal_id: int + The service principal id for the federation policy. + :param policy_id: str + The identifier for the federation policy. + :param policy: :class:`FederationPolicy` + :param update_mask: str (optional) + The field mask specifies which fields of the policy to update. To specify multiple fields in the + field mask, use comma as the separator (no space). The special value '*' indicates that all fields + should be updated (full replacement). If unspecified, all fields that are set in the policy provided + in the update request will overwrite the corresponding fields in the existing policy. Example value: + 'description,oidc_policy.audiences'. + + :returns: :class:`FederationPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: integer + description: The service principal id for the federation policy. + - name: policy_id + in: path + required: true + schema: + type: string + description: The identifier for the federation policy. + - name: update_mask + in: query + required: false + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + policy: + type: string + description: >- + :param update_mask: str (optional) The field mask specifies + which fields of the policy to update. To specify multiple + fields in the field mask, use comma as the separator (no + space). The special value '*' indicates that all fields + should be updated (full replacement). If unspecified, all + fields that are set in the policy provided in the update + request will overwrite the corresponding fields in the + existing policy. Example value: + 'description,oidc_policy.audiences'. + required: + - policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalFederationPolicyAPI + /api/2.0/accounts/{account_id}/servicePrincipals/{service_principal_id}/credentials/secrets: + post: + operationId: service_principal_secrets_create + summary: Create a secret for the given service principal. + tags: + - oauth2 + - service_principal_secrets + description: |- + Create a secret for the given service principal. + + :param service_principal_id: str + The service principal ID. + :param lifetime: str (optional) + The lifetime of the secret in seconds. If this parameter is not provided, the secret will have a + default lifetime of 730 days (63072000s). + + :returns: :class:`CreateServicePrincipalSecretResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: string + description: The service principal ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + lifetime: + type: string + description: >- + The lifetime of the secret in seconds. If this parameter is + not provided, the secret will have a default lifetime of 730 + days (63072000s). + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateServicePrincipalSecretResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalSecretsAPI + get: + operationId: service_principal_secrets_list + summary: >- + List all secrets associated with the given service principal. This + operation only returns information + tags: + - oauth2 + - service_principal_secrets + description: >- + List all secrets associated with the given service principal. This + operation only returns information + + about the secrets themselves and does not include the secret values. + + + :param service_principal_id: str + The service principal ID. + :param page_size: int (optional) + + :param page_token: str (optional) + An opaque page token which was the `next_page_token` in the response of the previous request to list + the secrets for this service principal. Provide this token to retrieve the next page of secret + entries. When providing a `page_token`, all other parameters provided to the request must match the + previous request. To list all of the secrets for a service principal, it is necessary to continue + requesting pages of entries until the response contains no `next_page_token`. Note that the number + of entries returned must not be used to determine when the listing is complete. + + :returns: Iterator over :class:`SecretInfo` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: string + description: The service principal ID. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + :param page_token: str (optional) An opaque page token which was the + `next_page_token` in the response of the previous request to list + the secrets for this service principal. Provide this token to + retrieve the next page of secret entries. When providing a + `page_token`, all other parameters provided to the request must + match the previous request. To list all of the secrets for a service + principal, it is necessary to continue requesting pages of entries + until the response contains no `next_page_token`. Note that the + number of entries returned must not be used to determine when the + listing is complete. + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListServicePrincipalSecretsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalSecretsAPI + /api/2.0/accounts/{account_id}/servicePrincipals/{service_principal_id}/credentials/secrets/{secret_id}: + delete: + operationId: service_principal_secrets_delete + summary: Delete a secret from the given service principal. + tags: + - oauth2 + - service_principal_secrets + description: |- + Delete a secret from the given service principal. + + :param service_principal_id: str + The service principal ID. + :param secret_id: str + The secret ID. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: service_principal_id + in: path + required: true + schema: + type: string + description: The service principal ID. + - name: secret_id + in: path + required: true + schema: + type: string + description: The secret ID. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalSecretsAPI +components: + schemas: + CreateCustomAppIntegrationOutput: + type: object + properties: + client_id: + type: string + client_secret: + type: string + description: >- + OAuth client-secret generated by the Databricks. If this is a + confidential OAuth app client-secret will be generated. + integration_id: + type: string + description: Unique integration id for the custom OAuth app + CreatePublishedAppIntegrationOutput: + type: object + properties: + integration_id: + type: string + CreateServicePrincipalSecretResponse: + type: object + properties: + create_time: + type: string + expire_time: + type: string + description: >- + UTC time when the secret will expire. If the field is not present, + the secret does not expire. + id: + type: string + description: ID of the secret + secret: + type: string + description: Secret Value + secret_hash: + type: string + description: Secret Hash + status: + type: string + description: Status of the secret + update_time: + type: string + description: UTC time when the secret was updated + DeleteCustomAppIntegrationOutput: + type: object + properties: {} + DeletePublishedAppIntegrationOutput: + type: object + properties: {} + FederationPolicy: + type: object + properties: + create_time: + type: string + description: + type: string + description: Description of the federation policy. + name: + type: string + description: >- + Resource name for the federation policy. Example values include + `accounts//federationPolicies/my-federation-policy` for + Account Federation Policies, and + `accounts//servicePrincipals//federationPolicies/my-federation-policy` + for Service Principal Federation Policies. Typically an output + parameter, which does not need to be specified in create or update + requests. If specified in a request, must match the value in the + request URL. + oidc_policy: + $ref: '#/components/schemas/OidcFederationPolicy' + policy_id: + type: string + description: The ID of the federation policy. Output only. + service_principal_id: + type: integer + description: >- + The service principal ID that this federation policy applies to. + Output only. Only set for service principal federation policies. + uid: + type: string + description: Unique, immutable id of the federation policy. + update_time: + type: string + description: Last update time of the federation policy. + GetCustomAppIntegrationOutput: + type: object + properties: + client_id: + type: string + confidential: + type: boolean + description: >- + This field indicates whether an OAuth client secret is required to + authenticate this client. + create_time: + type: string + created_by: + type: integer + creator_username: + type: string + integration_id: + type: string + description: ID of this custom app + name: + type: string + description: The display name of the custom OAuth app + redirect_urls: + type: array + items: + type: string + description: List of OAuth redirect urls + scopes: + type: array + items: + type: string + token_access_policy: + $ref: '#/components/schemas/TokenAccessPolicy' + description: Token access policy + user_authorized_scopes: + type: array + items: + type: string + description: >- + Scopes that will need to be consented by end user to mint the access + token. If the user does not authorize the access token will not be + minted. Must be a subset of scopes. + GetCustomAppIntegrationsOutput: + type: object + properties: + apps: + type: array + items: + $ref: '#/components/schemas/GetCustomAppIntegrationOutput' + next_page_token: + type: string + GetPublishedAppIntegrationOutput: + type: object + properties: + app_id: + type: string + create_time: + type: string + created_by: + type: integer + integration_id: + type: string + description: Unique integration id for the published OAuth app + name: + type: string + description: Display name of the published OAuth app + token_access_policy: + $ref: '#/components/schemas/TokenAccessPolicy' + description: Token access policy + GetPublishedAppIntegrationsOutput: + type: object + properties: + apps: + type: array + items: + $ref: '#/components/schemas/GetPublishedAppIntegrationOutput' + next_page_token: + type: string + GetPublishedAppsOutput: + type: object + properties: + apps: + type: array + items: + $ref: '#/components/schemas/PublishedAppOutput' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If not + present, there are no more results to show. + ListFederationPoliciesResponse: + type: object + properties: + next_page_token: + type: string + policies: + type: array + items: + $ref: '#/components/schemas/FederationPolicy' + ListServicePrincipalSecretsResponse: + type: object + properties: + next_page_token: + type: string + secrets: + type: array + items: + $ref: '#/components/schemas/SecretInfo' + description: List of the secrets + OidcFederationPolicy: + type: object + properties: + audiences: + type: array + items: + type: string + description: >- + The allowed token audiences, as specified in the 'aud' claim of + federated tokens. The audience identifier is intended to represent + the recipient of the token. Can be any non-empty string value. As + long as the audience in the token matches at least one audience in + the policy, the token is considered a match. If audiences is + unspecified, defaults to your Databricks account id. + issuer: + type: string + description: >- + The required token issuer, as specified in the 'iss' claim of + federated tokens. + jwks_json: + type: string + description: >- + The public keys used to validate the signature of federated tokens, + in JWKS format. Most use cases should not need to specify this + field. If jwks_uri and jwks_json are both unspecified (recommended), + Databricks automatically fetches the public keys from your issuer’s + well known endpoint. Databricks strongly recommends relying on your + issuer’s well known endpoint for discovering public keys. + jwks_uri: + type: string + description: >- + URL of the public keys used to validate the signature of federated + tokens, in JWKS format. Most use cases should not need to specify + this field. If jwks_uri and jwks_json are both unspecified + (recommended), Databricks automatically fetches the public keys from + your issuer’s well known endpoint. Databricks strongly recommends + relying on your issuer’s well known endpoint for discovering public + keys. + subject: + type: string + description: >- + The required token subject, as specified in the subject claim of + federated tokens. Must be specified for service principal federation + policies. Must not be specified for account federation policies. + subject_claim: + type: string + description: >- + The claim that contains the subject of the token. If unspecified, + the default value is 'sub'. + description: >- + Specifies the policy to use for validating OIDC claims in your federated + tokens. + PublishedAppOutput: + type: object + properties: + app_id: + type: string + client_id: + type: string + description: >- + Client ID of the published OAuth app. It is the client_id in the + OAuth flow + description: + type: string + description: Description of the published OAuth app. + is_confidential_client: + type: boolean + description: >- + Whether the published OAuth app is a confidential client. It is + always false for published OAuth apps. + name: + type: string + description: The display name of the published OAuth app. + redirect_urls: + type: array + items: + type: string + description: Redirect URLs of the published OAuth app. + scopes: + type: array + items: + type: string + description: Required scopes for the published OAuth app. + SecretInfo: + type: object + properties: + create_time: + type: string + expire_time: + type: string + description: >- + UTC time when the secret will expire. If the field is not present, + the secret does not expire. + id: + type: string + description: ID of the secret + secret_hash: + type: string + description: Secret Hash + status: + type: string + description: Status of the secret + update_time: + type: string + description: UTC time when the secret was updated + TokenAccessPolicy: + type: object + properties: + absolute_session_lifetime_in_minutes: + type: integer + access_token_ttl_in_minutes: + type: integer + description: access token time to live in minutes + enable_single_use_refresh_tokens: + type: boolean + description: >- + Whether to enable single-use refresh tokens (refresh token + rotation). If this feature is enabled, upon successfully getting a + new access token using a refresh token, Databricks will issue a new + refresh token along with the access token in the response and + invalidate the old refresh token. The client should use the new + refresh token to get access tokens in future requests. + refresh_token_ttl_in_minutes: + type: integer + description: >- + Refresh token time to live in minutes. When single-use refresh + tokens are enabled, this represents the TTL of an individual refresh + token. If the refresh token is used before it expires, a new one is + issued with a renewed individual TTL. + UpdateCustomAppIntegrationOutput: + type: object + properties: {} + UpdatePublishedAppIntegrationOutput: + type: object + properties: {} + x-stackQL-resources: + account_federation_policy: + id: databricks_account.oauth2.account_federation_policy + name: account_federation_policy + title: Account Federation Policy + methods: + account_federation_policy_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1federationPolicies/post + response: + mediaType: application/json + openAPIDocKey: '200' + account_federation_policy_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1federationPolicies/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.policies + account_federation_policy_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1federationPolicies~1{policy_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_federation_policy_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1federationPolicies~1{policy_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + account_federation_policy_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1federationPolicies~1{policy_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_federation_policy/methods/account_federation_policy_get + - $ref: >- + #/components/x-stackQL-resources/account_federation_policy/methods/account_federation_policy_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_federation_policy/methods/account_federation_policy_create + update: + - $ref: >- + #/components/x-stackQL-resources/account_federation_policy/methods/account_federation_policy_update + delete: + - $ref: >- + #/components/x-stackQL-resources/account_federation_policy/methods/account_federation_policy_delete + replace: [] + custom_app_integration: + id: databricks_account.oauth2.custom_app_integration + name: custom_app_integration + title: Custom App Integration + methods: + custom_app_integration_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations/post + response: + mediaType: application/json + openAPIDocKey: '200' + custom_app_integration_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.apps + custom_app_integration_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations~1{integration_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + custom_app_integration_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations~1{integration_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + custom_app_integration_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1custom-app-integrations~1{integration_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/custom_app_integration/methods/custom_app_integration_get + - $ref: >- + #/components/x-stackQL-resources/custom_app_integration/methods/custom_app_integration_list + insert: + - $ref: >- + #/components/x-stackQL-resources/custom_app_integration/methods/custom_app_integration_create + update: + - $ref: >- + #/components/x-stackQL-resources/custom_app_integration/methods/custom_app_integration_update + delete: + - $ref: >- + #/components/x-stackQL-resources/custom_app_integration/methods/custom_app_integration_delete + replace: [] + oauth_published_apps: + id: databricks_account.oauth2.oauth_published_apps + name: oauth_published_apps + title: Oauth Published Apps + methods: + oauth_published_apps_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-apps/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.apps + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/oauth_published_apps/methods/oauth_published_apps_list + insert: [] + update: [] + delete: [] + replace: [] + published_app_integration: + id: databricks_account.oauth2.published_app_integration + name: published_app_integration + title: Published App Integration + methods: + published_app_integration_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations/post + response: + mediaType: application/json + openAPIDocKey: '200' + published_app_integration_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.apps + published_app_integration_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations~1{integration_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + published_app_integration_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations~1{integration_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + published_app_integration_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1oauth2~1published-app-integrations~1{integration_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/published_app_integration/methods/published_app_integration_get + - $ref: >- + #/components/x-stackQL-resources/published_app_integration/methods/published_app_integration_list + insert: + - $ref: >- + #/components/x-stackQL-resources/published_app_integration/methods/published_app_integration_create + update: + - $ref: >- + #/components/x-stackQL-resources/published_app_integration/methods/published_app_integration_update + delete: + - $ref: >- + #/components/x-stackQL-resources/published_app_integration/methods/published_app_integration_delete + replace: [] + service_principal_federation_policy: + id: databricks_account.oauth2.service_principal_federation_policy + name: service_principal_federation_policy + title: Service Principal Federation Policy + methods: + service_principal_federation_policy_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1federationPolicies/post + response: + mediaType: application/json + openAPIDocKey: '200' + service_principal_federation_policy_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1federationPolicies/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.policies + service_principal_federation_policy_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1federationPolicies~1{policy_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + service_principal_federation_policy_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1federationPolicies~1{policy_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + service_principal_federation_policy_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1federationPolicies~1{policy_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/service_principal_federation_policy/methods/service_principal_federation_policy_get + - $ref: >- + #/components/x-stackQL-resources/service_principal_federation_policy/methods/service_principal_federation_policy_list + insert: + - $ref: >- + #/components/x-stackQL-resources/service_principal_federation_policy/methods/service_principal_federation_policy_create + update: + - $ref: >- + #/components/x-stackQL-resources/service_principal_federation_policy/methods/service_principal_federation_policy_update + delete: + - $ref: >- + #/components/x-stackQL-resources/service_principal_federation_policy/methods/service_principal_federation_policy_delete + replace: [] + service_principal_secrets: + id: databricks_account.oauth2.service_principal_secrets + name: service_principal_secrets + title: Service Principal Secrets + methods: + service_principal_secrets_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1credentials~1secrets/post + response: + mediaType: application/json + openAPIDocKey: '200' + service_principal_secrets_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1credentials~1secrets/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.secrets + service_principal_secrets_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1servicePrincipals~1{service_principal_id}~1credentials~1secrets~1{secret_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/service_principal_secrets/methods/service_principal_secrets_list + insert: + - $ref: >- + #/components/x-stackQL-resources/service_principal_secrets/methods/service_principal_secrets_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/service_principal_secrets/methods/service_principal_secrets_delete + replace: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/provisioning.yaml b/providers/src/databricks_account/v00.00.00000/services/provisioning.yaml index ab254281..1bb46d2d 100644 --- a/providers/src/databricks_account/v00.00.00000/services/provisioning.yaml +++ b/providers/src/databricks_account/v00.00.00000/services/provisioning.yaml @@ -1,2158 +1,3372 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Provisioning API - description: Account Provisioning -servers: -- url: https://accounts.cloud.databricks.com -paths: - /api/2.0/accounts/{account_id}/credentials: - post: - operationId: credentials-create - externalDocs: - url: https://docs.databricks.com/api/account/credentials/create - x-stackQL-resource: credentials - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The credential configuration creation request succeeded. - content: - application/json: - schema: - type: object - properties: - credentials_id: - type: string - credentials_name: - type: string - aws_credentials: - type: object - properties: - sts_role: - type: object - properties: - role_arn: - type: string - external_id: - type: string - account_id: - type: string - creation_time: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - credentials_name: - type: required - aws_credentials: - type: required - example: - credentials_name: credential_1 - aws_credentials: - sts_role: - role_arn: arn-aws-iam::111110000000:role/test_role - description: Properties of the new credential configuration. - description: Creates a Databricks credential configuration that represents cloud - cross-account credentials for a specified account. Databricks uses this to - set up network infrastructure properly to host Databricks clusters. For your - AWS IAM role, you need to trust the External ID (the Databricks Account API - account ID) in the returned credential object, and configure the required - access policy. - get: - operationId: credentials-list - externalDocs: - url: https://docs.databricks.com/api/account/credentials/list - x-stackQL-resource: credentials - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: Credential configurations were returned successfully. - content: - application/json: - schema: - type: array - items: - type: object - properties: - credentials_id: - type: string - credentials_name: - type: string - aws_credentials: - type: object - properties: - sts_role: - type: object - properties: - role_arn: - type: string - external_id: - type: string - account_id: - type: string - creation_time: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets all Databricks credential configurations associated with an - account specified by ID. - /api/2.0/accounts/{account_id}/credentials/{credentials_id}: - delete: - operationId: credentials-delete - externalDocs: - url: https://docs.databricks.com/api/account/credentials/delete - x-stackQL-resource: credentials - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: credentials_id - required: true - description: uuid - in: path - responses: - '200': - description: The credential configuration was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes a Databricks credential configuration object for an account, - both specified by ID. You cannot delete a credential that is associated with - any workspace. - get: - operationId: credentials-get - externalDocs: - url: https://docs.databricks.com/api/account/credentials/get - x-stackQL-resource: credentials - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: credentials_id - required: true - description: uuid - in: path - responses: - '200': - description: The credential configuration was successfully returned. - content: - application/json: - schema: - type: object - properties: - credentials_id: - type: string - credentials_name: - type: string - aws_credentials: - type: object - properties: - sts_role: - type: object - properties: - role_arn: - type: string - external_id: - type: string - account_id: - type: string - creation_time: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a Databricks credential configuration object for an account, - both specified by ID. - /api/2.0/accounts/{account_id}/customer-managed-keys: - post: - operationId: encryption-keys-create - externalDocs: - url: https://docs.databricks.com/api/account/encryptionkeys/create - x-stackQL-resource: encryption_keys - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The encryption key configuration was successfully created. - content: - application/json: - schema: - type: object - properties: - customer_managed_key_id: - type: string - account_id: - type: string - aws_key_info: - type: object - properties: - key_arn: - type: string - key_alias: - type: string - key_region: - type: string - reuse_key_for_cluster_volumes: - type: boolean - creation_time: - type: integer - use_cases: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - use_cases: - type: required - aws_key_info: - type: object - example: - aws_key_info: - key_arn: arn:aws:kms:us-west-2:111122223333:key/0987dcba-09fe-87dc-65ba-ab0987654321 - key_alias: alias/projectKey - reuse_key_for_cluster_volumes: true - use_cases: - - MANAGED_SERVICES - - STORAGE - description: Properties of the encryption key configuration. - description: Creates a customer-managed key configuration object for an account, - specified by ID. This operation uploads a reference to a customer-managed - key to Databricks. If the key is assigned as a workspace's customer-managed - key for managed services, Databricks uses the key to encrypt the workspaces - notebooks and secrets in the control plane, in addition to Databricks SQL - queries and query history. If it is specified as a workspace's customer-managed - key for workspace storage, the key encrypts the workspace's root S3 bucket - (which contains the workspace's root DBFS and system data) and, optionally, - cluster EBS volume data. - get: - operationId: encryption-keys-list - externalDocs: - url: https://docs.databricks.com/api/account/encryptionkeys/list - x-stackQL-resource: encryption_keys - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The encryption key configurations were successfully returned. - content: - application/json: - schema: - type: array - items: - type: object - properties: - customer_managed_key_id: - type: string - account_id: - type: string - aws_key_info: - type: object - properties: - key_arn: - type: string - key_alias: - type: string - key_region: - type: string - reuse_key_for_cluster_volumes: - type: boolean - creation_time: - type: integer - use_cases: - type: array - items: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets all customer-managed key configuration objects for an account. - If the key is specified as a workspace's managed services customer-managed - key, Databricks uses the key to encrypt the workspace's notebooks and secrets - in the control plane, in addition to Databricks SQL queries and query history. - If the key is specified as a workspace's storage customer-managed key, the - key is used to encrypt the workspace's root S3 bucket and optionally can encrypt - cluster EBS volumes data in the data plane. - /api/2.0/accounts/{account_id}/customer-managed-keys/{customer_managed_key_id}: - delete: - operationId: encryption-keys-delete - externalDocs: - url: https://docs.databricks.com/api/account/encryptionkeys/delete - x-stackQL-resource: encryption_keys - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: customer_managed_key_id - required: true - description: uuid - in: path - responses: - '200': - description: The encryption key configuration was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes a customer-managed key configuration object for an account. - You cannot delete a configuration that is associated with a running workspace. - get: - operationId: encryption-keys-get - externalDocs: - url: https://docs.databricks.com/api/account/encryptionkeys/get - x-stackQL-resource: encryption_keys - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: customer_managed_key_id - required: true - description: uuid - in: path - responses: - '200': - description: The encryption key configuration was successfully returned. - content: - application/json: - schema: - type: object - properties: - customer_managed_key_id: - type: string - account_id: - type: string - aws_key_info: - type: object - properties: - key_arn: - type: string - key_alias: - type: string - key_region: - type: string - reuse_key_for_cluster_volumes: - type: boolean - creation_time: - type: integer - use_cases: - type: array - items: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a customer-managed key configuration object for an account, - specified by ID. This operation uploads a reference to a customer-managed - key to Databricks. If assigned as a workspace's customer-managed key for managed - services, Databricks uses the key to encrypt the workspaces notebooks and - secrets in the control plane, in addition to Databricks SQL queries and query - history. If it is specified as a workspace's customer-managed key for storage, - the key encrypts the workspace's root S3 bucket (which contains the workspace's - root DBFS and system data) and, optionally, cluster EBS volume data. - /api/2.0/accounts/{account_id}/networks: - post: - operationId: networks-create - externalDocs: - url: https://docs.databricks.com/api/account/networks/create - x-stackQL-resource: networks - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The network configuration was successfully created. - content: - application/json: - schema: - type: object - properties: - network_id: - type: string - network_name: - type: string - vpc_id: - type: string - subnet_ids: - type: array - items: - type: string - security_group_ids: - type: array - items: - type: string - vpc_status: - type: string - warning_messages: - type: array - items: - type: object - properties: - warning_type: - type: string - warning_message: - type: string - error_messages: - type: array - items: - type: object - properties: - error_type: - type: string - error_message: - type: string - workspace_id: - type: integer - account_id: - type: string - creation_time: - type: integer - vpc_endpoints: - type: object - properties: - rest_api: - type: array - items: - type: string - dataplane_relay: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - network_name: - type: required - vpc_id: - type: string - subnet_ids: - type: string - security_group_ids: - type: Array of string - vpc_endpoints: - type: object - example: - network_name: string - vpc_id: string - subnet_ids: - - string - security_group_ids: - - string - vpc_endpoints: - rest_api: - - 497f6eca-6276-4993-bfeb-53cbbbba6f08 - dataplane_relay: - - 497f6eca-6276-4993-bfeb-53cbbbba6f08 - description: Properties of the new network configuration. - description: Creates a Databricks network configuration that represents an AWS - VPC and its resources. The VPC will be used for new Databricks clusters. This - requires a pre-existing VPC and subnets. For VPC requirements, see - get: - operationId: networks-list - externalDocs: - url: https://docs.databricks.com/api/account/networks/list - x-stackQL-resource: networks - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The network configurations were successfully returned. - content: - application/json: - schema: - type: array - items: - type: object - properties: - network_id: - type: string - network_name: - type: string - vpc_id: - type: string - subnet_ids: - type: array - items: - type: string - security_group_ids: - type: array - items: - type: string - vpc_status: - type: string - warning_messages: - type: array - items: - type: object - properties: - warning_type: - type: string - warning_message: - type: string - error_messages: - type: array - items: - type: object - properties: - error_type: - type: string - error_message: - type: string - workspace_id: - type: integer - account_id: - type: string - creation_time: - type: integer - vpc_endpoints: - type: object - properties: - rest_api: - type: array - items: - type: string - dataplane_relay: - type: array - items: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a list of all Databricks network configurations for an account, - specified by ID. - /api/2.0/accounts/{account_id}/networks/{network_id}: - delete: - operationId: networks-delete - externalDocs: - url: https://docs.databricks.com/api/account/networks/delete - x-stackQL-resource: networks - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: network_id - required: true - description: uuid - in: path - responses: - '200': - description: The network configuration was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes a Databricks network configuration, which represents a - cloud VPC and its resources. You cannot delete a network that is associated - with a workspace. - get: - operationId: networks-get - externalDocs: - url: https://docs.databricks.com/api/account/networks/get - x-stackQL-resource: networks - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: network_id - required: true - description: uuid - in: path - responses: - '200': - description: The network configuration was successfully returned. - content: - application/json: - schema: - type: object - properties: - network_id: - type: string - network_name: - type: string - vpc_id: - type: string - subnet_ids: - type: array - items: - type: string - security_group_ids: - type: array - items: - type: string - vpc_status: - type: string - warning_messages: - type: array - items: - type: object - properties: - warning_type: - type: string - warning_message: - type: string - error_messages: - type: array - items: - type: object - properties: - error_type: - type: string - error_message: - type: string - workspace_id: - type: integer - account_id: - type: string - creation_time: - type: integer - vpc_endpoints: - type: object - properties: - rest_api: - type: array - items: - type: string - dataplane_relay: - type: array - items: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a Databricks network configuration, which represents an AWS - VPC and its resources. This requires a pre-existing VPC and subnets. - /api/2.0/accounts/{account_id}/private-access-settings: - post: - operationId: private-access-create - externalDocs: - url: https://docs.databricks.com/api/account/privateaccess/create - x-stackQL-resource: private_access - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The private access settings object was successfully created. - content: - application/json: - schema: - type: object - properties: - private_access_settings_id: - type: string - private_access_settings_name: - type: string - public_access_enabled: - type: boolean - region: - type: string - account_id: - type: string - private_access_level: - type: string - allowed_vpc_endpoint_ids: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - private_access_settings_name: - type: required - region: - type: string - public_access_enabled: - type: required - private_access_level: - type: string - allowed_vpc_endpoint_ids: - type: boolean - example: - private_access_settings_name: string - region: string - public_access_enabled: false - private_access_level: ENDPOINT - allowed_vpc_endpoint_ids: - - 497f6eca-6276-4993-bfeb-53cbbbba6f08 - description: Properties of the new private access settings object. - description: Creates a private access settings object, which specifies how your - workspace is accessed over - get: - operationId: private-access-list - externalDocs: - url: https://docs.databricks.com/api/account/privateaccess/list - x-stackQL-resource: private_access - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The private access settings object was successfully returned. - content: - application/json: - schema: - type: array - items: - type: object - properties: - private_access_settings_id: - type: string - private_access_settings_name: - type: string - public_access_enabled: - type: boolean - region: - type: string - account_id: - type: string - private_access_level: - type: string - allowed_vpc_endpoint_ids: - type: array - items: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a list of all private access settings objects for an account, - specified by ID. - /api/2.0/accounts/{account_id}/private-access-settings/{private_access_settings_id}: - delete: - operationId: private-access-delete - externalDocs: - url: https://docs.databricks.com/api/account/privateaccess/delete - x-stackQL-resource: private_access - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: private_access_settings_id - required: true - description: uuid - in: path - responses: - '200': - description: The private access settings was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes a private access settings object, which determines how - your workspace is accessed over - get: - operationId: private-access-get - externalDocs: - url: https://docs.databricks.com/api/account/privateaccess/get - x-stackQL-resource: private_access - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: private_access_settings_id - required: true - description: uuid - in: path - responses: - '200': - description: The private access settings object was successfully returned. - content: - application/json: - schema: - type: object - properties: - private_access_settings_id: - type: string - private_access_settings_name: - type: string - public_access_enabled: - type: boolean - region: - type: string - account_id: - type: string - private_access_level: - type: string - allowed_vpc_endpoint_ids: - type: array - items: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a private access settings object, which specifies how your - workspace is accessed over - put: - operationId: private-access-replace - externalDocs: - url: https://docs.databricks.com/api/account/privateaccess/replace - x-stackQL-resource: private_access - x-stackQL-method: replace - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: private_access_settings_id - required: true - description: uuid - in: path - responses: - '200': - description: The private access settings object was successfully updated. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - private_access_settings_name: - type: required - region: - type: string - public_access_enabled: - type: required - private_access_level: - type: string - allowed_vpc_endpoint_ids: - type: boolean - example: - private_access_settings_name: string - region: string - public_access_enabled: false - private_access_level: ENDPOINT - allowed_vpc_endpoint_ids: - - 497f6eca-6276-4993-bfeb-53cbbbba6f08 - description: Properties of the new private access settings object. - description: Updates an existing private access settings object, which specifies - how your workspace is accessed over - /api/2.0/accounts/{account_id}/storage-configurations: - post: - operationId: storage-create - externalDocs: - url: https://docs.databricks.com/api/account/storage/create - x-stackQL-resource: storage - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The storage configuration was successfully created. - content: - application/json: - schema: - type: object - properties: - storage_configuration_id: - type: string - storage_configuration_name: - type: string - root_bucket_info: - type: object - properties: - bucket_name: - type: string - account_id: - type: string - creation_time: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - storage_configuration_name: - type: required - root_bucket_info: - type: required - example: - storage_configuration_name: storage_conf_1 - root_bucket_info: - bucket_name: string - description: Properties of the new storage configuration. - description: Creates new storage configuration for an account, specified by - ID. Uploads a storage configuration object that represents the root AWS S3 - bucket in your account. Databricks stores related workspace assets including - DBFS, cluster logs, and job results. For the AWS S3 bucket, you need to configure - the required bucket policy. - get: - operationId: storage-list - externalDocs: - url: https://docs.databricks.com/api/account/storage/list - x-stackQL-resource: storage - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The storage configurations were successfully returned. - content: - application/json: - schema: - type: array - items: - type: object - properties: - storage_configuration_id: - type: string - storage_configuration_name: - type: string - root_bucket_info: - type: object - properties: - bucket_name: - type: string - account_id: - type: string - creation_time: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a list of all Databricks storage configurations for your account, - specified by ID. - /api/2.0/accounts/{account_id}/storage-configurations/{storage_configuration_id}: - delete: - operationId: storage-delete - externalDocs: - url: https://docs.databricks.com/api/account/storage/delete - x-stackQL-resource: storage - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: storage_configuration_id - required: true - description: uuid - in: path - responses: - '200': - description: The storage configuration was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes a Databricks storage configuration. You cannot delete a - storage configuration that is associated with any workspace. - get: - operationId: storage-get - externalDocs: - url: https://docs.databricks.com/api/account/storage/get - x-stackQL-resource: storage - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: storage_configuration_id - required: true - description: uuid - in: path - responses: - '200': - description: The storage configuration was successfully returned. - content: - application/json: - schema: - type: object - properties: - storage_configuration_id: - type: string - storage_configuration_name: - type: string - root_bucket_info: - type: object - properties: - bucket_name: - type: string - account_id: - type: string - creation_time: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a Databricks storage configuration for an account, both specified - by ID. - /api/2.0/accounts/{account_id}/vpc-endpoints: - post: - operationId: vpc-endpoints-create - externalDocs: - url: https://docs.databricks.com/api/account/vpcendpoints/create - x-stackQL-resource: vpc_endpoints - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The VPC endpoint configuration was successfully created. - content: - application/json: - schema: - type: object - properties: - vpc_endpoint_id: - type: string - vpc_endpoint_name: - type: string - aws_vpc_endpoint_id: - type: string - aws_endpoint_service_id: - type: string - use_case: - type: string - region: - type: string - account_id: - type: string - aws_account_id: - type: string - state: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - vpc_endpoint_name: - type: required - aws_vpc_endpoint_id: - type: string - region: - type: string - example: - vpc_endpoint_name: string - aws_vpc_endpoint_id: string - region: string - description: Properties of the new VPC endpoint configuration. - description: Creates a VPC endpoint configuration, which represents a - get: - operationId: vpc-endpoints-list - externalDocs: - url: https://docs.databricks.com/api/account/vpcendpoints/list - x-stackQL-resource: vpc_endpoints - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The VPC endpoints were successfully returned. - content: - application/json: - schema: - type: array - items: - type: object - properties: - vpc_endpoint_id: - type: string - vpc_endpoint_name: - type: string - aws_vpc_endpoint_id: - type: string - aws_endpoint_service_id: - type: string - use_case: - type: string - region: - type: string - account_id: - type: string - aws_account_id: - type: string - state: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a list of all VPC endpoints for an account, specified by ID. - /api/2.0/accounts/{account_id}/vpc-endpoints/{vpc_endpoint_id}: - delete: - operationId: vpc-endpoints-delete - externalDocs: - url: https://docs.databricks.com/api/account/vpcendpoints/delete - x-stackQL-resource: vpc_endpoints - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: vpc_endpoint_id - required: true - description: uuid - in: path - responses: - '200': - description: The VPC endpoint configuration was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes a VPC endpoint configuration, which represents an - get: - operationId: vpc-endpoints-get - externalDocs: - url: https://docs.databricks.com/api/account/vpcendpoints/get - x-stackQL-resource: vpc_endpoints - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: vpc_endpoint_id - required: true - description: uuid - in: path - responses: - '200': - description: The VPC endpoint was successfully returned. - content: - application/json: - schema: - type: object - properties: - vpc_endpoint_id: - type: string - vpc_endpoint_name: - type: string - aws_vpc_endpoint_id: - type: string - aws_endpoint_service_id: - type: string - use_case: - type: string - region: - type: string - account_id: - type: string - aws_account_id: - type: string - state: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a VPC endpoint configuration, which represents a - /api/2.0/accounts/{account_id}/workspaces: - post: - operationId: workspaces-create - externalDocs: - url: https://docs.databricks.com/api/account/workspaces/create - x-stackQL-resource: workspaces - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: Workspace creation request was received. Check workspace status. - content: - application/json: - schema: - type: object - properties: - workspace_id: - type: integer - workspace_name: - type: string - deployment_name: - type: string - network_id: - type: string - aws_region: - type: string - credentials_id: - type: string - storage_configuration_id: - type: string - account_id: - type: string - workspace_status: - type: string - workspace_status_message: - type: string - managed_services_customer_managed_key_id: - type: string - private_access_settings_id: - type: string - creation_time: - type: integer - pricing_tier: - type: string - storage_customer_managed_key_id: - type: string - custom_tags: - type: object - properties: - property1: - type: string - property2: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - workspace_name: - type: required - network_id: - type: string - deployment_name: - type: uuid - aws_region: - type: string - credentials_id: - type: string - storage_configuration_id: - type: uuid - managed_services_customer_managed_key_id: - type: uuid - private_access_settings_id: - type: uuid - pricing_tier: - type: uuid - storage_customer_managed_key_id: - type: string - custom_tags: - type: object - example: - workspace_name: string - network_id: fd0cc5bc-683c-47e9-b15e-144d7744a496 - deployment_name: workspace_1 - aws_region: us-west-2 - credentials_id: ccc64f28-ebdc-4c89-add9-5dcb6d7727d8 - storage_configuration_id: b43a6064-04c1-4e1c-88b6-d91e5b136b13 - managed_services_customer_managed_key_id: 849b3d6b-e68e-468d-b3e5-deb08b03c56d - private_access_settings_id: 3b3bbcb5-46bd-4b03-944e-97eb44ed7991 - pricing_tier: PREMIUM - storage_customer_managed_key_id: 14138d0f-a575-4ae2-be71-ddfd0b602286 - custom_tags: - property1: string - property2: string - description: Properties of the new workspace. - description: Creates a new workspace using a credential configuration and a - storage configuration, an optional network configuration (if using a customer-managed - VPC), an optional managed services key configuration (if using customer-managed - keys for managed services), and an optional storage key configuration (if - using customer-managed keys for storage). The key configurations used for - managed services and storage encryption can be the same or different. - get: - operationId: workspaces-list - externalDocs: - url: https://docs.databricks.com/api/account/workspaces/list - x-stackQL-resource: workspaces - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The workspaces were returned successfully. - content: - application/json: - schema: - type: array - items: - type: object - properties: - workspace_id: - type: integer - workspace_name: - type: string - deployment_name: - type: string - network_id: - type: string - aws_region: - type: string - credentials_id: - type: string - storage_configuration_id: - type: string - account_id: - type: string - workspace_status: - type: string - workspace_status_message: - type: string - managed_services_customer_managed_key_id: - type: string - private_access_settings_id: - type: string - creation_time: - type: integer - pricing_tier: - type: string - storage_customer_managed_key_id: - type: string - custom_tags: - type: object - properties: - property1: - type: string - property2: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a list of all workspaces associated with an account, specified - by ID. - /api/2.0/accounts/{account_id}/workspaces/{workspace_id}: - delete: - operationId: workspaces-delete - externalDocs: - url: https://docs.databricks.com/api/account/workspaces/delete - x-stackQL-resource: workspaces - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: The workspace was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Terminates and deletes a Databricks workspace. From an API perspective, - deletion is immediate. However, it might take a few minutes for all workspaces - resources to be deleted, depending on the size and number of workspace resources. - get: - operationId: workspaces-get - externalDocs: - url: https://docs.databricks.com/api/account/workspaces/get - x-stackQL-resource: workspaces - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: The workspace configuration was successfully returned. - content: - application/json: - schema: - type: object - properties: - workspace_id: - type: integer - workspace_name: - type: string - deployment_name: - type: string - network_id: - type: string - aws_region: - type: string - credentials_id: - type: string - storage_configuration_id: - type: string - account_id: - type: string - workspace_status: - type: string - workspace_status_message: - type: string - managed_services_customer_managed_key_id: - type: string - private_access_settings_id: - type: string - creation_time: - type: integer - pricing_tier: - type: string - storage_customer_managed_key_id: - type: string - custom_tags: - type: object - properties: - property1: - type: string - property2: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets information including status for a Databricks workspace, specified - by ID. In the response, the - patch: - operationId: workspaces-update - externalDocs: - url: https://docs.databricks.com/api/account/workspaces/update - x-stackQL-resource: workspaces - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: The workspace update request is accepted. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - '509': - description: An external service is unavailable temporarily as it is being - updated. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - aws_region: - type: string - credentials_id: - type: uuid - storage_configuration_id: - type: uuid - network_id: - type: uuid - managed_services_customer_managed_key_id: - type: uuid - private_access_settings_id: - type: uuid - storage_customer_managed_key_id: - type: uuid - network_connectivity_config_id: - type: uuid - custom_tags: - type: object - example: - aws_region: us-west-2 - credentials_id: c7814269-df58-4ca3-85e9-f6672ef43d77 - storage_configuration_id: 04aae505-1b1e-4cb9-997d-e1c49282675d - network_id: d6797cf4-42b9-4cad-8591-9dd91c3f0fc3 - managed_services_customer_managed_key_id: faacdc79-6530-4583-a154-5d427a663e53 - private_access_settings_id: 3b3bbcb5-46bd-4b03-944e-97eb44ed7991 - storage_customer_managed_key_id: 14138d0f-a575-4ae2-be71-ddfd0b602286 - custom_tags: - property1: string - property2: string - network_connectivity_config_id: 390afeb4-1111-2222-af33-ca3965c04068 - description: Changes of the workspace properties. - description: Updates a workspace configuration for either a running workspace - or a failed workspace. The elements that can be updated varies between these - two use cases. -components: - x-stackQL-resources: - credentials: - id: databricks_account.provisioning.credentials - name: credentials - title: Credentials - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1credentials/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1credentials~1{credentials_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1credentials~1{credentials_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/credentials/methods/get' - - $ref: '#/components/x-stackQL-resources/credentials/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/credentials/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/credentials/methods/delete' - exec: [] - encryption_keys: - id: databricks_account.provisioning.encryption_keys - name: encryption_keys - title: Encryption_keys - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys~1{customer_managed_key_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys~1{customer_managed_key_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/encryption_keys/methods/get' - - $ref: '#/components/x-stackQL-resources/encryption_keys/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/encryption_keys/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/encryption_keys/methods/delete' - exec: [] - networks: - id: databricks_account.provisioning.networks - name: networks - title: Networks - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1networks/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1networks/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1networks~1{network_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1networks~1{network_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/networks/methods/get' - - $ref: '#/components/x-stackQL-resources/networks/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/networks/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/networks/methods/delete' - exec: [] - private_access: - id: databricks_account.provisioning.private_access - name: private_access - title: Private_access - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings~1{private_access_settings_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings~1{private_access_settings_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - replace: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings~1{private_access_settings_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/private_access/methods/get' - - $ref: '#/components/x-stackQL-resources/private_access/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/private_access/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/private_access/methods/replace' - delete: - - $ref: '#/components/x-stackQL-resources/private_access/methods/delete' - exec: [] - storage: - id: databricks_account.provisioning.storage - name: storage - title: Storage - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations~1{storage_configuration_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations~1{storage_configuration_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/storage/methods/get' - - $ref: '#/components/x-stackQL-resources/storage/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/storage/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/storage/methods/delete' - exec: [] - vpc_endpoints: - id: databricks_account.provisioning.vpc_endpoints - name: vpc_endpoints - title: Vpc_endpoints - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints~1{vpc_endpoint_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints~1{vpc_endpoint_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/vpc_endpoints/methods/get' - - $ref: '#/components/x-stackQL-resources/vpc_endpoints/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/vpc_endpoints/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/vpc_endpoints/methods/delete' - exec: [] - workspaces: - id: databricks_account.provisioning.workspaces - name: workspaces - title: Workspaces - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/workspaces/methods/get' - - $ref: '#/components/x-stackQL-resources/workspaces/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/workspaces/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/workspaces/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/workspaces/methods/delete' - exec: [] +openapi: 3.0.0 +info: + title: Databricks Provisioning API (account) + description: >- + OpenAPI specification for the Databricks provisioning service (account-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.provisioning +servers: + - url: https://accounts.cloud.databricks.com +paths: + /api/2.0/accounts/{account_id}/credentials: + post: + operationId: credentials_create + summary: >- + Creates a Databricks credential configuration that represents cloud + cross-account credentials for a + tags: + - provisioning + - credentials + description: >- + Creates a Databricks credential configuration that represents cloud + cross-account credentials for a + + specified account. Databricks uses this to set up network infrastructure + properly to host Databricks + + clusters. For your AWS IAM role, you need to trust the External ID (the + Databricks Account API account + + ID) in the returned credential object, and configure the required access + policy. + + + Save the response's `credentials_id` field, which is the ID for your new + credential configuration + + object. + + + For information about how to create a new workspace with this API, see + [Create a new workspace using + + the Account API] + + + [Create a new workspace using the Account API]: + http://docs.databricks.com/administration-guide/account-api/new-workspace.html + + + :param credentials_name: str + The human-readable name of the credential configuration object. + :param aws_credentials: :class:`CreateCredentialAwsCredentials` + + + :returns: :class:`Credential` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + credentials_name: + type: string + description: >- + The human-readable name of the credential configuration + object. + aws_credentials: + type: string + description: ':returns: :class:`Credential`' + required: + - credentials_name + - aws_credentials + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Credential' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + get: + operationId: credentials_list + summary: >- + List Databricks credential configuration objects for an account, + specified by ID. + tags: + - provisioning + - credentials + description: >- + List Databricks credential configuration objects for an account, + specified by ID. + + + + :returns: Iterator over :class:`Credential` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Credential' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + /api/2.0/accounts/{account_id}/credentials/{credentials_id}: + delete: + operationId: credentials_delete + summary: >- + Deletes a Databricks credential configuration object for an account, + both specified by ID. You cannot + tags: + - provisioning + - credentials + description: >- + Deletes a Databricks credential configuration object for an account, + both specified by ID. You cannot + + delete a credential that is associated with any workspace. + + + :param credentials_id: str + Databricks Account API credential configuration ID + + :returns: :class:`Credential` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: credentials_id + in: path + required: true + schema: + type: string + description: Databricks Account API credential configuration ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Credential' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + get: + operationId: credentials_get + summary: >- + Gets a Databricks credential configuration object for an account, both + specified by ID. + tags: + - provisioning + - credentials + description: >- + Gets a Databricks credential configuration object for an account, both + specified by ID. + + + :param credentials_id: str + Credential configuration ID + + :returns: :class:`Credential` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: credentials_id + in: path + required: true + schema: + type: string + description: Credential configuration ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Credential' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsAPI + /api/2.0/accounts/{account_id}/customer-managed-keys: + post: + operationId: encryption_keys_create + summary: >- + Creates a customer-managed key configuration object for an account, + specified by ID. This operation + tags: + - provisioning + - encryption_keys + description: >- + Creates a customer-managed key configuration object for an account, + specified by ID. This operation + + uploads a reference to a customer-managed key to Databricks. If the key + is assigned as a workspace's + + customer-managed key for managed services, Databricks uses the key to + encrypt the workspaces notebooks + + and secrets in the control plane, in addition to Databricks SQL queries + and query history. If it is + + specified as a workspace's customer-managed key for workspace storage, + the key encrypts the + + workspace's root S3 bucket (which contains the workspace's root DBFS and + system data) and, optionally, + + cluster EBS volume data. + + + **Important**: Customer-managed keys are supported only for some + deployment types, subscription types, + + and AWS regions that currently support creation of Databricks + workspaces. + + + This operation is available only if your account is on the E2 version of + the platform or on a select + + custom plan that allows multiple workspaces per account. + + + :param use_cases: List[:class:`KeyUseCase`] + The cases that the key can be used for. + :param aws_key_info: :class:`CreateAwsKeyInfo` (optional) + + :param gcp_key_info: :class:`CreateGcpKeyInfo` (optional) + + + :returns: :class:`CustomerManagedKey` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + use_cases: + type: string + description: The cases that the key can be used for. + aws_key_info: + type: string + description: ':param gcp_key_info: :class:`CreateGcpKeyInfo` (optional)' + gcp_key_info: + type: string + required: + - use_cases + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomerManagedKey' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EncryptionKeysAPI + get: + operationId: encryption_keys_list + summary: Lists Databricks customer-managed key configurations for an account. + tags: + - provisioning + - encryption_keys + description: |- + Lists Databricks customer-managed key configurations for an account. + + + :returns: Iterator over :class:`CustomerManagedKey` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomerManagedKey' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EncryptionKeysAPI + /api/2.0/accounts/{account_id}/customer-managed-keys/{customer_managed_key_id}: + delete: + operationId: encryption_keys_delete + summary: >- + Deletes a customer-managed key configuration object for an account. You + cannot delete a configuration + tags: + - provisioning + - encryption_keys + description: >- + Deletes a customer-managed key configuration object for an account. You + cannot delete a configuration + + that is associated with a running workspace. + + + :param customer_managed_key_id: str + Databricks encryption key configuration ID. + + :returns: :class:`CustomerManagedKey` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: customer_managed_key_id + in: path + required: true + schema: + type: string + description: Databricks encryption key configuration ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomerManagedKey' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EncryptionKeysAPI + get: + operationId: encryption_keys_get + summary: >- + Gets a customer-managed key configuration object for an account, + specified by ID. This operation + tags: + - provisioning + - encryption_keys + description: >- + Gets a customer-managed key configuration object for an account, + specified by ID. This operation + + uploads a reference to a customer-managed key to Databricks. If assigned + as a workspace's + + customer-managed key for managed services, Databricks uses the key to + encrypt the workspaces notebooks + + and secrets in the control plane, in addition to Databricks SQL queries + and query history. If it is + + specified as a workspace's customer-managed key for storage, the key + encrypts the workspace's root S3 + + bucket (which contains the workspace's root DBFS and system data) and, + optionally, cluster EBS volume + + data. + + + **Important**: Customer-managed keys are supported only for some + deployment types, subscription types, + + and AWS regions. + + + This operation is available only if your account is on the E2 version of + the platform.", + + + :param customer_managed_key_id: str + Databricks encryption key configuration ID. + + :returns: :class:`CustomerManagedKey` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: customer_managed_key_id + in: path + required: true + schema: + type: string + description: Databricks encryption key configuration ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomerManagedKey' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EncryptionKeysAPI + /api/2.0/accounts/{account_id}/networks: + post: + operationId: networks_create + summary: >- + Creates a Databricks network configuration that represents an VPC and + its resources. The VPC will be + tags: + - provisioning + - networks + description: >- + Creates a Databricks network configuration that represents an VPC and + its resources. The VPC will be + + used for new Databricks clusters. This requires a pre-existing VPC and + subnets. + + + :param gcp_network_info: :class:`GcpNetworkInfo` (optional) + + :param network_name: str (optional) + The human-readable name of the network configuration. + :param security_group_ids: List[str] (optional) + IDs of one to five security groups associated with this network. Security group IDs **cannot** be + used in multiple network configurations. + :param subnet_ids: List[str] (optional) + IDs of at least two subnets associated with this network. Subnet IDs **cannot** be used in multiple + network configurations. + :param vpc_endpoints: :class:`NetworkVpcEndpoints` (optional) + + :param vpc_id: str (optional) + The ID of the VPC associated with this network configuration. VPC IDs can be used in multiple + networks. + + :returns: :class:`Network` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + gcp_network_info: + type: string + description: >- + :param network_name: str (optional) The human-readable name + of the network configuration. + network_name: + type: string + security_group_ids: + type: string + description: >- + IDs of one to five security groups associated with this + network. Security group IDs **cannot** be used in multiple + network configurations. + subnet_ids: + type: string + description: >- + IDs of at least two subnets associated with this network. + Subnet IDs **cannot** be used in multiple network + configurations. + vpc_endpoints: + type: string + description: >- + :param vpc_id: str (optional) The ID of the VPC associated + with this network configuration. VPC IDs can be used in + multiple networks. + vpc_id: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Network' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworksAPI + get: + operationId: networks_list + summary: Lists Databricks network configurations for an account. + tags: + - provisioning + - networks + description: |- + Lists Databricks network configurations for an account. + + + :returns: Iterator over :class:`Network` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Network' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworksAPI + /api/2.0/accounts/{account_id}/networks/{network_id}: + delete: + operationId: networks_delete + summary: >- + Deletes a Databricks network configuration, which represents a cloud VPC + and its resources. You cannot + tags: + - provisioning + - networks + description: >- + Deletes a Databricks network configuration, which represents a cloud VPC + and its resources. You cannot + + delete a network that is associated with a workspace. + + + This operation is available only if your account is on the E2 version of + the platform. + + + :param network_id: str + Databricks Account API network configuration ID. + + :returns: :class:`Network` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_id + in: path + required: true + schema: + type: string + description: Databricks Account API network configuration ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Network' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworksAPI + get: + operationId: networks_get + summary: >- + Gets a Databricks network configuration, which represents a cloud VPC + and its resources. + tags: + - provisioning + - networks + description: >- + Gets a Databricks network configuration, which represents a cloud VPC + and its resources. + + + :param network_id: str + Databricks Account API network configuration ID. + + :returns: :class:`Network` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_id + in: path + required: true + schema: + type: string + description: Databricks Account API network configuration ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Network' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworksAPI + /api/2.0/accounts/{account_id}/private-access-settings: + post: + operationId: private_access_create + summary: >- + Creates a private access settings configuration, which represents + network access restrictions for + tags: + - provisioning + - private_access + description: >- + Creates a private access settings configuration, which represents + network access restrictions for + + workspace resources. Private access settings configure whether + workspaces can be accessed from the + + public internet or only from private endpoints. + + + :param allowed_vpc_endpoint_ids: List[str] (optional) + An array of Databricks VPC endpoint IDs. This is the Databricks ID returned when registering the VPC + endpoint configuration in your Databricks account. This is not the ID of the VPC endpoint in AWS. + Only used when private_access_level is set to ENDPOINT. This is an allow list of VPC endpoints + registered in your Databricks account that can connect to your workspace over AWS PrivateLink. Note: + If hybrid access to your workspace is enabled by setting public_access_enabled to true, this control + only works for PrivateLink connections. To control how your workspace is accessed via public + internet, see IP access lists. + :param private_access_level: :class:`PrivateAccessLevel` (optional) + The private access level controls which VPC endpoints can connect to the UI or API of any workspace + that attaches this private access settings object. `ACCOUNT` level access (the default) allows only + VPC endpoints that are registered in your Databricks account connect to your workspace. `ENDPOINT` + level access allows only specified VPC endpoints connect to your workspace. For details, see + allowed_vpc_endpoint_ids. + :param private_access_settings_name: str (optional) + The human-readable name of the private access settings object. + :param public_access_enabled: bool (optional) + Determines if the workspace can be accessed over public internet. For fully private workspaces, you + can optionally specify false, but only if you implement both the front-end and the back-end + PrivateLink connections. Otherwise, specify true, which means that public access is enabled. + :param region: str (optional) + The AWS region for workspaces attached to this private access settings object. + + :returns: :class:`PrivateAccessSettings` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allowed_vpc_endpoint_ids: + type: string + description: >- + An array of Databricks VPC endpoint IDs. This is the + Databricks ID returned when registering the VPC endpoint + configuration in your Databricks account. This is not the ID + of the VPC endpoint in AWS. Only used when + private_access_level is set to ENDPOINT. This is an allow + list of VPC endpoints registered in your Databricks account + that can connect to your workspace over AWS PrivateLink. + Note: If hybrid access to your workspace is enabled by + setting public_access_enabled to true, this control only + works for PrivateLink connections. To control how your + workspace is accessed via public internet, see IP access + lists. + private_access_level: + type: string + description: >- + The private access level controls which VPC endpoints can + connect to the UI or API of any workspace that attaches this + private access settings object. `ACCOUNT` level access (the + default) allows only VPC endpoints that are registered in + your Databricks account connect to your workspace. + `ENDPOINT` level access allows only specified VPC endpoints + connect to your workspace. For details, see + allowed_vpc_endpoint_ids. + private_access_settings_name: + type: string + description: >- + The human-readable name of the private access settings + object. + public_access_enabled: + type: string + description: >- + Determines if the workspace can be accessed over public + internet. For fully private workspaces, you can optionally + specify false, but only if you implement both the front-end + and the back-end PrivateLink connections. Otherwise, specify + true, which means that public access is enabled. + region: + type: string + description: >- + The AWS region for workspaces attached to this private + access settings object. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PrivateAccessSettings' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PrivateAccessAPI + get: + operationId: private_access_list + summary: Lists Databricks private access settings for an account. + tags: + - provisioning + - private_access + description: |- + Lists Databricks private access settings for an account. + + + :returns: Iterator over :class:`PrivateAccessSettings` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PrivateAccessSettings' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PrivateAccessAPI + /api/2.0/accounts/{account_id}/private-access-settings/{private_access_settings_id}: + delete: + operationId: private_access_delete + summary: >- + Deletes a Databricks private access settings configuration, both + specified by ID. + tags: + - provisioning + - private_access + description: >- + Deletes a Databricks private access settings configuration, both + specified by ID. + + + :param private_access_settings_id: str + + + :returns: :class:`PrivateAccessSettings` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: private_access_settings_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`PrivateAccessSettings`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PrivateAccessSettings' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PrivateAccessAPI + get: + operationId: private_access_get + summary: >- + Gets a Databricks private access settings configuration, both specified + by ID. + tags: + - provisioning + - private_access + description: >- + Gets a Databricks private access settings configuration, both specified + by ID. + + + :param private_access_settings_id: str + + + :returns: :class:`PrivateAccessSettings` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: private_access_settings_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`PrivateAccessSettings`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PrivateAccessSettings' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PrivateAccessAPI + put: + operationId: private_access_replace + summary: >- + Updates an existing private access settings object, which specifies how + your workspace is accessed + tags: + - provisioning + - private_access + description: >- + Updates an existing private access settings object, which specifies how + your workspace is accessed + + over AWS PrivateLink. To use AWS PrivateLink, a workspace must have a + private access settings object + + referenced by ID in the workspace's private_access_settings_id property. + This operation completely + + overwrites your existing private access settings object attached to your + workspaces. All workspaces + + attached to the private access settings are affected by any change. If + public_access_enabled, + + private_access_level, or allowed_vpc_endpoint_ids are updated, effects + of these changes might take + + several minutes to propagate to the workspace API. You can share one + private access settings object + + with multiple workspaces in a single account. However, private access + settings are specific to AWS + + regions, so only workspaces in the same AWS region can use a given + private access settings object. + + Before configuring PrivateLink, read the Databricks article about + PrivateLink. + + + :param private_access_settings_id: str + Databricks private access settings ID. + :param customer_facing_private_access_settings: + :class:`PrivateAccessSettings` + Properties of the new private access settings object. + + :returns: :class:`PrivateAccessSettings` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: private_access_settings_id + in: path + required: true + schema: + type: string + description: Databricks private access settings ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + customer_facing_private_access_settings: + type: string + description: Properties of the new private access settings object. + required: + - customer_facing_private_access_settings + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PrivateAccessSettings' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PrivateAccessAPI + /api/2.0/accounts/{account_id}/storage-configurations: + post: + operationId: storage_create + summary: Creates a Databricks storage configuration for an account. + tags: + - provisioning + - storage + description: |- + Creates a Databricks storage configuration for an account. + + :param storage_configuration_name: str + The human-readable name of the storage configuration. + :param root_bucket_info: :class:`RootBucketInfo` + Root S3 bucket information. + :param role_arn: str (optional) + Optional IAM role that is used to access the workspace catalog which is created during workspace + creation for UC by Default. If a storage configuration with this field populated is used to create a + workspace, then a workspace catalog is created together with the workspace. The workspace catalog + shares the root bucket with internal workspace storage (including DBFS root) but uses a dedicated + bucket path prefix. + + :returns: :class:`StorageConfiguration` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + storage_configuration_name: + type: string + description: The human-readable name of the storage configuration. + root_bucket_info: + type: string + description: Root S3 bucket information. + role_arn: + type: string + description: >- + Optional IAM role that is used to access the workspace + catalog which is created during workspace creation for UC by + Default. If a storage configuration with this field + populated is used to create a workspace, then a workspace + catalog is created together with the workspace. The + workspace catalog shares the root bucket with internal + workspace storage (including DBFS root) but uses a dedicated + bucket path prefix. + required: + - storage_configuration_name + - root_bucket_info + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StorageConfiguration' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageAPI + get: + operationId: storage_list + summary: Lists Databricks storage configurations for an account, specified by ID. + tags: + - provisioning + - storage + description: |- + Lists Databricks storage configurations for an account, specified by ID. + + + :returns: Iterator over :class:`StorageConfiguration` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StorageConfiguration' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageAPI + /api/2.0/accounts/{account_id}/storage-configurations/{storage_configuration_id}: + delete: + operationId: storage_delete + summary: >- + Deletes a Databricks storage configuration. You cannot delete a storage + configuration that is + tags: + - provisioning + - storage + description: >- + Deletes a Databricks storage configuration. You cannot delete a storage + configuration that is + + associated with any workspace. + + + :param storage_configuration_id: str + + + :returns: :class:`StorageConfiguration` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: storage_configuration_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`StorageConfiguration`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StorageConfiguration' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageAPI + get: + operationId: storage_get + summary: >- + Gets a Databricks storage configuration for an account, both specified + by ID. + tags: + - provisioning + - storage + description: >- + Gets a Databricks storage configuration for an account, both specified + by ID. + + + :param storage_configuration_id: str + + + :returns: :class:`StorageConfiguration` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: storage_configuration_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`StorageConfiguration`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StorageConfiguration' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageAPI + /api/2.0/accounts/{account_id}/vpc-endpoints: + post: + operationId: vpc_endpoints_create + summary: >- + Creates a VPC endpoint configuration, which represents a [VPC endpoint] + object in AWS used to + tags: + - provisioning + - vpc_endpoints + description: >- + Creates a VPC endpoint configuration, which represents a [VPC endpoint] + object in AWS used to + + communicate privately with Databricks over [AWS PrivateLink]. + + + After you create the VPC endpoint configuration, the Databricks + [endpoint service] automatically + + accepts the VPC endpoint. + + + Before configuring PrivateLink, read the [Databricks article about + PrivateLink]. + + + [AWS PrivateLink]: https://aws.amazon.com/privatelink + + [Databricks article about PrivateLink]: + https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html + + [VPC endpoint]: + https://docs.aws.amazon.com/vpc/latest/privatelink/vpc-endpoints.html + + [endpoint service]: + https://docs.aws.amazon.com/vpc/latest/privatelink/privatelink-share-your-services.html + + + :param aws_vpc_endpoint_id: str (optional) + The ID of the VPC endpoint object in AWS. + :param gcp_vpc_endpoint_info: :class:`GcpVpcEndpointInfo` (optional) + The cloud info of this vpc endpoint. + :param region: str (optional) + The region in which this VPC endpoint object exists. + :param vpc_endpoint_name: str (optional) + The human-readable name of the storage configuration. + + :returns: :class:`VpcEndpoint` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + aws_vpc_endpoint_id: + type: string + description: The ID of the VPC endpoint object in AWS. + gcp_vpc_endpoint_info: + type: string + description: The cloud info of this vpc endpoint. + region: + type: string + description: The region in which this VPC endpoint object exists. + vpc_endpoint_name: + type: string + description: The human-readable name of the storage configuration. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VpcEndpoint' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VpcEndpointsAPI + get: + operationId: vpc_endpoints_list + summary: Lists Databricks VPC endpoint configurations for an account. + tags: + - provisioning + - vpc_endpoints + description: |- + Lists Databricks VPC endpoint configurations for an account. + + + :returns: Iterator over :class:`VpcEndpoint` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VpcEndpoint' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VpcEndpointsAPI + /api/2.0/accounts/{account_id}/vpc-endpoints/{vpc_endpoint_id}: + delete: + operationId: vpc_endpoints_delete + summary: >- + Deletes a Databricks VPC endpoint configuration. You cannot delete a VPC + endpoint configuration that + tags: + - provisioning + - vpc_endpoints + description: >- + Deletes a Databricks VPC endpoint configuration. You cannot delete a VPC + endpoint configuration that + + is associated with any workspace. + + + :param vpc_endpoint_id: str + + + :returns: :class:`VpcEndpoint` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: vpc_endpoint_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`VpcEndpoint`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VpcEndpoint' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VpcEndpointsAPI + get: + operationId: vpc_endpoints_get + summary: >- + Gets a VPC endpoint configuration, which represents a [VPC endpoint] + object in AWS used to communicate + tags: + - provisioning + - vpc_endpoints + description: >- + Gets a VPC endpoint configuration, which represents a [VPC endpoint] + object in AWS used to communicate + + privately with Databricks over [AWS PrivateLink]. + + + [AWS PrivateLink]: https://aws.amazon.com/privatelink + + [VPC endpoint]: + https://docs.aws.amazon.com/vpc/latest/privatelink/concepts.html + + + :param vpc_endpoint_id: str + Databricks VPC endpoint ID. + + :returns: :class:`VpcEndpoint` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: vpc_endpoint_id + in: path + required: true + schema: + type: string + description: Databricks VPC endpoint ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VpcEndpoint' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VpcEndpointsAPI + /api/2.0/accounts/{account_id}/workspaces: + post: + operationId: workspaces_create + summary: >- + Creates a new workspace using a credential configuration and a storage + configuration, an optional + tags: + - provisioning + - workspaces + description: >- + Creates a new workspace using a credential configuration and a storage + configuration, an optional + + network configuration (if using a customer-managed VPC), an optional + managed services key + + configuration (if using customer-managed keys for managed services), and + an optional storage key + + configuration (if using customer-managed keys for storage). The key + configurations used for managed + + services and storage encryption can be the same or different. + + + Important: This operation is asynchronous. A response with HTTP status + code 200 means the request has + + been accepted and is in progress, but does not mean that the workspace + deployed successfully and is + + running. The initial workspace status is typically PROVISIONING. Use the + workspace ID (workspace_id) + + field in the response to identify the new workspace and make repeated + GET requests with the workspace + + ID and check its status. The workspace becomes available when the status + changes to RUNNING. + + + You can share one customer-managed VPC with multiple workspaces in a + single account. It is not + + required to create a new VPC for each workspace. However, you cannot + reuse subnets or Security Groups + + between workspaces. If you plan to share one VPC with multiple + workspaces, make sure you size your VPC + + and subnets accordingly. Because a Databricks Account API network + configuration encapsulates this + + information, you cannot reuse a Databricks Account API network + configuration across workspaces. + + + For information about how to create a new workspace with this API + including error handling, see + + [Create a new workspace using the Account API]. + + + Important: Customer-managed VPCs, PrivateLink, and customer-managed keys + are supported on a limited + + set of deployment and subscription types. If you have questions about + availability, contact your + + Databricks representative. + + + This operation is available only if your account is on the E2 version of + the platform or on a select + + custom plan that allows multiple workspaces per account. + + + [Create a new workspace using the Account API]: + http://docs.databricks.com/administration-guide/account-api/new-workspace.html + + + :param aws_region: str (optional) + + :param cloud: str (optional) + The cloud name. This field always has the value `gcp`. + :param cloud_resource_container: :class:`CloudResourceContainer` + (optional) + + :param compute_mode: :class:`CustomerFacingComputeMode` (optional) + If the compute mode is `SERVERLESS`, a serverless workspace is created that comes pre-configured + with serverless compute and default storage, providing a fully-managed, enterprise-ready SaaS + experience. This means you don't need to provide any resources managed by you, such as credentials, + storage, or network. If the compute mode is `HYBRID` (which is the default option), a classic + workspace is created that uses customer-managed resources. + :param credentials_id: str (optional) + ID of the workspace's credential configuration object. + :param custom_tags: Dict[str,str] (optional) + The custom tags key-value pairing that is attached to this workspace. The key-value pair is a string + of utf-8 characters. The value can be an empty string, with maximum length of 255 characters. The + key can be of maximum length of 127 characters, and cannot be empty. + :param deployment_name: str (optional) + The deployment name defines part of the subdomain for the workspace. The workspace URL for the web + application and REST APIs is .cloud.databricks.com. For example, if the + deployment name is abcsales, your workspace URL will be https://abcsales.cloud.databricks.com. + Hyphens are allowed. This property supports only the set of characters that are allowed in a + subdomain. To set this value, you must have a deployment name prefix. Contact your Databricks + account team to add an account deployment name prefix to your account. Workspace deployment names + follow the account prefix and a hyphen. For example, if your account's deployment prefix is acme and + the workspace deployment name is workspace-1, the JSON response for the deployment_name field + becomes acme-workspace-1. The workspace URL would be acme-workspace-1.cloud.databricks.com. You can + also set the deployment_name to the reserved keyword EMPTY if you want the deployment name to only + include the deployment prefix. For example, if your account's deployment prefix is acme and the + workspace deployment name is EMPTY, the deployment_name becomes acme only and the workspace URL is + acme.cloud.databricks.com. This value must be unique across all non-deleted deployments across all + AWS regions. If a new workspace omits this property, the server generates a unique deployment name + for you with the pattern dbc-xxxxxxxx-xxxx. + :param gcp_managed_network_config: :class:`GcpManagedNetworkConfig` + (optional) + + :param gke_config: :class:`GkeConfig` (optional) + + :param location: str (optional) + The Google Cloud region of the workspace data plane in your Google account (for example, + `us-east4`). + :param managed_services_customer_managed_key_id: str (optional) + The ID of the workspace's managed services encryption key configuration object. This is used to help + protect and control access to the workspace's notebooks, secrets, Databricks SQL queries, and query + history. The provided key configuration object property use_cases must contain MANAGED_SERVICES. + :param network_connectivity_config_id: str (optional) + The object ID of network connectivity config. Once assigned, the workspace serverless compute + resources use the same set of stable IP CIDR blocks and optional private link to access your + resources. + :param network_id: str (optional) + The ID of the workspace's network configuration object. To use AWS PrivateLink, this field is + required. + :param pricing_tier: :class:`PricingTier` (optional) + + :param private_access_settings_id: str (optional) + ID of the workspace's private access settings object. Only used for PrivateLink. You must specify + this ID if you are using [AWS PrivateLink] for either front-end (user-to-workspace connection), + back-end (data plane to control plane connection), or both connection types. Before configuring + PrivateLink, read the [Databricks article about PrivateLink].", + + [AWS PrivateLink]: https://aws.amazon.com/privatelink/ + [Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html + :param storage_configuration_id: str (optional) + ID of the workspace's storage configuration object. + :param storage_customer_managed_key_id: str (optional) + The ID of the workspace's storage encryption key configuration object. This is used to encrypt the + workspace's root S3 bucket (root DBFS and system data) and, optionally, cluster EBS volumes. The + provided key configuration object property use_cases must contain STORAGE. + :param workspace_name: str (optional) + The human-readable name of the workspace. + + :returns: + Long-running operation waiter for :class:`Workspace`. + See :method:wait_get_workspace_running for more details. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + aws_region: + type: string + description: >- + :param cloud: str (optional) The cloud name. This field + always has the value `gcp`. + cloud: + type: string + cloud_resource_container: + type: string + description: >- + :param compute_mode: :class:`CustomerFacingComputeMode` + (optional) If the compute mode is `SERVERLESS`, a serverless + workspace is created that comes pre-configured with + serverless compute and default storage, providing a + fully-managed, enterprise-ready SaaS experience. This means + you don't need to provide any resources managed by you, such + as credentials, storage, or network. If the compute mode is + `HYBRID` (which is the default option), a classic workspace + is created that uses customer-managed resources. + compute_mode: + type: string + credentials_id: + type: string + description: ID of the workspace's credential configuration object. + custom_tags: + type: string + description: >- + The custom tags key-value pairing that is attached to this + workspace. The key-value pair is a string of utf-8 + characters. The value can be an empty string, with maximum + length of 255 characters. The key can be of maximum length + of 127 characters, and cannot be empty. + deployment_name: + type: string + description: >- + The deployment name defines part of the subdomain for the + workspace. The workspace URL for the web application and + REST APIs is + .cloud.databricks.com. For + example, if the deployment name is abcsales, your workspace + URL will be https://abcsales.cloud.databricks.com. Hyphens + are allowed. This property supports only the set of + characters that are allowed in a subdomain. To set this + value, you must have a deployment name prefix. Contact your + Databricks account team to add an account deployment name + prefix to your account. Workspace deployment names follow + the account prefix and a hyphen. For example, if your + account's deployment prefix is acme and the workspace + deployment name is workspace-1, the JSON response for the + deployment_name field becomes acme-workspace-1. The + workspace URL would be + acme-workspace-1.cloud.databricks.com. You can also set the + deployment_name to the reserved keyword EMPTY if you want + the deployment name to only include the deployment prefix. + For example, if your account's deployment prefix is acme and + the workspace deployment name is EMPTY, the deployment_name + becomes acme only and the workspace URL is + acme.cloud.databricks.com. This value must be unique across + all non-deleted deployments across all AWS regions. If a new + workspace omits this property, the server generates a unique + deployment name for you with the pattern dbc-xxxxxxxx-xxxx. + gcp_managed_network_config: + type: string + description: ':param gke_config: :class:`GkeConfig` (optional)' + gke_config: + type: string + location: + type: string + description: >- + The Google Cloud region of the workspace data plane in your + Google account (for example, `us-east4`). + managed_services_customer_managed_key_id: + type: string + description: >- + The ID of the workspace's managed services encryption key + configuration object. This is used to help protect and + control access to the workspace's notebooks, secrets, + Databricks SQL queries, and query history. The provided key + configuration object property use_cases must contain + MANAGED_SERVICES. + network_connectivity_config_id: + type: string + description: >- + The object ID of network connectivity config. Once assigned, + the workspace serverless compute resources use the same set + of stable IP CIDR blocks and optional private link to access + your resources. + network_id: + type: string + description: >- + The ID of the workspace's network configuration object. To + use AWS PrivateLink, this field is required. + pricing_tier: + type: string + description: >- + :param private_access_settings_id: str (optional) ID of the + workspace's private access settings object. Only used for + PrivateLink. You must specify this ID if you are using [AWS + PrivateLink] for either front-end (user-to-workspace + connection), back-end (data plane to control plane + connection), or both connection types. Before configuring + PrivateLink, read the [Databricks article about + PrivateLink].", [AWS PrivateLink]: + https://aws.amazon.com/privatelink/ [Databricks article + about PrivateLink]: + https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html + private_access_settings_id: + type: string + storage_configuration_id: + type: string + description: ID of the workspace's storage configuration object. + storage_customer_managed_key_id: + type: string + description: >- + The ID of the workspace's storage encryption key + configuration object. This is used to encrypt the + workspace's root S3 bucket (root DBFS and system data) and, + optionally, cluster EBS volumes. The provided key + configuration object property use_cases must contain + STORAGE. + workspace_name: + type: string + description: The human-readable name of the workspace. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Workspace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspacesAPI + get: + operationId: workspaces_list + summary: Lists Databricks workspaces for an account. + tags: + - provisioning + - workspaces + description: |- + Lists Databricks workspaces for an account. + + + :returns: Iterator over :class:`Workspace` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Workspace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspacesAPI + /api/2.0/accounts/{account_id}/workspaces/{workspace_id}: + delete: + operationId: workspaces_delete + summary: Deletes a Databricks workspace, both specified by ID. + tags: + - provisioning + - workspaces + description: |- + Deletes a Databricks workspace, both specified by ID. + + :param workspace_id: int + + :returns: :class:`Workspace` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: ':returns: :class:`Workspace`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Workspace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspacesAPI + get: + operationId: workspaces_get + summary: >- + Gets information including status for a Databricks workspace, specified + by ID. In the response, the + tags: + - provisioning + - workspaces + description: >- + Gets information including status for a Databricks workspace, specified + by ID. In the response, the + + `workspace_status` field indicates the current status. After initial + workspace creation (which is + + asynchronous), make repeated `GET` requests with the workspace ID and + check its status. The workspace + + becomes available when the status changes to `RUNNING`. For information + about how to create a new + + workspace with this API **including error handling**, see [Create a new + workspace using the Account + + API]. + + + [Create a new workspace using the Account API]: + http://docs.databricks.com/administration-guide/account-api/new-workspace.html + + + :param workspace_id: int + + + :returns: :class:`Workspace` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: ':returns: :class:`Workspace`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Workspace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspacesAPI + patch: + operationId: workspaces_update + summary: Updates a workspace. + tags: + - provisioning + - workspaces + description: |- + Updates a workspace. + + :param workspace_id: int + A unique integer ID for the workspace + :param customer_facing_workspace: :class:`Workspace` + :param update_mask: str (optional) + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: + Long-running operation waiter for :class:`Workspace`. + See :method:wait_get_workspace_running for more details. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: A unique integer ID for the workspace + - name: update_mask + in: query + required: false + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + customer_facing_workspace: + type: string + description: >- + :param update_mask: str (optional) The field mask must be a + single string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + required: + - customer_facing_workspace + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Workspace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspacesAPI +components: + schemas: + AwsCredentials: + type: object + properties: + sts_role: + $ref: '#/components/schemas/StsRole' + AwsKeyInfo: + type: object + properties: + key_arn: + type: string + key_region: + type: string + description: The AWS KMS key region. + key_alias: + type: string + description: The AWS KMS key alias. + reuse_key_for_cluster_volumes: + type: boolean + description: >- + This field applies only if the `use_cases` property includes + `STORAGE`. If this is set to true or omitted, the key is also used + to encrypt cluster EBS volumes. If you do not want to use this key + for encrypting EBS volumes, set to false. + required: + - key_arn + - key_region + AzureKeyInfo: + type: object + properties: + disk_encryption_set_id: + type: string + key_access_configuration: + $ref: '#/components/schemas/KeyAccessConfiguration' + description: >- + The structure to store key access credential This is set if the + Managed Identity is being used to access the Azure Key Vault key. + key_name: + type: string + description: The name of the key in KeyVault. + key_vault_uri: + type: string + description: The base URI of the KeyVault. + tenant_id: + type: string + description: The tenant id where the KeyVault lives. + version: + type: string + description: The current key version. + AzureWorkspaceInfo: + type: object + properties: + resource_group: + type: string + subscription_id: + type: string + description: Azure Subscription ID + CloudResourceContainer: + type: object + properties: + gcp: + $ref: '#/components/schemas/CustomerFacingGcpCloudResourceContainer' + CreateAwsKeyInfo: + type: object + properties: + key_arn: + type: string + key_alias: + type: string + description: The AWS KMS key alias. + key_region: + type: string + description: The AWS KMS key region. + reuse_key_for_cluster_volumes: + type: boolean + description: >- + This field applies only if the `use_cases` property includes + `STORAGE`. If this is set to true or omitted, the key is also used + to encrypt cluster EBS volumes. If you do not want to use this key + for encrypting EBS volumes, set to false. + required: + - key_arn + CreateCredentialAwsCredentials: + type: object + properties: + sts_role: + $ref: '#/components/schemas/CreateCredentialStsRole' + CreateCredentialStsRole: + type: object + properties: + role_arn: + type: string + CreateGcpKeyInfo: + type: object + properties: + kms_key_id: + type: string + gcp_service_account: + $ref: '#/components/schemas/GcpServiceAccount' + description: >- + Globally unique service account email that has access to the KMS + key. The service account exists within the Databricks CP project. + required: + - kms_key_id + Credential: + type: object + properties: + account_id: + type: string + aws_credentials: + $ref: '#/components/schemas/AwsCredentials' + creation_time: + type: integer + description: Time in epoch milliseconds when the credential was created. + credentials_id: + type: string + description: Databricks credential configuration ID. + credentials_name: + type: string + description: The human-readable name of the credential configuration object. + CustomerFacingGcpCloudResourceContainer: + type: object + properties: + project_id: + type: string + CustomerManagedKey: + type: object + properties: + account_id: + type: string + aws_key_info: + $ref: '#/components/schemas/AwsKeyInfo' + azure_key_info: + $ref: '#/components/schemas/AzureKeyInfo' + creation_time: + type: integer + description: Time in epoch milliseconds when the customer key was created. + customer_managed_key_id: + type: string + description: ID of the encryption key configuration object. + gcp_key_info: + $ref: '#/components/schemas/GcpKeyInfo' + use_cases: + type: array + items: + $ref: '#/components/schemas/KeyUseCase' + description: The cases that the key can be used for. + GcpCommonNetworkConfig: + type: object + properties: + gke_cluster_master_ip_range: + type: string + description: >- + The IP range that will be used to allocate GKE cluster master + resources from. This field must not be set if + gke_cluster_type=PUBLIC_NODE_PUBLIC_MASTER. + gke_connectivity_type: + $ref: '#/components/schemas/GkeConfigConnectivityType' + description: The type of network connectivity of the GKE cluster. + description: >- + The shared network config for GCP workspace. This object has common + network configurations that + are network attributions of a workspace. DEPRECATED. Use GkeConfig instead. + GcpKeyInfo: + type: object + properties: + kms_key_id: + type: string + gcp_service_account: + $ref: '#/components/schemas/GcpServiceAccount' + description: >- + Globally unique service account email that has access to the KMS + key. The service account exists within the Databricks CP project. + required: + - kms_key_id + GcpManagedNetworkConfig: + type: object + properties: + gke_cluster_pod_ip_range: + type: string + description: The IP range that will be used to allocate GKE cluster Pods from. + gke_cluster_service_ip_range: + type: string + description: >- + The IP range that will be used to allocate GKE cluster Services + from. + subnet_cidr: + type: string + description: >- + The IP range which will be used to allocate GKE cluster nodes from. + Note: Pods, services and master IP range must be mutually exclusive. + description: The network configuration for the workspace. + GcpNetworkInfo: + type: object + properties: + network_project_id: + type: string + vpc_id: + type: string + description: The customer-provided VPC ID. + subnet_id: + type: string + description: >- + The customer-provided Subnet ID that will be available to Clusters + in Workspaces using this Network. + subnet_region: + type: string + pod_ip_range_name: + type: string + description: >- + Name of the secondary range within the subnet that will be used by + GKE as Pod IP range. This is BYO VPC specific. DB VPC uses + network.getGcpManagedNetworkConfig.getGkeClusterPodIpRange + service_ip_range_name: + type: string + description: >- + Name of the secondary range within the subnet that will be used by + GKE as Service IP range. + required: + - network_project_id + - vpc_id + - subnet_id + - subnet_region + - pod_ip_range_name + - service_ip_range_name + GcpServiceAccount: + type: object + properties: + service_account_email: + type: string + GcpVpcEndpointInfo: + type: object + properties: + project_id: + type: string + psc_endpoint_name: + type: string + endpoint_region: + type: string + psc_connection_id: + type: string + service_attachment_id: + type: string + required: + - project_id + - psc_endpoint_name + - endpoint_region + GkeConfig: + type: object + properties: + connectivity_type: + $ref: '#/components/schemas/GkeConfigConnectivityType' + description: The type of network connectivity of the GKE cluster. + master_ip_range: + type: string + description: >- + The IP range that will be used to allocate GKE cluster master + resources from. This field must not be set if + gke_cluster_type=PUBLIC_NODE_PUBLIC_MASTER. + description: The configurations of the GKE cluster used by the GCP workspace. + KeyAccessConfiguration: + type: object + properties: + credential_id: + type: string + description: The credential ID that is used to access the key vault. + Network: + type: object + properties: + account_id: + type: string + creation_time: + type: integer + description: Time in epoch milliseconds when the network was created. + error_messages: + type: array + items: + $ref: '#/components/schemas/NetworkHealth' + description: Array of error messages about the network configuration. + gcp_network_info: + $ref: '#/components/schemas/GcpNetworkInfo' + network_id: + type: string + description: The Databricks network configuration ID. + network_name: + type: string + description: The human-readable name of the network configuration. + security_group_ids: + type: array + items: + type: string + description: >- + IDs of one to five security groups associated with this network. + Security group IDs **cannot** be used in multiple network + configurations. + subnet_ids: + type: array + items: + type: string + description: >- + IDs of at least two subnets associated with this network. Subnet IDs + **cannot** be used in multiple network configurations. + vpc_endpoints: + $ref: '#/components/schemas/NetworkVpcEndpoints' + vpc_id: + type: string + description: >- + The ID of the VPC associated with this network configuration. VPC + IDs can be used in multiple networks. + vpc_status: + $ref: '#/components/schemas/VpcStatus' + warning_messages: + type: array + items: + $ref: '#/components/schemas/NetworkWarning' + description: Array of warning messages about the network configuration. + workspace_id: + type: integer + description: Workspace ID associated with this network configuration. + NetworkHealth: + type: object + properties: + error_message: + type: string + error_type: + $ref: '#/components/schemas/ErrorType' + NetworkVpcEndpoints: + type: object + properties: + dataplane_relay: + type: array + items: + type: string + rest_api: + type: array + items: + type: string + description: >- + The VPC endpoint ID used by this network to access the Databricks + REST API. + NetworkWarning: + type: object + properties: + warning_message: + type: string + warning_type: + $ref: '#/components/schemas/WarningType' + PrivateAccessSettings: + type: object + properties: + account_id: + type: string + description: The Databricks account ID that hosts the private access settings. + allowed_vpc_endpoint_ids: + type: array + items: + type: string + description: >- + An array of Databricks VPC endpoint IDs. This is the Databricks ID + that is returned when registering the VPC endpoint configuration in + your Databricks account. This is not the ID of the VPC endpoint in + AWS. Only used when private_access_level is set to ENDPOINT. This is + an allow list of VPC endpoints that in your account that can connect + to your workspace over AWS PrivateLink. If hybrid access to your + workspace is enabled by setting public_access_enabled to true, this + control only works for PrivateLink connections. To control how your + workspace is accessed via public internet, see IP access lists. + private_access_level: + $ref: '#/components/schemas/PrivateAccessLevel' + description: >- + The private access level controls which VPC endpoints can connect to + the UI or API of any workspace that attaches this private access + settings object. `ACCOUNT` level access (the default) allows only + VPC endpoints that are registered in your Databricks account connect + to your workspace. `ENDPOINT` level access allows only specified VPC + endpoints connect to your workspace. For details, see + allowed_vpc_endpoint_ids. + private_access_settings_id: + type: string + description: Databricks private access settings ID. + private_access_settings_name: + type: string + description: The human-readable name of the private access settings object. + public_access_enabled: + type: boolean + description: >- + Determines if the workspace can be accessed over public internet. + For fully private workspaces, you can optionally specify false, but + only if you implement both the front-end and the back-end + PrivateLink connections. Otherwise, specify true, which means that + public access is enabled. + region: + type: string + description: >- + The AWS region for workspaces attached to this private access + settings object. + description: '*' + RootBucketInfo: + type: object + properties: + bucket_name: + type: string + StorageConfiguration: + type: object + properties: + account_id: + type: string + creation_time: + type: integer + description: >- + Time in epoch milliseconds when the storage configuration was + created. + role_arn: + type: string + description: >- + Optional IAM role that is used to access the workspace catalog which + is created during workspace creation for UC by Default. If a storage + configuration with this field populated is used to create a + workspace, then a workspace catalog is created together with the + workspace. The workspace catalog shares the root bucket with + internal workspace storage (including DBFS root) but uses a + dedicated bucket path prefix. + root_bucket_info: + $ref: '#/components/schemas/RootBucketInfo' + description: The root bucket information for the storage configuration. + storage_configuration_id: + type: string + description: Databricks storage configuration ID. + storage_configuration_name: + type: string + description: The human-readable name of the storage configuration. + StsRole: + type: object + properties: + role_arn: + type: string + VpcEndpoint: + type: object + properties: + account_id: + type: string + description: >- + The Databricks account ID that hosts the VPC endpoint configuration. + TODO - This may signal an OpenAPI diff; it does not show up in the + generated spec + aws_account_id: + type: string + description: The AWS Account in which the VPC endpoint object exists. + aws_endpoint_service_id: + type: string + description: >- + The ID of the Databricks [endpoint service] that this VPC endpoint + is connected to. For a list of endpoint service IDs for each + supported AWS region, see the [Databricks PrivateLink + documentation]. [Databricks PrivateLink documentation]: + https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html + [endpoint service]: + https://docs.aws.amazon.com/vpc/latest/privatelink/endpoint-service.html + aws_vpc_endpoint_id: + type: string + description: The ID of the VPC endpoint object in AWS. + gcp_vpc_endpoint_info: + $ref: '#/components/schemas/GcpVpcEndpointInfo' + description: The cloud info of this vpc endpoint. Info for a GCP vpc endpoint. + region: + type: string + description: The AWS region in which this VPC endpoint object exists. + state: + type: string + description: >- + The current state (such as `available` or `rejected`) of the VPC + endpoint. Derived from AWS. For the full set of values, see [AWS + DescribeVpcEndpoint documentation]. [AWS DescribeVpcEndpoint + documentation]: + https://docs.aws.amazon.com/cli/latest/reference/ec2/describe-vpc-endpoints.html + use_case: + $ref: '#/components/schemas/EndpointUseCase' + description: >- + This enumeration represents the type of Databricks VPC endpoint + service that was used when creating this VPC endpoint. If the VPC + endpoint connects to the Databricks control plane for either the + front-end connection or the back-end REST API connection, the value + is WORKSPACE_ACCESS. If the VPC endpoint connects to the Databricks + workspace for the back-end secure cluster connectivity relay, the + value is DATAPLANE_RELAY_ACCESS. + vpc_endpoint_id: + type: string + description: >- + Databricks VPC endpoint ID. This is the Databricks-specific name of + the VPC endpoint. Do not confuse this with the + `aws_vpc_endpoint_id`, which is the ID within AWS of the VPC + endpoint. + vpc_endpoint_name: + type: string + description: The human-readable name of the storage configuration. + description: '*' + Workspace: + type: object + properties: + account_id: + type: string + aws_region: + type: string + azure_workspace_info: + $ref: '#/components/schemas/AzureWorkspaceInfo' + cloud: + type: string + description: The cloud name. This field can have values like `azure`, `gcp`. + cloud_resource_container: + $ref: '#/components/schemas/CloudResourceContainer' + compute_mode: + $ref: '#/components/schemas/CustomerFacingComputeMode' + description: The compute mode of the workspace. + creation_time: + type: integer + description: Time in epoch milliseconds when the workspace was created. + credentials_id: + type: string + description: ID of the workspace's credential configuration object. + custom_tags: + type: object + description: >- + The custom tags key-value pairing that is attached to this + workspace. The key-value pair is a string of utf-8 characters. The + value can be an empty string, with maximum length of 255 characters. + The key can be of maximum length of 127 characters, and cannot be + empty. + deployment_name: + type: string + expected_workspace_status: + $ref: '#/components/schemas/WorkspaceStatus' + description: >- + A client owned field used to indicate the workspace status that the + client expects to be in. For now this is only used to unblock + Temporal workflow for GCP least privileged workspace. + gcp_managed_network_config: + $ref: '#/components/schemas/GcpManagedNetworkConfig' + gke_config: + $ref: '#/components/schemas/GkeConfig' + location: + type: string + description: >- + The Google Cloud region of the workspace data plane in your Google + account (for example, `us-east4`). + managed_services_customer_managed_key_id: + type: string + description: ID of the key configuration for encrypting managed services. + network: + $ref: '#/components/schemas/WorkspaceNetwork' + description: >- + The network configuration for the workspace. DEPRECATED. Use + `network_id` instead. + network_connectivity_config_id: + type: string + description: The object ID of network connectivity config. + network_id: + type: string + description: >- + If this workspace is BYO VPC, then the network_id will be populated. + If this workspace is not BYO VPC, then the network_id will be empty. + pricing_tier: + $ref: '#/components/schemas/PricingTier' + private_access_settings_id: + type: string + description: >- + ID of the workspace's private access settings object. Only used for + PrivateLink. You must specify this ID if you are using [AWS + PrivateLink] for either front-end (user-to-workspace connection), + back-end (data plane to control plane connection), or both + connection types. Before configuring PrivateLink, read the + [Databricks article about PrivateLink].", [AWS PrivateLink]: + https://aws.amazon.com/privatelink/ [Databricks article about + PrivateLink]: + https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html + storage_configuration_id: + type: string + description: ID of the workspace's storage configuration object. + storage_customer_managed_key_id: + type: string + description: ID of the key configuration for encrypting workspace storage. + storage_mode: + $ref: '#/components/schemas/CustomerFacingStorageMode' + description: The storage mode of the workspace. + workspace_id: + type: integer + description: A unique integer ID for the workspace + workspace_name: + type: string + description: The human-readable name of the workspace. + workspace_status: + $ref: '#/components/schemas/WorkspaceStatus' + description: The status of a workspace + workspace_status_message: + type: string + description: Message describing the current workspace status. + WorkspaceNetwork: + type: object + properties: + gcp_common_network_config: + $ref: '#/components/schemas/GcpCommonNetworkConfig' + description: >- + The shared network config for GCP workspace. This object has common + network configurations that are network attributions of a workspace. + This object is input-only. + gcp_managed_network_config: + $ref: '#/components/schemas/GcpManagedNetworkConfig' + description: >- + The mutually exclusive network deployment modes. The option decides + which network mode the workspace will use. The network config for + GCP workspace with Databricks managed network. This object is + input-only and will not be provided when listing workspaces. See + go/gcp-byovpc-alpha-design for interface decisions. + network_id: + type: string + description: >- + The ID of the network object, if the workspace is a BYOVPC + workspace. This should apply to workspaces on all clouds in internal + services. In accounts-rest-api, user will use workspace.network_id + for input and output instead. Currently (2021-06-19) the network ID + is only used by GCP. + description: The network configuration for workspaces. + CustomerFacingComputeMode: + type: string + enum: + - HYBRID + - SERVERLESS + description: >- + Corresponds to compute mode defined here: + + https://src.dev.databricks.com/databricks/universe@9076536b18479afd639d1c1f9dd5a59f72215e69/-/blob/central/api/common.proto?L872 + CustomerFacingStorageMode: + type: string + enum: + - CUSTOMER_HOSTED + - DEFAULT_STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EndpointUseCase: + type: string + enum: + - DATAPLANE_RELAY_ACCESS + - WORKSPACE_ACCESS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ErrorType: + type: string + enum: + - credentials + - networkAcl + - securityGroup + - subnet + - vpc + description: >- + ErrorType and WarningType are used to represent the type of error or + warning by NetworkHealth + + and NetworkWarning defined in central/api/accounts/accounts.proto + GkeConfigConnectivityType: + type: string + enum: + - PRIVATE_NODE_PUBLIC_MASTER + - PUBLIC_NODE_PUBLIC_MASTER + description: >- + Specifies the network connectivity types for the GKE nodes and the GKE + master network. + + + Set to `PRIVATE_NODE_PUBLIC_MASTER` for a private GKE cluster for the + workspace. The GKE nodes + + will not have public IPs. + + + Set to `PUBLIC_NODE_PUBLIC_MASTER` for a public GKE cluster. The nodes + of a public GKE cluster + + have public IP addresses. + KeyUseCase: + type: string + enum: + - MANAGED_SERVICES + - STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PricingTier: + type: string + enum: + - COMMUNITY_EDITION + - DEDICATED + - ENTERPRISE + - PREMIUM + - STANDARD + - UNKNOWN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PrivateAccessLevel: + type: string + enum: + - ACCOUNT + - ENDPOINT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + VpcStatus: + type: string + enum: + - BROKEN + - UNATTACHED + - VALID + - WARNED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + WarningType: + type: string + enum: + - securityGroup + - subnet + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + WorkspaceStatus: + type: string + enum: + - BANNED + - CANCELLING + - FAILED + - NOT_PROVISIONED + - PROVISIONING + - RUNNING + description: >- + The different statuses of a workspace. The following represents the + current set of valid + + transitions from status to status: NOT_PROVISIONED -> PROVISIONING -> + CANCELLED PROVISIONING -> + + RUNNING -> FAILED -> CANCELLED (note that this transition is disallowed + in the MultiWorkspace + + Project) RUNNING -> PROVISIONING -> BANNED -> CANCELLED FAILED -> + PROVISIONING -> CANCELLED + + BANNED -> RUNNING -> CANCELLED Note that a transition from any state to + itself is also valid. + + TODO(PLAT-5867): add a transition from CANCELLED to some other value + (e.g. RECOVERING) + x-stackQL-resources: + credentials: + id: databricks_account.provisioning.credentials + name: credentials + title: Credentials + methods: + credentials_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + credentials_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1credentials/get' + response: + mediaType: application/json + openAPIDocKey: '200' + credentials_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1credentials~1{credentials_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + credentials_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1credentials~1{credentials_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/credentials_get + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/credentials_list + insert: + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/credentials_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/credentials/methods/credentials_delete + replace: [] + encryption_keys: + id: databricks_account.provisioning.encryption_keys + name: encryption_keys + title: Encryption Keys + methods: + encryption_keys_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys/post + response: + mediaType: application/json + openAPIDocKey: '200' + encryption_keys_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys/get + response: + mediaType: application/json + openAPIDocKey: '200' + encryption_keys_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys~1{customer_managed_key_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + encryption_keys_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1customer-managed-keys~1{customer_managed_key_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/encryption_keys/methods/encryption_keys_get + - $ref: >- + #/components/x-stackQL-resources/encryption_keys/methods/encryption_keys_list + insert: + - $ref: >- + #/components/x-stackQL-resources/encryption_keys/methods/encryption_keys_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/encryption_keys/methods/encryption_keys_delete + replace: [] + networks: + id: databricks_account.provisioning.networks + name: networks + title: Networks + methods: + networks_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1networks/post' + response: + mediaType: application/json + openAPIDocKey: '200' + networks_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1networks/get' + response: + mediaType: application/json + openAPIDocKey: '200' + networks_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1networks~1{network_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + networks_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1networks~1{network_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/networks/methods/networks_get' + - $ref: '#/components/x-stackQL-resources/networks/methods/networks_list' + insert: + - $ref: '#/components/x-stackQL-resources/networks/methods/networks_create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/networks/methods/networks_delete' + replace: [] + private_access: + id: databricks_account.provisioning.private_access + name: private_access + title: Private Access + methods: + private_access_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings/post + response: + mediaType: application/json + openAPIDocKey: '200' + private_access_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings/get + response: + mediaType: application/json + openAPIDocKey: '200' + private_access_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings~1{private_access_settings_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + private_access_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings~1{private_access_settings_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + private_access_replace: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1private-access-settings~1{private_access_settings_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/private_access/methods/private_access_get + - $ref: >- + #/components/x-stackQL-resources/private_access/methods/private_access_list + insert: + - $ref: >- + #/components/x-stackQL-resources/private_access/methods/private_access_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/private_access/methods/private_access_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/private_access/methods/private_access_replace + storage: + id: databricks_account.provisioning.storage + name: storage + title: Storage + methods: + storage_create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations/post + response: + mediaType: application/json + openAPIDocKey: '200' + storage_list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations/get + response: + mediaType: application/json + openAPIDocKey: '200' + storage_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations~1{storage_configuration_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + storage_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1storage-configurations~1{storage_configuration_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/storage/methods/storage_get' + - $ref: '#/components/x-stackQL-resources/storage/methods/storage_list' + insert: + - $ref: '#/components/x-stackQL-resources/storage/methods/storage_create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/storage/methods/storage_delete' + replace: [] + vpc_endpoints: + id: databricks_account.provisioning.vpc_endpoints + name: vpc_endpoints + title: Vpc Endpoints + methods: + vpc_endpoints_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints/post' + response: + mediaType: application/json + openAPIDocKey: '200' + vpc_endpoints_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints/get' + response: + mediaType: application/json + openAPIDocKey: '200' + vpc_endpoints_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints~1{vpc_endpoint_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + vpc_endpoints_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1vpc-endpoints~1{vpc_endpoint_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/vpc_endpoints/methods/vpc_endpoints_get + - $ref: >- + #/components/x-stackQL-resources/vpc_endpoints/methods/vpc_endpoints_list + insert: + - $ref: >- + #/components/x-stackQL-resources/vpc_endpoints/methods/vpc_endpoints_create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/vpc_endpoints/methods/vpc_endpoints_delete + replace: [] + workspaces: + id: databricks_account.provisioning.workspaces + name: workspaces + title: Workspaces + methods: + workspaces_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces/post' + response: + mediaType: application/json + openAPIDocKey: '200' + workspaces_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces/get' + response: + mediaType: application/json + openAPIDocKey: '200' + workspaces_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + workspaces_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + workspaces_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/workspaces/methods/workspaces_get' + - $ref: >- + #/components/x-stackQL-resources/workspaces/methods/workspaces_list + insert: + - $ref: >- + #/components/x-stackQL-resources/workspaces/methods/workspaces_create + update: + - $ref: >- + #/components/x-stackQL-resources/workspaces/methods/workspaces_update + delete: + - $ref: >- + #/components/x-stackQL-resources/workspaces/methods/workspaces_delete + replace: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/settings.yaml b/providers/src/databricks_account/v00.00.00000/services/settings.yaml index ae5a1c1a..c1e8d62d 100644 --- a/providers/src/databricks_account/v00.00.00000/services/settings.yaml +++ b/providers/src/databricks_account/v00.00.00000/services/settings.yaml @@ -1,985 +1,4340 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Settings API - description: Account Settings -servers: -- url: https://accounts.cloud.databricks.com -paths: - /api/2.0/accounts/{account_id}/settings/types/shield_csp_enablement_ac/names/default: - get: - operationId: compliance-security-profile-get - externalDocs: - url: https://docs.databricks.com/api/account/cspenablementaccount/get - x-stackQL-resource: compliance_security_profile - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - csp_enablement_account: - type: object - properties: - is_enforced: - type: boolean - compliance_standards: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the compliance security profile setting for new workspaces. - patch: - operationId: compliance-security-profile-update - externalDocs: - url: https://docs.databricks.com/api/account/cspenablementaccount/update - x-stackQL-resource: compliance_security_profile - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - csp_enablement_account: - type: object - properties: - is_enforced: - type: boolean - compliance_standards: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - csp_enablement_account: - is_enforced: true - compliance_standards: - - NONE - field_mask: string - description: Details required to update a setting. - description: Updates the value of the compliance security profile setting for - new workspaces. - /api/2.0/accounts/{account_id}/settings/types/shield_esm_enablement_ac/names/default: - get: - operationId: enhanced-security-monitoring-get - externalDocs: - url: https://docs.databricks.com/api/account/esmenablementaccount/get - x-stackQL-resource: enhanced_security_monitoring - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - esm_enablement_account: - type: object - properties: - is_enforced: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the enhanced security monitoring setting for new workspaces. - patch: - operationId: enhanced-security-monitoring-update - externalDocs: - url: https://docs.databricks.com/api/account/esmenablementaccount/update - x-stackQL-resource: enhanced_security_monitoring - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - esm_enablement_account: - type: object - properties: - is_enforced: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - esm_enablement_account: - is_enforced: true - field_mask: string - description: Details required to update a setting. - description: Updates the value of the enhanced security monitoring setting for - new workspaces. - /api/2.0/accounts/{account_id}/ip-access-lists: - post: - operationId: ip-access-lists-create - externalDocs: - url: https://docs.databricks.com/api/account/accountipaccesslists/create - x-stackQL-resource: ip_access_lists - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - ip_access_list: - type: object - properties: - list_id: - type: string - label: - type: string - ip_addresses: - type: array - items: - type: string - address_count: - type: integer - list_type: - type: string - created_at: - type: integer - created_by: - type: integer - updated_at: - type: integer - updated_by: - type: integer - enabled: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - label: - type: required - list_type: - type: string - ip_addresses: - type: required - example: - label: Office VPN - list_type: ALLOW - ip_addresses: - - 192.168.100.0/22 - description: Details required to configure a block list or allow list. - description: Creates an IP access list for the account. - get: - operationId: ip-access-lists-list - externalDocs: - url: https://docs.databricks.com/api/account/accountipaccesslists/list - x-stackQL-resource: ip_access_lists - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - ip_access_lists: - type: array - items: - type: object - properties: - address_count: - type: integer - created_at: - type: integer - created_by: - type: integer - enabled: - type: boolean - ip_addresses: - type: array - items: - type: string - label: - type: string - list_id: - type: string - list_type: - type: string - updated_at: - type: integer - updated_by: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.ip_access_lists - description: Gets all IP access lists for the specified account. - /api/2.0/accounts/{account_id}/ip-access-lists/{ip_access_list_id}: - delete: - operationId: ip-access-lists-delete - externalDocs: - url: https://docs.databricks.com/api/account/accountipaccesslists/delete - x-stackQL-resource: ip_access_lists - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes an IP access list, specified by its list ID. - get: - operationId: ip-access-lists-get - externalDocs: - url: https://docs.databricks.com/api/account/accountipaccesslists/get - x-stackQL-resource: ip_access_lists - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - ip_access_list: - type: object - properties: - list_id: - type: string - label: - type: string - ip_addresses: - type: array - items: - type: string - address_count: - type: integer - list_type: - type: string - created_at: - type: integer - created_by: - type: integer - updated_at: - type: integer - updated_by: - type: integer - enabled: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.ip_access_list - description: Gets an IP access list, specified by its list ID. - put: - operationId: ip-access-lists-replace - externalDocs: - url: https://docs.databricks.com/api/account/accountipaccesslists/replace - x-stackQL-resource: ip_access_lists - x-stackQL-method: replace - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - label: - type: required - list_type: - type: string - ip_addresses: - type: required - enabled: - type: string - example: - label: Office VPN - list_type: ALLOW - ip_addresses: - - 192.168.100.0/22 - enabled: true - description: Details required to replace an IP access list. - description: Replaces an IP access list, specified by its ID. - patch: - operationId: ip-access-lists-update - externalDocs: - url: https://docs.databricks.com/api/account/accountipaccesslists/update - x-stackQL-resource: ip_access_lists - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: string - in: path - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - label: - type: string - list_type: - type: string - ip_addresses: - type: Array of string - enabled: - type: boolean - example: - label: Office VPN - list_type: ALLOW - ip_addresses: - - 192.168.100.0/22 - enabled: true - description: Details required to update an IP access list. - description: Updates an existing IP access list, specified by its ID. - /api/2.0/accounts/{account_id}/network-connectivity-configs: - post: - operationId: network-connectivity-createnetworkconnectivityconfiguration - externalDocs: - url: https://docs.databricks.com/api/account/networkconnectivity/createnetworkconnectivityconfiguration - x-stackQL-resource: network_connectivity - x-stackQL-method: createnetworkconnectivityconfiguration - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: The network connectivity configuration was successfully created. - content: - application/json: - schema: - type: object - properties: - network_connectivity_config_id: - type: string - account_id: - type: string - name: - type: string - region: - type: string - egress_config: - type: object - properties: - default_rules: - type: object - properties: - aws_stable_ip_rule: - type: object - properties: - cidr_blocks: - type: array - items: - type: string - creation_time: - type: integer - updated_time: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - region: - type: uuid - example: - name: 9ffa9185-7453-4fb2-aa6a-3105a6ae83a8 - region: string - description: Properties of the new network connectivity configuration. - description: Creates a network connectivity configuration (NCC), which provides - stable IP CIDR blocks that are associated with your workspace. You can assign - an NCC to one or more workspaces in the same region. Once assigned, the workspace - serverless compute resources use the same set of stable IP CIDR blocks to - access your resources. - get: - operationId: network-connectivity-listnetworkconnectivityconfigurations - externalDocs: - url: https://docs.databricks.com/api/account/networkconnectivity/listnetworkconnectivityconfigurations - x-stackQL-resource: network_connectivity - x-stackQL-method: listnetworkconnectivityconfigurations - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: page_token - in: query - responses: - '200': - description: The network connectivity configuration list was successfully - retrieved. - content: - application/json: - schema: - type: object - properties: - items: - type: array - items: - type: object - properties: - network_connectivity_config_id: - type: string - account_id: - type: string - name: - type: string - region: - type: string - egress_config: - type: object - properties: - default_rules: - type: object - properties: - aws_stable_ip_rule: - type: object - properties: - cidr_blocks: - type: array - items: - type: string - creation_time: - type: integer - updated_time: - type: integer - next_page_token: - type: string - description: Gets an array of network connectivity configurations. - /api/2.0/accounts/{account_id}/network-connectivity-configs/{network_connectivity_config_id}: - delete: - operationId: network-connectivity-deletenetworkconnectivityconfiguration - externalDocs: - url: https://docs.databricks.com/api/account/networkconnectivity/deletenetworkconnectivityconfiguration - x-stackQL-resource: network_connectivity - x-stackQL-method: deletenetworkconnectivityconfiguration - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: network_connectivity_config_id - required: true - description: uuid - in: path - responses: - '200': - description: The network connectivity config was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes a network connectivity configuration. - get: - operationId: network-connectivity-getnetworkconnectivityconfiguration - externalDocs: - url: https://docs.databricks.com/api/account/networkconnectivity/getnetworkconnectivityconfiguration - x-stackQL-resource: network_connectivity - x-stackQL-method: getnetworkconnectivityconfiguration - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: network_connectivity_config_id - required: true - description: uuid - in: path - responses: - '200': - description: The network connectivity configuration was successfully returned. - content: - application/json: - schema: - type: object - properties: - network_connectivity_config_id: - type: string - account_id: - type: string - name: - type: string - region: - type: string - egress_config: - type: object - properties: - default_rules: - type: object - properties: - aws_stable_ip_rule: - type: object - properties: - cidr_blocks: - type: array - items: - type: string - creation_time: - type: integer - updated_time: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a network connectivity configuration. -components: - x-stackQL-resources: - compliance_security_profile: - id: databricks_account.settings.compliance_security_profile - name: compliance_security_profile - title: Compliance_security_profile - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_csp_enablement_ac~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_csp_enablement_ac~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/compliance_security_profile/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/compliance_security_profile/methods/update' - replace: [] - delete: [] - exec: [] - enhanced_security_monitoring: - id: databricks_account.settings.enhanced_security_monitoring - name: enhanced_security_monitoring - title: Enhanced_security_monitoring - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_esm_enablement_ac~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_esm_enablement_ac~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/enhanced_security_monitoring/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/enhanced_security_monitoring/methods/update' - replace: [] - delete: [] - exec: [] - ip_access_lists: - id: databricks_account.settings.ip_access_lists - name: ip_access_lists - title: Ip_access_lists - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.ip_access_lists - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.ip_access_list - replace: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/get' - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/update' - replace: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/replace' - delete: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/delete' - exec: [] - network_connectivity: - id: databricks_account.settings.network_connectivity - name: network_connectivity - title: Network_connectivity - methods: - createnetworkconnectivityconfiguration: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listnetworkconnectivityconfigurations: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletenetworkconnectivityconfiguration: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getnetworkconnectivityconfiguration: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/network_connectivity/methods/getnetworkconnectivityconfiguration' - - $ref: '#/components/x-stackQL-resources/network_connectivity/methods/listnetworkconnectivityconfigurations' - insert: - - $ref: '#/components/x-stackQL-resources/network_connectivity/methods/createnetworkconnectivityconfiguration' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/network_connectivity/methods/deletenetworkconnectivityconfiguration' - exec: [] +openapi: 3.0.0 +info: + title: Databricks Settings API (account) + description: >- + OpenAPI specification for the Databricks settings service (account-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.settings +servers: + - url: https://accounts.cloud.databricks.com +paths: + /api/2.0/accounts/{account_id}/ip-access-lists: + post: + operationId: account_ip_access_lists_create + summary: Creates an IP access list for the account. + tags: + - settings + - account_ip_access_lists + description: >- + Creates an IP access list for the account. + + + A list can be an allow list or a block list. See the top of this file + for a description of how the + + server treats allow lists and block lists at runtime. + + + When creating or updating an IP access list: + + + * For all allow lists and block lists combined, the API supports a + maximum of 1000 IP/CIDR values, + + where one CIDR counts as a single value. Attempts to exceed that number + return error 400 with + + `error_code` value `QUOTA_EXCEEDED`. * If the new list would block the + calling user's current IP, + + error 400 is returned with `error_code` value `INVALID_STATE`. + + + It can take a few minutes for the changes to take effect. + + + :param label: str + Label for the IP access list. This **cannot** be empty. + :param list_type: :class:`ListType` + + :param ip_addresses: List[str] (optional) + + + :returns: :class:`CreateIpAccessListResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + label: + type: string + description: Label for the IP access list. This **cannot** be empty. + list_type: + type: string + description: ':param ip_addresses: List[str] (optional)' + ip_addresses: + type: string + required: + - label + - list_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateIpAccessListResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIpAccessListsAPI + get: + operationId: account_ip_access_lists_list + summary: Gets all IP access lists for the specified account. + tags: + - settings + - account_ip_access_lists + description: |- + Gets all IP access lists for the specified account. + + + :returns: Iterator over :class:`IpAccessListInfo` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetIpAccessListsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIpAccessListsAPI + /api/2.0/accounts/{account_id}/ip-access-lists/{ip_access_list_id}: + delete: + operationId: account_ip_access_lists_delete + summary: Deletes an IP access list, specified by its list ID. + tags: + - settings + - account_ip_access_lists + description: |- + Deletes an IP access list, specified by its list ID. + + :param ip_access_list_id: str + The ID for the corresponding IP access list + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIpAccessListsAPI + get: + operationId: account_ip_access_lists_get + summary: Gets an IP access list, specified by its list ID. + tags: + - settings + - account_ip_access_lists + description: |- + Gets an IP access list, specified by its list ID. + + :param ip_access_list_id: str + The ID for the corresponding IP access list + + :returns: :class:`GetIpAccessListResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetIpAccessListResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIpAccessListsAPI + put: + operationId: account_ip_access_lists_replace + summary: Replaces an IP access list, specified by its ID. + tags: + - settings + - account_ip_access_lists + description: >- + Replaces an IP access list, specified by its ID. + + + A list can include allow lists and block lists. See the top of this file + for a description of how the + + server treats allow lists and block lists at run time. When replacing an + IP access list: * For all + + allow lists and block lists combined, the API supports a maximum of 1000 + IP/CIDR values, where one + + CIDR counts as a single value. Attempts to exceed that number return + error 400 with `error_code` value + + `QUOTA_EXCEEDED`. * If the resulting list would block the calling user's + current IP, error 400 is + + returned with `error_code` value `INVALID_STATE`. It can take a few + minutes for the changes to take + + effect. + + + :param ip_access_list_id: str + The ID for the corresponding IP access list + :param label: str + Label for the IP access list. This **cannot** be empty. + :param list_type: :class:`ListType` + + :param enabled: bool + Specifies whether this IP access list is enabled. + :param ip_addresses: List[str] (optional) + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + requestBody: + content: + application/json: + schema: + type: object + properties: + label: + type: string + description: Label for the IP access list. This **cannot** be empty. + list_type: + type: string + description: >- + :param enabled: bool Specifies whether this IP access list + is enabled. + enabled: + type: boolean + ip_addresses: + type: string + description: List[str] (optional) + required: + - label + - list_type + - enabled + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIpAccessListsAPI + patch: + operationId: account_ip_access_lists_update + summary: Updates an existing IP access list, specified by its ID. + tags: + - settings + - account_ip_access_lists + description: >- + Updates an existing IP access list, specified by its ID. + + + A list can include allow lists and block lists. See the top of this file + for a description of how the + + server treats allow lists and block lists at run time. + + + When updating an IP access list: + + + * For all allow lists and block lists combined, the API supports a + maximum of 1000 IP/CIDR values, + + where one CIDR counts as a single value. Attempts to exceed that number + return error 400 with + + `error_code` value `QUOTA_EXCEEDED`. * If the updated list would block + the calling user's current IP, + + error 400 is returned with `error_code` value `INVALID_STATE`. + + + It can take a few minutes for the changes to take effect. + + + :param ip_access_list_id: str + The ID for the corresponding IP access list + :param enabled: bool (optional) + Specifies whether this IP access list is enabled. + :param ip_addresses: List[str] (optional) + + :param label: str (optional) + Label for the IP access list. This **cannot** be empty. + :param list_type: :class:`ListType` (optional) + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + requestBody: + content: + application/json: + schema: + type: object + properties: + enabled: + type: string + description: Specifies whether this IP access list is enabled. + ip_addresses: + type: string + description: >- + :param label: str (optional) Label for the IP access list. + This **cannot** be empty. + label: + type: string + list_type: + type: string + description: ':class:`ListType` (optional)' + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountIpAccessListsAPI + /api/2.0/accounts/{account_id}/network-connectivity-configs: + post: + operationId: network_connectivity_create_network_connectivity_configuration + summary: >- + Creates a network connectivity configuration (NCC), which provides + stable Azure service subnets when + tags: + - settings + - network_connectivity + description: >- + Creates a network connectivity configuration (NCC), which provides + stable Azure service subnets when + + accessing your Azure Storage accounts. You can also use a network + connectivity configuration to create + + Databricks managed private endpoints so that Databricks serverless + compute resources privately access + + your resources. + + + **IMPORTANT**: After you create the network connectivity configuration, + you must assign one or more + + workspaces to the new network connectivity configuration. You can share + one network connectivity + + configuration with multiple workspaces from the same Azure region within + the same Databricks account. + + See [configure serverless secure connectivity]. + + + [configure serverless secure connectivity]: + https://learn.microsoft.com/azure/databricks/security/network/serverless-network-security + + + :param network_connectivity_config: + :class:`CreateNetworkConnectivityConfiguration` + + + :returns: :class:`NetworkConnectivityConfiguration` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + network_connectivity_config: + type: string + description: ':returns: :class:`NetworkConnectivityConfiguration`' + required: + - network_connectivity_config + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NetworkConnectivityConfiguration' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + get: + operationId: network_connectivity_list_network_connectivity_configurations + summary: Gets an array of network connectivity configurations. + tags: + - settings + - network_connectivity + description: |- + Gets an array of network connectivity configurations. + + :param page_token: str (optional) + Pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`NetworkConnectivityConfiguration` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/ListNetworkConnectivityConfigurationsResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + /api/2.0/accounts/{account_id}/network-connectivity-configs/{network_connectivity_config_id}/private-endpoint-rules: + post: + operationId: network_connectivity_create_private_endpoint_rule + summary: >- + Create a private endpoint rule for the specified network connectivity + config object. Once the object + tags: + - settings + - network_connectivity + description: >- + Create a private endpoint rule for the specified network connectivity + config object. Once the object + + is created, Databricks asynchronously provisions a new Azure private + endpoint to your specified Azure + + resource. + + + **IMPORTANT**: You must use Azure portal or other Azure tools to approve + the private endpoint to + + complete the connection. To get the information of the private endpoint + created, make a `GET` request + + on the new private endpoint rule. See [serverless private link]. + + + [serverless private link]: + https://learn.microsoft.com/azure/databricks/security/network/serverless-network-security/serverless-private-link + + + :param network_connectivity_config_id: str + Your Network Connectivity Configuration ID. + :param private_endpoint_rule: :class:`CreatePrivateEndpointRule` + + + :returns: :class:`NccPrivateEndpointRule` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_connectivity_config_id + in: path + required: true + schema: + type: string + description: Your Network Connectivity Configuration ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + private_endpoint_rule: + type: string + description: ':returns: :class:`NccPrivateEndpointRule`' + required: + - private_endpoint_rule + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NccPrivateEndpointRule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + get: + operationId: network_connectivity_list_private_endpoint_rules + summary: Gets an array of private endpoint rules. + tags: + - settings + - network_connectivity + description: |- + Gets an array of private endpoint rules. + + :param network_connectivity_config_id: str + Your Network Connectvity Configuration ID. + :param page_token: str (optional) + Pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`NccPrivateEndpointRule` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_connectivity_config_id + in: path + required: true + schema: + type: string + description: Your Network Connectvity Configuration ID. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListPrivateEndpointRulesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + /api/2.0/accounts/{account_id}/network-connectivity-configs/{network_connectivity_config_id}: + delete: + operationId: network_connectivity_delete_network_connectivity_configuration + summary: Deletes a network connectivity configuration. + tags: + - settings + - network_connectivity + description: |- + Deletes a network connectivity configuration. + + :param network_connectivity_config_id: str + Your Network Connectivity Configuration ID. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_connectivity_config_id + in: path + required: true + schema: + type: string + description: Your Network Connectivity Configuration ID. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + get: + operationId: network_connectivity_get_network_connectivity_configuration + summary: Gets a network connectivity configuration. + tags: + - settings + - network_connectivity + description: |- + Gets a network connectivity configuration. + + :param network_connectivity_config_id: str + Your Network Connectivity Configuration ID. + + :returns: :class:`NetworkConnectivityConfiguration` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_connectivity_config_id + in: path + required: true + schema: + type: string + description: Your Network Connectivity Configuration ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NetworkConnectivityConfiguration' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + /api/2.0/accounts/{account_id}/network-connectivity-configs/{network_connectivity_config_id}/private-endpoint-rules/{private_endpoint_rule_id}: + delete: + operationId: network_connectivity_delete_private_endpoint_rule + summary: >- + Initiates deleting a private endpoint rule. If the connection state is + PENDING or EXPIRED, the private + tags: + - settings + - network_connectivity + description: >- + Initiates deleting a private endpoint rule. If the connection state is + PENDING or EXPIRED, the private + + endpoint is immediately deleted. Otherwise, the private endpoint is + deactivated and will be deleted + + after seven days of deactivation. When a private endpoint is + deactivated, the `deactivated` field is + + set to `true` and the private endpoint is not available to your + serverless compute resources. + + + :param network_connectivity_config_id: str + Your Network Connectvity Configuration ID. + :param private_endpoint_rule_id: str + Your private endpoint rule ID. + + :returns: :class:`NccPrivateEndpointRule` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_connectivity_config_id + in: path + required: true + schema: + type: string + description: Your Network Connectvity Configuration ID. + - name: private_endpoint_rule_id + in: path + required: true + schema: + type: string + description: Your private endpoint rule ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NccPrivateEndpointRule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + get: + operationId: network_connectivity_get_private_endpoint_rule + summary: Gets the private endpoint rule. + tags: + - settings + - network_connectivity + description: |- + Gets the private endpoint rule. + + :param network_connectivity_config_id: str + Your Network Connectvity Configuration ID. + :param private_endpoint_rule_id: str + Your private endpoint rule ID. + + :returns: :class:`NccPrivateEndpointRule` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_connectivity_config_id + in: path + required: true + schema: + type: string + description: Your Network Connectvity Configuration ID. + - name: private_endpoint_rule_id + in: path + required: true + schema: + type: string + description: Your private endpoint rule ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NccPrivateEndpointRule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + patch: + operationId: network_connectivity_update_private_endpoint_rule + summary: >- + Updates a private endpoint rule. Currently only a private endpoint rule + to customer-managed resources + tags: + - settings + - network_connectivity + description: >- + Updates a private endpoint rule. Currently only a private endpoint rule + to customer-managed resources + + is allowed to be updated. + + + :param network_connectivity_config_id: str + The ID of a network connectivity configuration, which is the parent resource of this private + endpoint rule object. + :param private_endpoint_rule_id: str + Your private endpoint rule ID. + :param private_endpoint_rule: :class:`UpdatePrivateEndpointRule` + + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + :returns: :class:`NccPrivateEndpointRule` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_connectivity_config_id + in: path + required: true + schema: + type: string + description: >- + The ID of a network connectivity configuration, which is the parent + resource of this private endpoint rule object. + - name: private_endpoint_rule_id + in: path + required: true + schema: + type: string + description: Your private endpoint rule ID. + - name: update_mask + in: query + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + private_endpoint_rule: + type: string + description: >- + :param update_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. + required: + - private_endpoint_rule + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NccPrivateEndpointRule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkConnectivityAPI + /api/2.0/accounts/{account_id}/network-policies: + post: + operationId: network_policies_create_network_policy_rpc + summary: >- + Creates a new network policy to manage which network destinations can be + accessed from the Databricks + tags: + - settings + - network_policies + description: >- + Creates a new network policy to manage which network destinations can be + accessed from the Databricks + + environment. + + + :param network_policy: :class:`AccountNetworkPolicy` + Network policy configuration details. + + :returns: :class:`AccountNetworkPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + network_policy: + type: string + description: Network policy configuration details. + required: + - network_policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountNetworkPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkPoliciesAPI + get: + operationId: network_policies_list_network_policies_rpc + summary: Gets an array of network policies. + tags: + - settings + - network_policies + description: |- + Gets an array of network policies. + + :param page_token: str (optional) + Pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`AccountNetworkPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListNetworkPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkPoliciesAPI + /api/2.0/accounts/{account_id}/network-policies/{network_policy_id}: + delete: + operationId: network_policies_delete_network_policy_rpc + summary: Deletes a network policy. Cannot be called on 'default-policy'. + tags: + - settings + - network_policies + description: |- + Deletes a network policy. Cannot be called on 'default-policy'. + + :param network_policy_id: str + The unique identifier of the network policy to delete. + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_policy_id + in: path + required: true + schema: + type: string + description: The unique identifier of the network policy to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkPoliciesAPI + get: + operationId: network_policies_get_network_policy_rpc + summary: Gets a network policy. + tags: + - settings + - network_policies + description: |- + Gets a network policy. + + :param network_policy_id: str + The unique identifier of the network policy to retrieve. + + :returns: :class:`AccountNetworkPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_policy_id + in: path + required: true + schema: + type: string + description: The unique identifier of the network policy to retrieve. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountNetworkPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkPoliciesAPI + put: + operationId: network_policies_update_network_policy_rpc + summary: >- + Updates a network policy. This allows you to modify the configuration of + a network policy. + tags: + - settings + - network_policies + description: >- + Updates a network policy. This allows you to modify the configuration of + a network policy. + + + :param network_policy_id: str + The unique identifier for the network policy. + :param network_policy: :class:`AccountNetworkPolicy` + Updated network policy configuration details. + + :returns: :class:`AccountNetworkPolicy` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: network_policy_id + in: path + required: true + schema: + type: string + description: The unique identifier for the network policy. + requestBody: + content: + application/json: + schema: + type: object + properties: + network_policy: + type: string + description: Updated network policy configuration details. + required: + - network_policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountNetworkPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NetworkPoliciesAPI + /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/network: + get: + operationId: workspace_network_configuration_get_workspace_network_option_rpc + summary: >- + Gets the network option for a workspace. Every workspace has exactly one + network policy binding, with + tags: + - settings + - workspace_network_configuration + description: >- + Gets the network option for a workspace. Every workspace has exactly one + network policy binding, with + + 'default-policy' used if no explicit assignment exists. + + + :param workspace_id: int + The workspace ID. + + :returns: :class:`WorkspaceNetworkOption` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: The workspace ID. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspaceNetworkOption' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceNetworkConfigurationAPI + put: + operationId: workspace_network_configuration_update_workspace_network_option_rpc + summary: >- + Updates the network option for a workspace. This operation associates + the workspace with the specified + tags: + - settings + - workspace_network_configuration + description: >- + Updates the network option for a workspace. This operation associates + the workspace with the specified + + network policy. To revert to the default policy, specify + 'default-policy' as the network_policy_id. + + + :param workspace_id: int + The workspace ID. + :param workspace_network_option: :class:`WorkspaceNetworkOption` + The network option details for the workspace. + + :returns: :class:`WorkspaceNetworkOption` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: workspace_id + in: path + required: true + schema: + type: integer + description: The workspace ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + workspace_network_option: + type: string + description: The network option details for the workspace. + required: + - workspace_network_option + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspaceNetworkOption' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceNetworkConfigurationAPI +components: + schemas: + AccountIpAccessEnable: + type: object + properties: + acct_ip_acl_enable: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - acct_ip_acl_enable + AccountNetworkPolicy: + type: object + properties: + account_id: + type: string + egress: + $ref: '#/components/schemas/NetworkPolicyEgress' + description: The network policies applying for egress traffic. + network_policy_id: + type: string + description: The unique identifier for the network policy. + AibiDashboardEmbeddingAccessPolicy: + type: object + properties: + access_policy_type: + $ref: >- + #/components/schemas/AibiDashboardEmbeddingAccessPolicyAccessPolicyType + required: + - access_policy_type + AibiDashboardEmbeddingAccessPolicySetting: + type: object + properties: + aibi_dashboard_embedding_access_policy: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicy' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - aibi_dashboard_embedding_access_policy + AibiDashboardEmbeddingApprovedDomains: + type: object + properties: + approved_domains: + type: array + items: + type: string + AibiDashboardEmbeddingApprovedDomainsSetting: + type: object + properties: + aibi_dashboard_embedding_approved_domains: + $ref: '#/components/schemas/AibiDashboardEmbeddingApprovedDomains' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - aibi_dashboard_embedding_approved_domains + AutomaticClusterUpdateSetting: + type: object + properties: + automatic_cluster_update_workspace: + $ref: '#/components/schemas/ClusterAutoRestartMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - automatic_cluster_update_workspace + BooleanMessage: + type: object + properties: + value: + type: boolean + ClusterAutoRestartMessage: + type: object + properties: + can_toggle: + type: boolean + enabled: + type: boolean + enablement_details: + $ref: '#/components/schemas/ClusterAutoRestartMessageEnablementDetails' + maintenance_window: + $ref: '#/components/schemas/ClusterAutoRestartMessageMaintenanceWindow' + restart_even_if_no_updates_available: + type: boolean + ClusterAutoRestartMessageEnablementDetails: + type: object + properties: + forced_for_compliance_mode: + type: boolean + description: The feature is force enabled if compliance mode is active + unavailable_for_disabled_entitlement: + type: boolean + description: >- + The feature is unavailable if the corresponding entitlement disabled + (see getShieldEntitlementEnable) + unavailable_for_non_enterprise_tier: + type: boolean + description: >- + The feature is unavailable if the customer doesn't have enterprise + tier + description: >- + Contains an information about the enablement status judging (e.g. + whether the enterprise tier is + enabled) This is only additional information that MUST NOT be used to decide whether the setting + is enabled or not. This is intended to use only for purposes like showing an error message to + the customer with the additional details. For example, using these details we can check why + exactly the feature is disabled for this customer. + ClusterAutoRestartMessageMaintenanceWindow: + type: object + properties: + week_day_based_schedule: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule + ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule: + type: object + properties: + day_of_week: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowDayOfWeek + frequency: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency + window_start_time: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWindowStartTime + ClusterAutoRestartMessageMaintenanceWindowWindowStartTime: + type: object + properties: + hours: + type: integer + minutes: + type: integer + ComplianceSecurityProfile: + type: object + properties: + compliance_standards: + type: array + items: + $ref: '#/components/schemas/ComplianceStandard' + description: Set by customers when they request Compliance Security Profile (CSP) + is_enabled: + type: boolean + description: 'SHIELD feature: CSP' + ComplianceSecurityProfileSetting: + type: object + properties: + compliance_security_profile_workspace: + $ref: '#/components/schemas/ComplianceSecurityProfile' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - compliance_security_profile_workspace + Config: + type: object + properties: + email: + $ref: '#/components/schemas/EmailConfig' + generic_webhook: + $ref: '#/components/schemas/GenericWebhookConfig' + microsoft_teams: + $ref: '#/components/schemas/MicrosoftTeamsConfig' + pagerduty: + $ref: '#/components/schemas/PagerdutyConfig' + slack: + $ref: '#/components/schemas/SlackConfig' + CreateIpAccessListResponse: + type: object + properties: + ip_access_list: + $ref: '#/components/schemas/IpAccessListInfo' + description: An IP access list was successfully created. + CreateNetworkConnectivityConfiguration: + type: object + properties: + name: + type: string + description: >- + The name of the network connectivity configuration. The name can + contain alphanumeric characters, hyphens, and underscores. The + length must be between 3 and 30 characters. The name must match the + regular expression ^[0-9a-zA-Z-_]{3,30}$ + region: + type: string + description: >- + The region for the network connectivity configuration. Only + workspaces in the same region can be attached to the network + connectivity configuration. + required: + - name + - region + description: Properties of the new network connectivity configuration. + CreateOboTokenResponse: + type: object + properties: + token_info: + $ref: '#/components/schemas/TokenInfo' + token_value: + type: string + description: Value of the token. + description: An on-behalf token was successfully created for the service principal. + CreatePrivateEndpointRule: + type: object + properties: + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints to customer-managed private endpoint + services. Domain names of target private link service. When updating + this field, the full list of target domain_names must be specified. + endpoint_service: + type: string + description: >- + The full target AWS endpoint service name that connects to the + destination resources of the private endpoint. + error_message: + type: string + group_id: + type: string + description: >- + Not used by customer-managed private endpoint services. The + sub-resource type (group ID) of the target resource. Note that to + connect to workspace root storage (root DBFS), you need two + endpoints, one for blob and one for dfs. + resource_id: + type: string + description: The Azure resource ID of the target resource. + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + CreateTokenResponse: + type: object + properties: + token_info: + $ref: '#/components/schemas/PublicTokenInfo' + token_value: + type: string + description: The value of the new token. + CspEnablementAccount: + type: object + properties: + compliance_standards: + type: array + items: + $ref: '#/components/schemas/ComplianceStandard' + description: >- + Set by customers when they request Compliance Security Profile (CSP) + Invariants are enforced in Settings policy. + is_enforced: + type: boolean + description: Enforced = it cannot be overriden at workspace level. + description: Account level policy for CSP + CspEnablementAccountSetting: + type: object + properties: + csp_enablement_account: + $ref: '#/components/schemas/CspEnablementAccount' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - csp_enablement_account + CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRule: + type: object + properties: + account_id: + type: string + description: >- + Databricks account ID. You can find your account ID from the + Accounts Console. + connection_state: + $ref: >- + #/components/schemas/CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRulePrivateLinkConnectionState + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + deactivated: + type: boolean + description: Whether this private endpoint is deactivated. + deactivated_at: + type: integer + description: Time in epoch milliseconds when this object was deactivated. + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards a VPC endpoint service for + customer-managed VPC endpoint service. The target AWS resource FQDNs + accessible via the VPC endpoint service. When updating this field, + we perform full update on this field. Please ensure a full list of + desired domain_names is provided. + enabled: + type: boolean + description: >- + Only used by private endpoints towards an AWS S3 service. Update + this field to activate/deactivate this private endpoint to allow + egress access from serverless compute resources. + endpoint_service: + type: string + description: >- + The full target AWS endpoint service name that connects to the + destination resources of the private endpoint. + error_message: + type: string + network_connectivity_config_id: + type: string + description: >- + The ID of a network connectivity configuration, which is the parent + resource of this private endpoint rule object. + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + rule_id: + type: string + description: The ID of a private endpoint rule. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + vpc_endpoint_id: + type: string + description: >- + The AWS VPC endpoint ID. You can use this ID to identify VPC + endpoint created by Databricks. + description: >- + Properties of the new private endpoint rule. Note that for private + endpoints towards a VPC + endpoint service behind a customer-managed NLB, you must approve the endpoint in AWS console + after initialization. + DashboardEmailSubscriptions: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + DefaultNamespaceSetting: + type: object + properties: + namespace: + $ref: '#/components/schemas/StringMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - namespace + description: >- + This represents the setting configuration for the default namespace in + the Databricks workspace. + Setting the default catalog for the workspace determines the catalog that is used when queries + do not reference a fully qualified 3 level name. For example, if the default catalog is set to + 'retail_prod' then a query 'SELECT * FROM myTable' would reference the object + 'retail_prod.default.myTable' (the schema 'default' is always assumed). This setting requires a + restart of clusters and SQL warehouses to take effect. Additionally, the default namespace only + applies when using Unity Catalog-enabled compute. + DefaultWarehouseId: + type: object + properties: + string_val: + $ref: '#/components/schemas/StringMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - string_val + DeleteAccountIpAccessEnableResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteAibiDashboardEmbeddingAccessPolicySettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteAibiDashboardEmbeddingApprovedDomainsSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDashboardEmailSubscriptionsResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDefaultNamespaceSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDefaultWarehouseIdResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDisableLegacyAccessResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDisableLegacyDbfsResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDisableLegacyFeaturesResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteLlmProxyPartnerPoweredWorkspaceResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeletePersonalComputeSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteRestrictWorkspaceAdminsSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteSqlResultsDownloadResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DisableLegacyAccess: + type: object + properties: + disable_legacy_access: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - disable_legacy_access + DisableLegacyDbfs: + type: object + properties: + disable_legacy_dbfs: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - disable_legacy_dbfs + DisableLegacyFeatures: + type: object + properties: + disable_legacy_features: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - disable_legacy_features + EgressNetworkPolicy: + type: object + properties: + internet_access: + $ref: '#/components/schemas/EgressNetworkPolicyInternetAccessPolicy' + description: The access policy enforced for egress traffic to the internet. + description: >- + The network policies applying for egress traffic. This message is used + by the UI/REST API. We + translate this message to the format expected by the dataplane in Lakehouse Network Manager (for + the format expected by the dataplane, see networkconfig.textproto). + EgressNetworkPolicyInternetAccessPolicy: + type: object + properties: + allowed_internet_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyInternetDestination + allowed_storage_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyStorageDestination + log_only_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyLogOnlyMode + description: >- + Optional. If not specified, assume the policy is enforced for all + workloads. + restriction_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyRestrictionMode + EgressNetworkPolicyInternetAccessPolicyInternetDestination: + type: object + properties: + destination: + type: string + protocol: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationFilteringProtocol + type: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationType + description: >- + Users can specify accessible internet destinations when outbound access + is restricted. We only + support domain name (FQDN) destinations for the time being, though going forwards we want to + support host names and IP addresses. + EgressNetworkPolicyInternetAccessPolicyLogOnlyMode: + type: object + properties: + log_only_mode_type: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyLogOnlyModeLogOnlyModeType + workloads: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyLogOnlyModeWorkloadType + EgressNetworkPolicyInternetAccessPolicyStorageDestination: + type: object + properties: + allowed_paths: + type: array + items: + type: string + azure_container: + type: string + azure_dns_zone: + type: string + azure_storage_account: + type: string + azure_storage_service: + type: string + bucket_name: + type: string + region: + type: string + type: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyStorageDestinationStorageDestinationType + description: Users can specify accessible storage destinations. + EgressNetworkPolicyNetworkAccessPolicy: + type: object + properties: + restriction_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyRestrictionMode + allowed_internet_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyInternetDestination + description: >- + List of internet destinations that serverless workloads are allowed + to access when in RESTRICTED_ACCESS mode. + allowed_storage_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyStorageDestination + description: >- + List of storage destinations that serverless workloads are allowed + to access when in RESTRICTED_ACCESS mode. + policy_enforcement: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcement + description: >- + Optional. When policy_enforcement is not provided, we default to + ENFORCE_MODE_ALL_SERVICES + required: + - restriction_mode + EgressNetworkPolicyNetworkAccessPolicyInternetDestination: + type: object + properties: + destination: + type: string + description: >- + The internet destination to which access will be allowed. Format + dependent on the destination type. + internet_destination_type: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyInternetDestinationInternetDestinationType + description: >- + Users can specify accessible internet destinations when outbound access + is restricted. We only + support DNS_NAME (FQDN format) destinations for the time being. Going forward we may extend + support to host names and IP addresses. + EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcement: + type: object + properties: + dry_run_mode_product_filter: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementDryRunModeProductFilter + enforcement_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementEnforcementMode + description: >- + The mode of policy enforcement. ENFORCED blocks traffic that + violates policy, while DRY_RUN only logs violations without + blocking. When not specified, defaults to ENFORCED. + EgressNetworkPolicyNetworkAccessPolicyStorageDestination: + type: object + properties: + azure_storage_account: + type: string + description: The Azure storage account name. + azure_storage_service: + type: string + description: The Azure storage service type (blob, dfs, etc.). + bucket_name: + type: string + region: + type: string + storage_destination_type: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyStorageDestinationStorageDestinationType + description: Users can specify accessible storage destinations. + EmailConfig: + type: object + properties: + addresses: + type: array + items: + type: string + Empty: + type: object + properties: {} + EnableExportNotebook: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + EnableNotebookTableClipboard: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + EnableResultsDownloading: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + EnhancedSecurityMonitoring: + type: object + properties: + is_enabled: + type: boolean + description: 'SHIELD feature: ESM' + EnhancedSecurityMonitoringSetting: + type: object + properties: + enhanced_security_monitoring_workspace: + $ref: '#/components/schemas/EnhancedSecurityMonitoring' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - enhanced_security_monitoring_workspace + EsmEnablementAccount: + type: object + properties: + is_enforced: + type: boolean + description: Account level policy for ESM + EsmEnablementAccountSetting: + type: object + properties: + esm_enablement_account: + $ref: '#/components/schemas/EsmEnablementAccount' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - esm_enablement_account + ExchangeToken: + type: object + properties: + credential: + type: string + description: The requested token. + credentialEolTime: + type: integer + description: >- + The end-of-life timestamp of the token. The value is in milliseconds + since the Unix epoch. + ownerId: + type: integer + description: User ID of the user that owns this token. + scopes: + type: array + items: + type: string + description: The scopes of access granted in the token. + tokenType: + $ref: '#/components/schemas/TokenType' + description: The type of this exchange token + description: The exchange token is the result of the token exchange with the IdP + ExchangeTokenResponse: + type: object + properties: + values: + type: array + items: + $ref: '#/components/schemas/ExchangeToken' + description: Exhanged tokens were successfully returned. + FetchIpAccessListResponse: + type: object + properties: + ip_access_list: + $ref: '#/components/schemas/IpAccessListInfo' + description: An IP access list was successfully returned. + GenericWebhookConfig: + type: object + properties: + password: + type: string + password_set: + type: boolean + description: '[Output-Only] Whether password is set.' + url: + type: string + description: '[Input-Only] URL for webhook.' + url_set: + type: boolean + description: '[Output-Only] Whether URL is set.' + username: + type: string + description: '[Input-Only][Optional] Username for webhook.' + username_set: + type: boolean + description: '[Output-Only] Whether username is set.' + GetIpAccessListResponse: + type: object + properties: + ip_access_list: + $ref: '#/components/schemas/IpAccessListInfo' + GetIpAccessListsResponse: + type: object + properties: + ip_access_lists: + type: array + items: + $ref: '#/components/schemas/IpAccessListInfo' + description: IP access lists were successfully returned. + GetTokenPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/TokenPermissionsDescription' + GetTokenResponse: + type: object + properties: + token_info: + $ref: '#/components/schemas/TokenInfo' + description: Token with specified Token ID was successfully returned. + IpAccessListInfo: + type: object + properties: + address_count: + type: integer + description: Total number of IP or CIDR values. + created_at: + type: integer + description: Creation timestamp in milliseconds. + created_by: + type: integer + description: User ID of the user who created this list. + enabled: + type: boolean + description: Specifies whether this IP access list is enabled. + ip_addresses: + type: array + items: + type: string + label: + type: string + description: Label for the IP access list. This **cannot** be empty. + list_id: + type: string + description: Universally unique identifier (UUID) of the IP access list. + list_type: + $ref: '#/components/schemas/ListType' + updated_at: + type: integer + description: Update timestamp in milliseconds. + updated_by: + type: integer + description: User ID of the user who updated this list. + description: Definition of an IP Access list + ListIpAccessListResponse: + type: object + properties: + ip_access_lists: + type: array + items: + $ref: '#/components/schemas/IpAccessListInfo' + description: IP access lists were successfully returned. + ListNetworkConnectivityConfigurationsResponse: + type: object + properties: + items: + type: array + items: + $ref: '#/components/schemas/NetworkConnectivityConfiguration' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If null, + there are no more results to show. + description: The network connectivity configuration list was successfully retrieved. + ListNetworkPoliciesResponse: + type: object + properties: + items: + type: array + items: + $ref: '#/components/schemas/AccountNetworkPolicy' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If null, + there are no more results to show. + ListNotificationDestinationsResponse: + type: object + properties: + next_page_token: + type: string + results: + type: array + items: + $ref: '#/components/schemas/ListNotificationDestinationsResult' + ListNotificationDestinationsResult: + type: object + properties: + destination_type: + $ref: '#/components/schemas/DestinationType' + display_name: + type: string + description: The display name for the notification destination. + id: + type: string + description: UUID identifying notification destination. + ListPrivateEndpointRulesResponse: + type: object + properties: + items: + type: array + items: + $ref: '#/components/schemas/NccPrivateEndpointRule' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If null, + there are no more results to show. + description: The private endpoint rule list was successfully retrieved. + ListPublicTokensResponse: + type: object + properties: + token_infos: + type: array + items: + $ref: '#/components/schemas/PublicTokenInfo' + ListTokensResponse: + type: object + properties: + token_infos: + type: array + items: + $ref: '#/components/schemas/TokenInfo' + description: Token metadata of each user-created token in the workspace + description: Tokens were successfully returned. + LlmProxyPartnerPoweredAccount: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + LlmProxyPartnerPoweredEnforce: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + LlmProxyPartnerPoweredWorkspace: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + MicrosoftTeamsConfig: + type: object + properties: + app_id: + type: string + app_id_set: + type: boolean + description: '[Output-Only] Whether App ID is set.' + auth_secret: + type: string + description: '[Input-Only] Secret for Microsoft Teams App authentication.' + auth_secret_set: + type: boolean + description: '[Output-Only] Whether secret is set.' + channel_url: + type: string + description: '[Input-Only] Channel URL for Microsoft Teams App.' + channel_url_set: + type: boolean + description: '[Output-Only] Whether Channel URL is set.' + tenant_id: + type: string + description: '[Input-Only] Tenant ID for Microsoft Teams App.' + tenant_id_set: + type: boolean + description: '[Output-Only] Whether Tenant ID is set.' + url: + type: string + description: '[Input-Only] URL for Microsoft Teams webhook.' + url_set: + type: boolean + description: '[Output-Only] Whether URL is set.' + NccAwsStableIpRule: + type: object + properties: + cidr_blocks: + type: array + items: + type: string + description: >- + The list of stable IP CIDR blocks from which Databricks network + traffic originates when accessing your resources. + description: >- + The stable AWS IP CIDR blocks. You can use these to configure the + firewall of your resources to + allow traffic from your Databricks workspace. + NccAzurePrivateEndpointRule: + type: object + properties: + connection_state: + $ref: '#/components/schemas/NccAzurePrivateEndpointRuleConnectionState' + description: >- + The current status of this private endpoint. The private endpoint + rules are effective only if the connection state is ESTABLISHED. + Remember that you must approve new endpoints on your resources in + the Azure portal before they take effect. The possible values are: - + INIT: (deprecated) The endpoint has been created and pending + approval. - PENDING: The endpoint has been created and pending + approval. - ESTABLISHED: The endpoint has been approved and is ready + to use in your serverless compute resources. - REJECTED: Connection + was rejected by the private link resource owner. - DISCONNECTED: + Connection was removed by the private link resource owner, the + private endpoint becomes informative and should be deleted for + clean-up. - EXPIRED: If the endpoint was created but not approved in + 14 days, it will be EXPIRED. + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + deactivated: + type: boolean + description: Whether this private endpoint is deactivated. + deactivated_at: + type: integer + description: Time in epoch milliseconds when this object was deactivated. + domain_names: + type: array + items: + type: string + description: >- + Not used by customer-managed private endpoint services. Domain names + of target private link service. When updating this field, the full + list of target domain_names must be specified. + endpoint_name: + type: string + description: The name of the Azure private endpoint resource. + error_message: + type: string + group_id: + type: string + description: >- + Only used by private endpoints to Azure first-party services. The + sub-resource type (group ID) of the target resource. Note that to + connect to workspace root storage (root DBFS), you need two + endpoints, one for blob and one for dfs. + network_connectivity_config_id: + type: string + description: >- + The ID of a network connectivity configuration, which is the parent + resource of this private endpoint rule object. + resource_id: + type: string + description: The Azure resource ID of the target resource. + rule_id: + type: string + description: The ID of a private endpoint rule. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + NccAzureServiceEndpointRule: + type: object + properties: + subnets: + type: array + items: + type: string + description: >- + The list of subnets from which Databricks network traffic originates + when accessing your Azure resources. + target_region: + type: string + description: The Azure region in which this service endpoint rule applies.. + target_services: + type: array + items: + $ref: '#/components/schemas/EgressResourceType' + description: The Azure services to which this service endpoint rule applies to. + description: >- + The stable Azure service endpoints. You can configure the firewall of + your Azure resources to + allow traffic from your Databricks serverless compute resources. + NccEgressConfig: + type: object + properties: + default_rules: + $ref: '#/components/schemas/NccEgressDefaultRules' + target_rules: + $ref: '#/components/schemas/NccEgressTargetRules' + description: >- + The network connectivity rules that configured for each + destinations. These rules override default rules. + NccEgressDefaultRules: + type: object + properties: + aws_stable_ip_rule: + $ref: '#/components/schemas/NccAwsStableIpRule' + azure_service_endpoint_rule: + $ref: '#/components/schemas/NccAzureServiceEndpointRule' + description: Default rules don't have specific targets. + NccEgressTargetRules: + type: object + properties: + aws_private_endpoint_rules: + type: array + items: + $ref: >- + #/components/schemas/CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRule + description: >- + AWS private endpoint rule controls the AWS private endpoint based + egress rules. + azure_private_endpoint_rules: + type: array + items: + $ref: '#/components/schemas/NccAzurePrivateEndpointRule' + description: >- + Target rule controls the egress rules that are dedicated to specific + resources. + NccPrivateEndpointRule: + type: object + properties: + account_id: + type: string + description: >- + Databricks account ID. You can find your account ID from the + Accounts Console. + connection_state: + $ref: >- + #/components/schemas/NccPrivateEndpointRulePrivateLinkConnectionState + description: >- + The current status of this private endpoint. The private endpoint + rules are effective only if the connection state is ESTABLISHED. + Remember that you must approve new endpoints on your resources in + the Cloud console before they take effect. The possible values are: + - PENDING: The endpoint has been created and pending approval. - + ESTABLISHED: The endpoint has been approved and is ready to use in + your serverless compute resources. - REJECTED: Connection was + rejected by the private link resource owner. - DISCONNECTED: + Connection was removed by the private link resource owner, the + private endpoint becomes informative and should be deleted for + clean-up. - EXPIRED: If the endpoint was created but not approved in + 14 days, it will be EXPIRED. - CREATING: The endpoint creation is in + progress. Once successfully created, the state will transition to + PENDING. - CREATE_FAILED: The endpoint creation failed. You can + check the error_message field for more details. + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + deactivated: + type: boolean + description: Whether this private endpoint is deactivated. + deactivated_at: + type: integer + description: Time in epoch milliseconds when this object was deactivated. + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints to customer-managed private endpoint + services. Domain names of target private link service. When updating + this field, the full list of target domain_names must be specified. + enabled: + type: boolean + description: >- + Only used by private endpoints towards an AWS S3 service. Update + this field to activate/deactivate this private endpoint to allow + egress access from serverless compute resources. + endpoint_name: + type: string + description: The name of the Azure private endpoint resource. + endpoint_service: + type: string + description: >- + The full target AWS endpoint service name that connects to the + destination resources of the private endpoint. + error_message: + type: string + group_id: + type: string + description: >- + Not used by customer-managed private endpoint services. The + sub-resource type (group ID) of the target resource. Note that to + connect to workspace root storage (root DBFS), you need two + endpoints, one for blob and one for dfs. + network_connectivity_config_id: + type: string + description: >- + The ID of a network connectivity configuration, which is the parent + resource of this private endpoint rule object. + resource_id: + type: string + description: The Azure resource ID of the target resource. + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + rule_id: + type: string + description: The ID of a private endpoint rule. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + vpc_endpoint_id: + type: string + description: >- + The AWS VPC endpoint ID. You can use this ID to identify the VPC + endpoint created by Databricks. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + NetworkConnectivityConfiguration: + type: object + properties: + account_id: + type: string + description: >- + Your Databricks account ID. You can find your account ID in your + Databricks accounts console. + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + egress_config: + $ref: '#/components/schemas/NccEgressConfig' + description: >- + The network connectivity rules that apply to network traffic from + your serverless compute resources. + name: + type: string + description: >- + The name of the network connectivity configuration. The name can + contain alphanumeric characters, hyphens, and underscores. The + length must be between 3 and 30 characters. The name must match the + regular expression ^[0-9a-zA-Z-_]{3,30}$ + network_connectivity_config_id: + type: string + description: Databricks network connectivity configuration ID. + region: + type: string + description: >- + The region for the network connectivity configuration. Only + workspaces in the same region can be attached to the network + connectivity configuration. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + description: Properties of the new network connectivity configuration. + NetworkPolicyEgress: + type: object + properties: + network_access: + $ref: '#/components/schemas/EgressNetworkPolicyNetworkAccessPolicy' + description: The access policy enforced for egress traffic to the internet. + description: >- + The network policies applying for egress traffic. This message is used + by the UI/REST API. We + translate this message to the format expected by the dataplane in Lakehouse Network Manager (for + the format expected by the dataplane, see networkconfig.textproto). This policy should be + consistent with [[com.databricks.api.proto.settingspolicy.EgressNetworkPolicy]]. Details see + API-design: https://docs.google.com/document/d/1DKWO_FpZMCY4cF2O62LpwII1lx8gsnDGG-qgE3t3TOA/ + NotificationDestination: + type: object + properties: + config: + $ref: '#/components/schemas/Config' + destination_type: + $ref: '#/components/schemas/DestinationType' + description: >- + [Output-only] The type of the notification destination. The type can + not be changed once set. + display_name: + type: string + description: The display name for the notification destination. + id: + type: string + description: UUID identifying notification destination. + PagerdutyConfig: + type: object + properties: + integration_key: + type: string + integration_key_set: + type: boolean + description: '[Output-Only] Whether integration key is set.' + PartitionId: + type: object + properties: + workspaceId: + type: integer + description: The ID of the workspace. + description: Partition by workspace or account + PersonalComputeMessage: + type: object + properties: + value: + $ref: '#/components/schemas/PersonalComputeMessageEnum' + required: + - value + PersonalComputeSetting: + type: object + properties: + personal_compute: + $ref: '#/components/schemas/PersonalComputeMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - personal_compute + PublicTokenInfo: + type: object + properties: + comment: + type: string + creation_time: + type: integer + description: Server time (in epoch milliseconds) when the token was created. + expiry_time: + type: integer + description: >- + Server time (in epoch milliseconds) when the token will expire, or + -1 if not applicable. + token_id: + type: string + description: The ID of this token. + RestrictWorkspaceAdminsMessage: + type: object + properties: + status: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessageStatus' + required: + - status + RestrictWorkspaceAdminsSetting: + type: object + properties: + restrict_workspace_admins: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - restrict_workspace_admins + RevokeTokenResponse: + type: object + properties: {} + SlackConfig: + type: object + properties: + channel_id: + type: string + channel_id_set: + type: boolean + description: '[Output-Only] Whether channel ID is set.' + oauth_token: + type: string + description: '[Input-Only] OAuth token for Slack authentication.' + oauth_token_set: + type: boolean + description: '[Output-Only] Whether OAuth token is set.' + url: + type: string + description: '[Input-Only] URL for Slack destination.' + url_set: + type: boolean + description: '[Output-Only] Whether URL is set.' + SqlResultsDownload: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + StringMessage: + type: object + properties: + value: + type: string + TokenAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/TokenPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + TokenAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/TokenPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + TokenInfo: + type: object + properties: + comment: + type: string + created_by_id: + type: integer + description: User ID of the user that created the token. + created_by_username: + type: string + description: Username of the user that created the token. + creation_time: + type: integer + description: Timestamp when the token was created. + expiry_time: + type: integer + description: Timestamp when the token expires. + last_used_day: + type: integer + description: >- + Approximate timestamp for the day the token was last used. Accurate + up to 1 day. + owner_id: + type: integer + description: User ID of the user that owns the token. + token_id: + type: string + description: ID of the token. + workspace_id: + type: integer + description: >- + If applicable, the ID of the workspace that the token was created + in. + TokenPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/TokenPermissionLevel' + TokenPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/TokenAccessControlResponse' + object_id: + type: string + object_type: + type: string + TokenPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/TokenPermissionLevel' + UpdatePrivateEndpointRule: + type: object + properties: + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints to customer-managed private endpoint + services. Domain names of target private link service. When updating + this field, the full list of target domain_names must be specified. + enabled: + type: boolean + description: >- + Only used by private endpoints towards an AWS S3 service. Update + this field to activate/deactivate this private endpoint to allow + egress access from serverless compute resources. + error_message: + type: string + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + WorkspaceNetworkOption: + type: object + properties: + network_policy_id: + type: string + workspace_id: + type: integer + description: The workspace ID. + AibiDashboardEmbeddingAccessPolicyAccessPolicyType: + type: string + enum: + - ALLOW_ALL_DOMAINS + - ALLOW_APPROVED_DOMAINS + - DENY_ALL_DOMAINS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowDayOfWeek: + type: string + enum: + - FRIDAY + - MONDAY + - SATURDAY + - SUNDAY + - THURSDAY + - TUESDAY + - WEDNESDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency: + type: string + enum: + - EVERY_WEEK + - FIRST_AND_THIRD_OF_MONTH + - FIRST_OF_MONTH + - FOURTH_OF_MONTH + - SECOND_AND_FOURTH_OF_MONTH + - SECOND_OF_MONTH + - THIRD_OF_MONTH + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ComplianceStandard: + type: string + enum: + - CANADA_PROTECTED_B + - CYBER_ESSENTIAL_PLUS + - FEDRAMP_HIGH + - FEDRAMP_IL5 + - FEDRAMP_MODERATE + - GERMANY_C5 + - GERMANY_TISAX + - HIPAA + - HITRUST + - IRAP_PROTECTED + - ISMAP + - ITAR_EAR + - K_FSI + - NONE + - PCI_DSS + description: >- + Compliance standard for SHIELD customers. See README.md for how + instructions of how to add new + + standards. + CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRulePrivateLinkConnectionState: + type: string + enum: + - CREATE_FAILED + - CREATING + - DISCONNECTED + - ESTABLISHED + - EXPIRED + - PENDING + - REJECTED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DestinationType: + type: string + enum: + - EMAIL + - MICROSOFT_TEAMS + - PAGERDUTY + - SLACK + - WEBHOOK + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationFilteringProtocol: + type: string + enum: + - TCP + description: >- + The filtering protocol used by the DP. For private and public preview, + SEG will only support TCP + + filtering (i.e. DNS based filtering, filtering by destination IP + address), so protocol will be + + set to TCP by default and hidden from the user. In the future, users may + be able to select HTTP + + filtering (i.e. SNI based filtering, filtering by FQDN). + EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationType: + type: string + enum: + - FQDN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyInternetAccessPolicyLogOnlyModeLogOnlyModeType: + type: string + enum: + - ALL_SERVICES + - SELECTED_SERVICES + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyInternetAccessPolicyLogOnlyModeWorkloadType: + type: string + enum: + - DBSQL + - ML_SERVING + description: >- + The values should match the list of workloads used in + networkconfig.proto + EgressNetworkPolicyInternetAccessPolicyRestrictionMode: + type: string + enum: + - FULL_ACCESS + - PRIVATE_ACCESS_ONLY + - RESTRICTED_ACCESS + description: >- + At which level can Databricks and Databricks managed compute access + Internet. FULL_ACCESS: + + Databricks can access Internet. No blocking rules will apply. + RESTRICTED_ACCESS: Databricks can + + only access explicitly allowed internet and storage destinations, as + well as UC connections and + + external locations. PRIVATE_ACCESS_ONLY (not used): Databricks can only + access destinations via + + private link. + EgressNetworkPolicyInternetAccessPolicyStorageDestinationStorageDestinationType: + type: string + enum: + - AWS_S3 + - AZURE_STORAGE + - CLOUDFLARE_R2 + - GOOGLE_CLOUD_STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyNetworkAccessPolicyInternetDestinationInternetDestinationType: + type: string + enum: + - DNS_NAME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementDryRunModeProductFilter: + type: string + enum: + - DBSQL + - ML_SERVING + description: >- + The values should match the list of workloads used in + networkconfig.proto + EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementEnforcementMode: + type: string + enum: + - DRY_RUN + - ENFORCED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyNetworkAccessPolicyRestrictionMode: + type: string + enum: + - FULL_ACCESS + - RESTRICTED_ACCESS + description: >- + At which level can Databricks and Databricks managed compute access + Internet. FULL_ACCESS: + + Databricks can access Internet. No blocking rules will apply. + RESTRICTED_ACCESS: Databricks can + + only access explicitly allowed internet and storage destinations, as + well as UC connections and + + external locations. + EgressNetworkPolicyNetworkAccessPolicyStorageDestinationStorageDestinationType: + type: string + enum: + - AWS_S3 + - AZURE_STORAGE + - GOOGLE_CLOUD_STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressResourceType: + type: string + enum: + - AZURE_BLOB_STORAGE + description: >- + The target resources that are supported by Network Connectivity Config. + Note: some egress types + + can support general types that are not defined in EgressResourceType. + E.g.: Azure private + + endpoint supports private link enabled Azure services. + ListType: + type: string + enum: + - ALLOW + - BLOCK + description: >- + Type of IP access list. Valid values are as follows and are + case-sensitive: + + + * `ALLOW`: An allow list. Include this IP or range. * `BLOCK`: A block + list. Exclude this IP or + + range. IP addresses in the block list are excluded even if they are + included in an allow list. + NccAzurePrivateEndpointRuleConnectionState: + type: string + enum: + - CREATE_FAILED + - CREATING + - DISCONNECTED + - ESTABLISHED + - EXPIRED + - INIT + - PENDING + - REJECTED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + NccPrivateEndpointRulePrivateLinkConnectionState: + type: string + enum: + - CREATE_FAILED + - CREATING + - DISCONNECTED + - ESTABLISHED + - EXPIRED + - PENDING + - REJECTED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PersonalComputeMessageEnum: + type: string + enum: + - DELEGATE + - 'ON' + description: >- + ON: Grants all users in all workspaces access to the Personal Compute + default policy, allowing + + all users to create single-machine compute resources. DELEGATE: Moves + access control for the + + Personal Compute default policy to individual workspaces and requires a + workspace’s users or + + groups to be added to the ACLs of that workspace’s Personal Compute + default policy before they + + will be able to create compute resources through that policy. + RestrictWorkspaceAdminsMessageStatus: + type: string + enum: + - ALLOW_ALL + - RESTRICT_TOKENS_AND_JOB_RUN_AS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TokenPermissionLevel: + type: string + enum: + - CAN_USE + description: Permission level + TokenType: + type: string + enum: + - ARCLIGHT_AZURE_EXCHANGE_TOKEN + - ARCLIGHT_AZURE_EXCHANGE_TOKEN_WITH_USER_DELEGATION_KEY + - ARCLIGHT_MULTI_TENANT_AZURE_EXCHANGE_TOKEN + - ARCLIGHT_MULTI_TENANT_AZURE_EXCHANGE_TOKEN_WITH_USER_DELEGATION_KEY + - AZURE_ACTIVE_DIRECTORY_TOKEN + description: >- + The type of token request. As of now, only + `AZURE_ACTIVE_DIRECTORY_TOKEN` is supported. + x-stackQL-resources: + account_ip_access_lists: + id: databricks_account.settings.account_ip_access_lists + name: account_ip_access_lists + title: Account Ip Access Lists + methods: + account_ip_access_lists_create: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists/post' + response: + mediaType: application/json + openAPIDocKey: '200' + account_ip_access_lists_list: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.ip_access_lists + account_ip_access_lists_delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + account_ip_access_lists_get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.ip_access_list + account_ip_access_lists_replace: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + account_ip_access_lists_update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1ip-access-lists~1{ip_access_list_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_ip_access_lists/methods/account_ip_access_lists_get + - $ref: >- + #/components/x-stackQL-resources/account_ip_access_lists/methods/account_ip_access_lists_list + insert: + - $ref: >- + #/components/x-stackQL-resources/account_ip_access_lists/methods/account_ip_access_lists_create + update: + - $ref: >- + #/components/x-stackQL-resources/account_ip_access_lists/methods/account_ip_access_lists_update + delete: + - $ref: >- + #/components/x-stackQL-resources/account_ip_access_lists/methods/account_ip_access_lists_delete + replace: + - $ref: >- + #/components/x-stackQL-resources/account_ip_access_lists/methods/account_ip_access_lists_replace + network_connectivity: + id: databricks_account.settings.network_connectivity + name: network_connectivity + title: Network Connectivity + methods: + create_network_connectivity_configuration: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs/post + response: + mediaType: application/json + openAPIDocKey: '200' + list_network_connectivity_configurations: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.items + delete_network_connectivity_configuration: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get_network_connectivity_configuration: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/network_connectivity/methods/get_network_connectivity_configuration + - $ref: >- + #/components/x-stackQL-resources/network_connectivity/methods/list_network_connectivity_configurations + insert: + - $ref: >- + #/components/x-stackQL-resources/network_connectivity/methods/create_network_connectivity_configuration + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/network_connectivity/methods/delete_network_connectivity_configuration + replace: [] + private_endpoint_rules: + id: databricks_account.settings.private_endpoint_rules + name: private_endpoint_rules + title: Private Endpoint Rules + methods: + create_private_endpoint_rule: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}~1private-endpoint-rules/post + response: + mediaType: application/json + openAPIDocKey: '200' + list_private_endpoint_rules: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}~1private-endpoint-rules/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.items + delete_private_endpoint_rule: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}~1private-endpoint-rules~1{private_endpoint_rule_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get_private_endpoint_rule: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}~1private-endpoint-rules~1{private_endpoint_rule_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update_private_endpoint_rule: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-connectivity-configs~1{network_connectivity_config_id}~1private-endpoint-rules~1{private_endpoint_rule_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/private_endpoint_rules/methods/get_private_endpoint_rule + - $ref: >- + #/components/x-stackQL-resources/private_endpoint_rules/methods/list_private_endpoint_rules + insert: + - $ref: >- + #/components/x-stackQL-resources/private_endpoint_rules/methods/create_private_endpoint_rule + update: + - $ref: >- + #/components/x-stackQL-resources/private_endpoint_rules/methods/update_private_endpoint_rule + delete: + - $ref: >- + #/components/x-stackQL-resources/private_endpoint_rules/methods/delete_private_endpoint_rule + replace: [] + network_policies: + id: databricks_account.settings.network_policies + name: network_policies + title: Network Policies + methods: + create_network_policy_rpc: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1network-policies/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list_network_policies_rpc: + operation: + $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1network-policies/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.items + delete_network_policy_rpc: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-policies~1{network_policy_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get_network_policy_rpc: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-policies~1{network_policy_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update_network_policy_rpc: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1network-policies~1{network_policy_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/network_policies/methods/get_network_policy_rpc + - $ref: >- + #/components/x-stackQL-resources/network_policies/methods/list_network_policies_rpc + insert: + - $ref: >- + #/components/x-stackQL-resources/network_policies/methods/create_network_policy_rpc + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/network_policies/methods/delete_network_policy_rpc + replace: + - $ref: >- + #/components/x-stackQL-resources/network_policies/methods/update_network_policy_rpc + workspace_network_configuration: + id: databricks_account.settings.workspace_network_configuration + name: workspace_network_configuration + title: Workspace Network Configuration + methods: + get_workspace_network_option_rpc: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1network/get + response: + mediaType: application/json + openAPIDocKey: '200' + update_workspace_network_option_rpc: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1network/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/workspace_network_configuration/methods/get_workspace_network_option_rpc + insert: [] + update: [] + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/workspace_network_configuration/methods/update_workspace_network_option_rpc diff --git a/providers/src/databricks_account/v00.00.00000/services/settingsv2.yaml b/providers/src/databricks_account/v00.00.00000/services/settingsv2.yaml new file mode 100644 index 00000000..8eb312f0 --- /dev/null +++ b/providers/src/databricks_account/v00.00.00000/services/settingsv2.yaml @@ -0,0 +1,995 @@ +openapi: 3.0.0 +info: + title: Databricks Settingsv2 API (account) + description: >- + OpenAPI specification for the Databricks settingsv2 service (account-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.settingsv2 +servers: + - url: https://accounts.cloud.databricks.com +paths: + /api/2.1/accounts/{account_id}/settings/{name}: + get: + operationId: account_settings_v2_get_public_account_setting + summary: >- + Get a setting value at account level. See + :method:settingsv2/listaccountsettingsmetadata for list of + tags: + - settingsv2 + - account_settings_v2 + description: >- + Get a setting value at account level. See + :method:settingsv2/listaccountsettingsmetadata for list of + + setting available via public APIs at account level. + + + :param name: str + + + :returns: :class:`Setting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: ':returns: :class:`Setting`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Setting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountSettingsV2API + patch: + operationId: account_settings_v2_patch_public_account_setting + summary: >- + Patch a setting value at account level. See + :method:settingsv2/listaccountsettingsmetadata for list of + tags: + - settingsv2 + - account_settings_v2 + description: >- + Patch a setting value at account level. See + :method:settingsv2/listaccountsettingsmetadata for list of + + setting available via public APIs at account level. To determine the + correct field to include in a + + patch request, refer to the type field of the setting returned in the + + :method:settingsv2/listaccountsettingsmetadata response. + + + Note: Page refresh is required for changes to take effect in UI. + + + :param name: str + + :param setting: :class:`Setting` + + + :returns: :class:`Setting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: ':param setting: :class:`Setting`' + requestBody: + content: + application/json: + schema: + type: object + properties: + setting: + type: string + required: + - setting + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Setting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountSettingsV2API + /api/2.1/accounts/{account_id}/users/{user_id}/settings/{name}: + get: + operationId: account_settings_v2_get_public_account_user_preference + summary: >- + Get a user preference for a specific user. User preferences are personal + settings that allow + tags: + - settingsv2 + - account_settings_v2 + description: >- + Get a user preference for a specific user. User preferences are personal + settings that allow + + individual customization without affecting other users. See + + :method:settingsv2/listaccountuserpreferencesmetadata for list of user + preferences available via + + public APIs. + + + :param user_id: str + User ID of the user whose setting is being retrieved. + :param name: str + User Setting name. + + :returns: :class:`UserPreference` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: user_id + in: path + required: true + schema: + type: string + description: User ID of the user whose setting is being retrieved. + - name: name + in: path + required: true + schema: + type: string + description: User Setting name. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UserPreference' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountSettingsV2API + patch: + operationId: account_settings_v2_patch_public_account_user_preference + summary: >- + Update a user preference for a specific user. User preferences are + personal settings that allow + tags: + - settingsv2 + - account_settings_v2 + description: >- + Update a user preference for a specific user. User preferences are + personal settings that allow + + individual customization without affecting other users. See + + :method:settingsv2/listaccountuserpreferencesmetadata for list of user + preferences available via + + public APIs. + + + Note: Page refresh is required for changes to take effect in UI. + + + :param user_id: str + User ID of the user whose setting is being updated. + :param name: str + + :param setting: :class:`UserPreference` + + + :returns: :class:`UserPreference` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: user_id + in: path + required: true + schema: + type: string + description: User ID of the user whose setting is being updated. + - name: name + in: path + required: true + schema: + type: string + description: ':param setting: :class:`UserPreference`' + requestBody: + content: + application/json: + schema: + type: object + properties: + setting: + type: string + required: + - setting + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UserPreference' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountSettingsV2API + /api/2.1/accounts/{account_id}/settings-metadata: + get: + operationId: account_settings_v2_list_account_settings_metadata + summary: >- + List valid setting keys and metadata. These settings are available to be + referenced via GET + tags: + - settingsv2 + - account_settings_v2 + description: >- + List valid setting keys and metadata. These settings are available to be + referenced via GET + + :method:settingsv2/getpublicaccountsetting and PATCH + :method:settingsv2/patchpublicaccountsetting APIs + + + :param page_size: int (optional) + The maximum number of settings to return. The service may return fewer than this value. If + unspecified, at most 200 settings will be returned. The maximum value is 1000; values above 1000 + will be coerced to 1000. + :param page_token: str (optional) + A page token, received from a previous `ListAccountSettingsMetadataRequest` call. Provide this to + retrieve the subsequent page. + + When paginating, all other parameters provided to `ListAccountSettingsMetadataRequest` must match + the call that provided the page token. + + :returns: Iterator over :class:`SettingsMetadata` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of settings to return. The service may return + fewer than this value. If unspecified, at most 200 settings will be + returned. The maximum value is 1000; values above 1000 will be + coerced to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous + `ListAccountSettingsMetadataRequest` call. Provide this to retrieve + the subsequent page. When paginating, all other parameters provided + to `ListAccountSettingsMetadataRequest` must match the call that + provided the page token. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountSettingsMetadataResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountSettingsV2API + /api/2.1/accounts/{account_id}/users/{user_id}/settings-metadata: + get: + operationId: account_settings_v2_list_account_user_preferences_metadata + summary: >- + List valid user preferences and their metadata for a specific user. User + preferences are personal + tags: + - settingsv2 + - account_settings_v2 + description: >- + List valid user preferences and their metadata for a specific user. User + preferences are personal + + settings that allow individual customization without affecting other + users. These settings are + + available to be referenced via GET + :method:settingsv2/getpublicaccountuserpreference and PATCH + + :method:settingsv2/patchpublicaccountuserpreference APIs + + + :param user_id: str + User ID of the user whose settings metadata is being retrieved. + :param page_size: int (optional) + The maximum number of settings to return. The service may return fewer than this value. If + unspecified, at most 200 settings will be returned. The maximum value is 1000; values above 1000 + will be coerced to 1000. + :param page_token: str (optional) + A page token, received from a previous `ListAccountUserPreferencesMetadataRequest` call. Provide + this to retrieve the subsequent page. + + When paginating, all other parameters provided to `ListAccountUserPreferencesMetadataRequest` must + match the call that provided the page token. + + :returns: Iterator over :class:`SettingsMetadata` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: user_id + in: path + required: true + schema: + type: string + description: User ID of the user whose settings metadata is being retrieved. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of settings to return. The service may return + fewer than this value. If unspecified, at most 200 settings will be + returned. The maximum value is 1000; values above 1000 will be + coerced to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous + `ListAccountUserPreferencesMetadataRequest` call. Provide this to + retrieve the subsequent page. When paginating, all other parameters + provided to `ListAccountUserPreferencesMetadataRequest` must match + the call that provided the page token. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountSettingsMetadataResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountSettingsV2API +components: + schemas: + AibiDashboardEmbeddingAccessPolicy: + type: object + properties: + access_policy_type: + $ref: >- + #/components/schemas/AibiDashboardEmbeddingAccessPolicyAccessPolicyType + required: + - access_policy_type + AibiDashboardEmbeddingApprovedDomains: + type: object + properties: + approved_domains: + type: array + items: + type: string + BooleanMessage: + type: object + properties: + value: + type: boolean + ClusterAutoRestartMessage: + type: object + properties: + can_toggle: + type: boolean + enabled: + type: boolean + enablement_details: + $ref: '#/components/schemas/ClusterAutoRestartMessageEnablementDetails' + maintenance_window: + $ref: '#/components/schemas/ClusterAutoRestartMessageMaintenanceWindow' + restart_even_if_no_updates_available: + type: boolean + ClusterAutoRestartMessageEnablementDetails: + type: object + properties: + forced_for_compliance_mode: + type: boolean + description: The feature is force enabled if compliance mode is active + unavailable_for_disabled_entitlement: + type: boolean + description: >- + The feature is unavailable if the corresponding entitlement disabled + (see getShieldEntitlementEnable) + unavailable_for_non_enterprise_tier: + type: boolean + description: >- + The feature is unavailable if the customer doesn't have enterprise + tier + description: >- + Contains an information about the enablement status judging (e.g. + whether the enterprise tier is + enabled) This is only additional information that MUST NOT be used to decide whether the setting + is enabled or not. This is intended to use only for purposes like showing an error message to + the customer with the additional details. For example, using these details we can check why + exactly the feature is disabled for this customer. + ClusterAutoRestartMessageMaintenanceWindow: + type: object + properties: + week_day_based_schedule: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule + ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule: + type: object + properties: + day_of_week: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowDayOfWeek + frequency: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency + window_start_time: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWindowStartTime + ClusterAutoRestartMessageMaintenanceWindowWindowStartTime: + type: object + properties: + hours: + type: integer + minutes: + type: integer + IntegerMessage: + type: object + properties: + value: + type: integer + ListAccountSettingsMetadataResponse: + type: object + properties: + next_page_token: + type: string + settings_metadata: + type: array + items: + $ref: '#/components/schemas/SettingsMetadata' + description: List of all settings available via public APIs and their metadata + ListAccountUserPreferencesMetadataResponse: + type: object + properties: + next_page_token: + type: string + settings_metadata: + type: array + items: + $ref: '#/components/schemas/SettingsMetadata' + description: List of all settings available via public APIs and their metadata + ListWorkspaceSettingsMetadataResponse: + type: object + properties: + next_page_token: + type: string + settings_metadata: + type: array + items: + $ref: '#/components/schemas/SettingsMetadata' + description: List of all settings available via public APIs and their metadata + PersonalComputeMessage: + type: object + properties: + value: + $ref: >- + #/components/schemas/PersonalComputeMessagePersonalComputeMessageEnum + RestrictWorkspaceAdminsMessage: + type: object + properties: + status: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessageStatus' + required: + - status + Setting: + type: object + properties: + aibi_dashboard_embedding_access_policy: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicy' + aibi_dashboard_embedding_approved_domains: + $ref: '#/components/schemas/AibiDashboardEmbeddingApprovedDomains' + description: >- + Setting value for aibi_dashboard_embedding_approved_domains setting. + This is the setting value set by consumers, check + effective_aibi_dashboard_embedding_approved_domains for final + setting value. + automatic_cluster_update_workspace: + $ref: '#/components/schemas/ClusterAutoRestartMessage' + description: >- + Setting value for automatic_cluster_update_workspace setting. This + is the setting value set by consumers, check + effective_automatic_cluster_update_workspace for final setting + value. + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + description: >- + Setting value for boolean type setting. This is the setting value + set by consumers, check effective_boolean_val for final setting + value. + effective_aibi_dashboard_embedding_access_policy: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicy' + description: >- + Effective setting value for aibi_dashboard_embedding_access_policy + setting. This is the final effective value of setting. To set a + value use aibi_dashboard_embedding_access_policy. + effective_aibi_dashboard_embedding_approved_domains: + $ref: '#/components/schemas/AibiDashboardEmbeddingApprovedDomains' + description: >- + Effective setting value for + aibi_dashboard_embedding_approved_domains setting. This is the final + effective value of setting. To set a value use + aibi_dashboard_embedding_approved_domains. + effective_automatic_cluster_update_workspace: + $ref: '#/components/schemas/ClusterAutoRestartMessage' + description: >- + Effective setting value for automatic_cluster_update_workspace + setting. This is the final effective value of setting. To set a + value use automatic_cluster_update_workspace. + effective_boolean_val: + $ref: '#/components/schemas/BooleanMessage' + description: >- + Effective setting value for boolean type setting. This is the final + effective value of setting. To set a value use boolean_val. + effective_integer_val: + $ref: '#/components/schemas/IntegerMessage' + description: >- + Effective setting value for integer type setting. This is the final + effective value of setting. To set a value use integer_val. + effective_personal_compute: + $ref: '#/components/schemas/PersonalComputeMessage' + description: >- + Effective setting value for personal_compute setting. This is the + final effective value of setting. To set a value use + personal_compute. + effective_restrict_workspace_admins: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessage' + description: >- + Effective setting value for restrict_workspace_admins setting. This + is the final effective value of setting. To set a value use + restrict_workspace_admins. + effective_string_val: + $ref: '#/components/schemas/StringMessage' + description: >- + Effective setting value for string type setting. This is the final + effective value of setting. To set a value use string_val. + integer_val: + $ref: '#/components/schemas/IntegerMessage' + description: >- + Setting value for integer type setting. This is the setting value + set by consumers, check effective_integer_val for final setting + value. + name: + type: string + description: Name of the setting. + personal_compute: + $ref: '#/components/schemas/PersonalComputeMessage' + description: >- + Setting value for personal_compute setting. This is the setting + value set by consumers, check effective_personal_compute for final + setting value. + restrict_workspace_admins: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessage' + description: >- + Setting value for restrict_workspace_admins setting. This is the + setting value set by consumers, check + effective_restrict_workspace_admins for final setting value. + string_val: + $ref: '#/components/schemas/StringMessage' + description: >- + Setting value for string type setting. This is the setting value set + by consumers, check effective_string_val for final setting value. + SettingsMetadata: + type: object + properties: + description: + type: string + docs_link: + type: string + description: Link to databricks documentation for the setting + name: + type: string + description: Name of the setting. + type: + type: string + description: >- + Sample message depicting the type of the setting. To set this + setting, the value sent must match this type. + StringMessage: + type: object + properties: + value: + type: string + UserPreference: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + effective_boolean_val: + $ref: '#/components/schemas/BooleanMessage' + effective_string_val: + $ref: '#/components/schemas/StringMessage' + name: + type: string + description: Name of the setting. + string_val: + $ref: '#/components/schemas/StringMessage' + user_id: + type: string + description: User ID of the user. + description: >- + User Preference represents a user-specific setting scoped to an + individual user within an + account. Unlike workspace or account settings that apply to all users, user preferences allow + personal customization (e.g., UI theme, editor preferences) without affecting other users. + AibiDashboardEmbeddingAccessPolicyAccessPolicyType: + type: string + enum: + - ALLOW_ALL_DOMAINS + - ALLOW_APPROVED_DOMAINS + - DENY_ALL_DOMAINS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowDayOfWeek: + type: string + enum: + - FRIDAY + - MONDAY + - SATURDAY + - SUNDAY + - THURSDAY + - TUESDAY + - WEDNESDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency: + type: string + enum: + - EVERY_WEEK + - FIRST_AND_THIRD_OF_MONTH + - FIRST_OF_MONTH + - FOURTH_OF_MONTH + - SECOND_AND_FOURTH_OF_MONTH + - SECOND_OF_MONTH + - THIRD_OF_MONTH + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PersonalComputeMessagePersonalComputeMessageEnum: + type: string + enum: + - DELEGATE + - 'ON' + description: >- + ON: Grants all users in all workspaces access to the Personal Compute + default policy, allowing + + all users to create single-machine compute resources. DELEGATE: Moves + access control for the + + Personal Compute default policy to individual workspaces and requires a + workspace’s users or + + groups to be added to the ACLs of that workspace’s Personal Compute + default policy before they + + will be able to create compute resources through that policy. + RestrictWorkspaceAdminsMessageStatus: + type: string + enum: + - ALLOW_ALL + - RESTRICT_TOKENS_AND_JOB_RUN_AS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + account_settings: + id: databricks_account.settingsv2.account_settings + name: account_settings + title: Account Settings + methods: + get_public_account_setting: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1settings~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + patch_public_account_setting: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1settings~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + list_account_settings_metadata: + operation: + $ref: '#/paths/~1api~12.1~1accounts~1{account_id}~1settings-metadata/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.settings_metadata + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_settings/methods/get_public_account_setting + - $ref: >- + #/components/x-stackQL-resources/account_settings/methods/list_account_settings_metadata + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/account_settings/methods/patch_public_account_setting + delete: [] + replace: [] + account_user_preferences: + id: databricks_account.settingsv2.account_user_preferences + name: account_user_preferences + title: Account User Preferences + methods: + get_public_account_user_preference: + operation: + $ref: >- + #/paths/~1api~12.1~1accounts~1{account_id}~1users~1{user_id}~1settings~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + patch_public_account_user_preference: + operation: + $ref: >- + #/paths/~1api~12.1~1accounts~1{account_id}~1users~1{user_id}~1settings~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_user_preferences/methods/get_public_account_user_preference + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/account_user_preferences/methods/patch_public_account_user_preference + delete: [] + replace: [] + account_user_preferences_metadata: + id: databricks_account.settingsv2.account_user_preferences_metadata + name: account_user_preferences_metadata + title: Account User Preferences Metadata + methods: + list_account_user_preferences_metadata: + operation: + $ref: >- + #/paths/~1api~12.1~1accounts~1{account_id}~1users~1{user_id}~1settings-metadata/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.settings_metadata + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/account_user_preferences_metadata/methods/list_account_user_preferences_metadata + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_account/v00.00.00000/services/unity_catalog.yaml b/providers/src/databricks_account/v00.00.00000/services/unity_catalog.yaml deleted file mode 100644 index 792fe362..00000000 --- a/providers/src/databricks_account/v00.00.00000/services/unity_catalog.yaml +++ /dev/null @@ -1,1158 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Unity_catalog API - description: Unity Catalog -servers: -- url: https://accounts.cloud.databricks.com -paths: - /api/2.0/accounts/{account_id}/metastores: - post: - operationId: metastores-create - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastores/create - x-stackQL-resource: metastores - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '201': - description: The metastore was successfully created. - content: - application/json: - schema: - type: object - properties: - metastore_info: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - metastore_info: - type: object - example: - metastore_info: - name: string - storage_root: string - region: string - description: Properties of the new metastore. - description: Creates a Unity Catalog metastore. - get: - operationId: metastores-list - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastores/list - x-stackQL-resource: metastores - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: account_id - required: true - description: uuid - in: path - responses: - '200': - description: Metastores were returned successfully. - content: - application/json: - schema: - type: object - properties: - metastores: - type: array - items: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - x-stackQL-objectKey: $.metastores - description: Gets all Unity Catalog metastores associated with an account specified - by ID. - /api/2.0/accounts/{account_id}/metastores/{metastore_id}: - delete: - operationId: metastores-delete - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastores/delete - x-stackQL-resource: metastores - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - - name: force - in: query - responses: - '200': - description: The metastore was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a Unity Catalog metastore for an account, both specified - by ID. - get: - operationId: metastores-get - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastores/get - x-stackQL-resource: metastores - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '200': - description: The metastore was successfully returned. - content: - application/json: - schema: - type: object - properties: - metastore_info: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.metastore_info - description: Gets a Unity Catalog metastore from an account, both specified - by ID. - put: - operationId: metastores-update - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastores/update - x-stackQL-resource: metastores - x-stackQL-method: update - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '201': - description: The metastore update request succeeded. - content: - application/json: - schema: - type: object - properties: - metastore_info: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - metastore_info: - type: object - example: - metastore_info: - storage_root_credential_id: string - delta_sharing_recipient_token_lifetime_in_seconds: 0 - delta_sharing_organization_name: string - owner: string - privilege_model_version: string - new_name: string - delta_sharing_scope: INTERNAL - description: Properties of the metastore to change. - description: Updates an existing Unity Catalog metastore. - /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/metastores/{metastore_id}: - post: - operationId: metastore-assignments-create - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastoreassignments/create - x-stackQL-resource: metastore_assignments - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: workspace_id - required: true - description: int64 - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '201': - description: The metastore assignment was successfully created. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - metastore_assignment: - type: object - example: - metastore_assignment: - metastore_id: string - default_catalog_name: string - description: The mapping from workspace to metastore. - description: Creates an assignment to a metastore for a workspace - delete: - operationId: metastore-assignments-delete - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastoreassignments/delete - x-stackQL-resource: metastore_assignments - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: workspace_id - required: true - description: int64 - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '200': - description: The metastore assignment was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a metastore assignment to a workspace, leaving the workspace - with no metastore. - put: - operationId: metastore-assignments-update - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastoreassignments/update - x-stackQL-resource: metastore_assignments - x-stackQL-method: update - x-stackQL-verb: replace - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: workspace_id - required: true - description: int64 - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '201': - description: The metastore assignment was successfully updated. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - metastore_assignment: - type: object - example: - metastore_assignment: - metastore_id: string - default_catalog_name: string - description: The metastore assignment to update. - description: Updates an assignment to a metastore for a workspace. Currently, - only the default catalog may be updated. - /api/2.0/accounts/{account_id}/workspaces/{workspace_id}/metastore: - get: - operationId: metastore-assignments-get - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastoreassignments/get - x-stackQL-resource: metastore_assignments - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: The workspace metastore assignment was successfully returned. - content: - application/json: - schema: - type: object - properties: - metastore_assignment: - type: object - properties: - metastore_id: - type: string - workspace_id: - type: integer - default_catalog_name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.metastore_assignment - description: Gets the metastore assignment, if any, for the workspace specified - by ID. If the workspace is assigned a metastore, the mappig will be returned. - If no metastore is assigned to the workspace, the assignment will not be found - and a 404 returned. - /api/2.0/accounts/{account_id}/metastores/{metastore_id}/storage-credentials: - post: - operationId: storage-credentials-create - externalDocs: - url: https://docs.databricks.com/api/account/accountstoragecredentials/create - x-stackQL-resource: storage_credentials - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '200': - description: The new storage credential was successfully created. - content: - application/json: - schema: - type: object - properties: - credential_info: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - credential_info: - type: object - example: - credential_info: - name: string - comment: string - read_only: true - aws_iam_role: - role_arn: string - skip_validation: false - description: 'Creates a new storage credential. The request object is specific - to the cloud:' - get: - operationId: storage-credentials-list - externalDocs: - url: https://docs.databricks.com/api/account/accountstoragecredentials/list - x-stackQL-resource: storage_credentials - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '200': - description: The metastore storage credentials were successfully returned. - content: - application/json: - schema: - type: object - properties: - storage_credentials: - type: array - items: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - x-stackQL-objectKey: $.storage_credentials - description: Gets a list of all storage credentials that have been assigned - to given metastore. - /api/2.0/accounts/{account_id}/metastores/{metastore_id}/storage-credentials/{storage_credential_name}: - delete: - operationId: storage-credentials-delete - externalDocs: - url: https://docs.databricks.com/api/account/accountstoragecredentials/delete - x-stackQL-resource: storage_credentials - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - - name: storage_credential_name - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: The storage credential was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a storage credential from the metastore. The caller must - be an owner of the storage credential. - get: - operationId: storage-credentials-get - externalDocs: - url: https://docs.databricks.com/api/account/accountstoragecredentials/get - x-stackQL-resource: storage_credentials - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - - name: storage_credential_name - required: true - description: string - in: path - responses: - '200': - description: The storage credential was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - credential_info: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.credential_info - description: Gets a storage credential from the metastore. The caller must be - a metastore admin, the owner of the storage credential, or have a level of - privilege on the storage credential. - put: - operationId: storage-credentials-update - externalDocs: - url: https://docs.databricks.com/api/account/accountstoragecredentials/update - x-stackQL-resource: storage_credentials - x-stackQL-method: update - x-stackQL-verb: replace - x-numReqParams: 3 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - - name: storage_credential_name - required: true - description: string - in: path - responses: - '201': - description: The storage credential was successfully updated. - content: - application/json: - schema: - type: object - properties: - credential_info: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - credential_info: - type: object - example: - credential_info: - comment: string - read_only: true - owner: string - aws_iam_role: - role_arn: string - isolation_mode: ISOLATION_MODE_OPEN - new_name: string - skip_validation: false - force: true - description: The storage credential to update. - description: Updates a storage credential on the metastore. The caller must - be the owner of the storage credential. If the caller is a metastore admin, - only the - /api/2.0/accounts/{account_id}/metastores/{metastore_id}/workspaces: - get: - operationId: workspace-assignments-list - externalDocs: - url: https://docs.databricks.com/api/account/accountmetastoreassignments/list - x-stackQL-resource: workspace_assignments - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: account_id - required: true - description: uuid - in: path - - name: metastore_id - required: true - description: uuid - in: path - responses: - '200': - description: The metastore assignments were successfully returned. - content: - application/json: - schema: - type: object - properties: - workspace_ids: - type: array - items: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a list of all Databricks workspace IDs that have been assigned - to given metastore. -components: - x-stackQL-resources: - metastores: - id: databricks_account.unity_catalog.metastores - name: metastores - title: Metastores - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.metastores - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.metastore_info - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/metastores/methods/get' - - $ref: '#/components/x-stackQL-resources/metastores/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/metastores/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/metastores/methods/update' - delete: - - $ref: '#/components/x-stackQL-resources/metastores/methods/delete' - exec: [] - metastore_assignments: - id: databricks_account.unity_catalog.metastore_assignments - name: metastore_assignments - title: Metastore_assignments - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastores~1{metastore_id}/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastores~1{metastore_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastores~1{metastore_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1workspaces~1{workspace_id}~1metastore/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.metastore_assignment - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/update' - delete: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/delete' - exec: [] - storage_credentials: - id: databricks_account.unity_catalog.storage_credentials - name: storage_credentials - title: Storage_credentials - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.storage_credentials - delete: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials~1{storage_credential_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials~1{storage_credential_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.credential_info - update: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1storage-credentials~1{storage_credential_name}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/get' - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/update' - delete: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/delete' - exec: [] - workspace_assignments: - id: databricks_account.unity_catalog.workspace_assignments - name: workspace_assignments - title: Workspace_assignments - methods: - list: - operation: - $ref: '#/paths/~1api~12.0~1accounts~1{account_id}~1metastores~1{metastore_id}~1workspaces/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/workspace_assignments/methods/list' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/provider.yaml b/providers/src/databricks_workspace/v00.00.00000/provider.yaml index eeeb005f..5b5a7fc8 100644 --- a/providers/src/databricks_workspace/v00.00.00000/provider.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/provider.yaml @@ -1,176 +1,301 @@ -id: databricks_workspace -name: databricks_workspace -version: v00.00.00000 -providerServices: - apps: - id: apps:v00.00.00000 - name: apps - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/apps.yaml - title: Apps - version: v00.00.00000 - description: Apps - cleanrooms: - id: cleanrooms:v00.00.00000 - name: cleanrooms - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/cleanrooms.yaml - title: Cleanrooms - version: v00.00.00000 - description: Cleanrooms - compute: - id: compute:v00.00.00000 - name: compute - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/compute.yaml - title: Compute - version: v00.00.00000 - description: Compute - dbsql: - id: dbsql:v00.00.00000 - name: dbsql - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/dbsql.yaml - title: Dbsql - version: v00.00.00000 - description: Dbsql - deltalivetables: - id: deltalivetables:v00.00.00000 - name: deltalivetables - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/deltalivetables.yaml - title: Deltalivetables - version: v00.00.00000 - description: Deltalivetables - deltasharing: - id: deltasharing:v00.00.00000 - name: deltasharing - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/deltasharing.yaml - title: Deltasharing - version: v00.00.00000 - description: Deltasharing - filemanagement: - id: filemanagement:v00.00.00000 - name: filemanagement - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/filemanagement.yaml - title: Filemanagement - version: v00.00.00000 - description: Filemanagement - iam: - id: iam:v00.00.00000 - name: iam - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/iam.yaml - title: Iam - version: v00.00.00000 - description: Iam - lakeview: - id: lakeview:v00.00.00000 - name: lakeview - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/lakeview.yaml - title: Lakeview - version: v00.00.00000 - description: Lakeview - machinelearning: - id: machinelearning:v00.00.00000 - name: machinelearning - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/machinelearning.yaml - title: Machinelearning - version: v00.00.00000 - description: Machinelearning - marketplace: - id: marketplace:v00.00.00000 - name: marketplace - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/marketplace.yaml - title: Marketplace - version: v00.00.00000 - description: Marketplace - realtimeserving: - id: realtimeserving:v00.00.00000 - name: realtimeserving - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/realtimeserving.yaml - title: Realtimeserving - version: v00.00.00000 - description: Realtimeserving - repos: - id: repos:v00.00.00000 - name: repos - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/repos.yaml - title: Repos - version: v00.00.00000 - description: Repos - secrets: - id: secrets:v00.00.00000 - name: secrets - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/secrets.yaml - title: Secrets - version: v00.00.00000 - description: Secrets - unitycatalog: - id: unitycatalog:v00.00.00000 - name: unitycatalog - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/unitycatalog.yaml - title: Unitycatalog - version: v00.00.00000 - description: Unitycatalog - vectorsearch: - id: vectorsearch:v00.00.00000 - name: vectorsearch - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/vectorsearch.yaml - title: Vectorsearch - version: v00.00.00000 - description: Vectorsearch - workflows: - id: workflows:v00.00.00000 - name: workflows - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/workflows.yaml - title: Workflows - version: v00.00.00000 - description: Workflows - workspace: - id: workspace:v00.00.00000 - name: workspace - preferred: true - service: - $ref: databricks_workspace/v00.00.00000/services/workspace.yaml - title: Workspace - version: v00.00.00000 - description: Workspace -config: - auth: - type: oauth2 - client_id_env_var: DATABRICKS_CLIENT_ID - client_secret_env_var: DATABRICKS_CLIENT_SECRET - grant_type: client_credentials - token_url: https://accounts.cloud.databricks.com/oidc/accounts/{{ .__env__DATABRICKS_ACCOUNT_ID - }}/v1/token - scopes: - - all-apis +id: databricks_workspace +name: databricks_workspace +version: v00.00.00000 +providerServices: + agentbricks: + id: agentbricks:v00.00.00000 + name: agentbricks + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/agentbricks.yaml + title: Databricks Agentbricks API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks agentbricks service + (workspace-level APIs), generated from the Databricks Python SDK. + apps: + id: apps:v00.00.00000 + name: apps + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/apps.yaml + title: Databricks Apps API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks apps service (workspace-level + APIs), generated from the Databricks Python SDK. + catalog: + id: catalog:v00.00.00000 + name: catalog + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/catalog.yaml + title: Databricks Catalog API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks catalog service (workspace-level + APIs), generated from the Databricks Python SDK. + cleanrooms: + id: cleanrooms:v00.00.00000 + name: cleanrooms + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/cleanrooms.yaml + title: Databricks Cleanrooms API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks cleanrooms service + (workspace-level APIs), generated from the Databricks Python SDK. + compute: + id: compute:v00.00.00000 + name: compute + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/compute.yaml + title: Databricks Compute API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks compute service (workspace-level + APIs), generated from the Databricks Python SDK. + dashboards: + id: dashboards:v00.00.00000 + name: dashboards + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/dashboards.yaml + title: Databricks Dashboards API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks dashboards service + (workspace-level APIs), generated from the Databricks Python SDK. + database: + id: database:v00.00.00000 + name: database + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/database.yaml + title: Databricks Database API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks database service (workspace-level + APIs), generated from the Databricks Python SDK. + dataquality: + id: dataquality:v00.00.00000 + name: dataquality + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/dataquality.yaml + title: Databricks Dataquality API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks dataquality service + (workspace-level APIs), generated from the Databricks Python SDK. + files: + id: files:v00.00.00000 + name: files + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/files.yaml + title: Databricks Files API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks files service (workspace-level + APIs), generated from the Databricks Python SDK. + iam: + id: iam:v00.00.00000 + name: iam + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/iam.yaml + title: Databricks Iam API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks iam service (workspace-level + APIs), generated from the Databricks Python SDK. + iamv2: + id: iamv2:v00.00.00000 + name: iamv2 + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/iamv2.yaml + title: Databricks Iamv2 API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks iamv2 service (workspace-level + APIs), generated from the Databricks Python SDK. + jobs: + id: jobs:v00.00.00000 + name: jobs + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/jobs.yaml + title: Databricks Jobs API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks jobs service (workspace-level + APIs), generated from the Databricks Python SDK. + marketplace: + id: marketplace:v00.00.00000 + name: marketplace + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/marketplace.yaml + title: Databricks Marketplace API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks marketplace service + (workspace-level APIs), generated from the Databricks Python SDK. + ml: + id: ml:v00.00.00000 + name: ml + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/ml.yaml + title: Databricks Ml API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks ml service (workspace-level + APIs), generated from the Databricks Python SDK. + oauth2: + id: oauth2:v00.00.00000 + name: oauth2 + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/oauth2.yaml + title: Databricks Oauth2 API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks oauth2 service (workspace-level + APIs), generated from the Databricks Python SDK. + pipelines: + id: pipelines:v00.00.00000 + name: pipelines + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/pipelines.yaml + title: Databricks Pipelines API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks pipelines service + (workspace-level APIs), generated from the Databricks Python SDK. + postgres: + id: postgres:v00.00.00000 + name: postgres + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/postgres.yaml + title: Databricks Postgres API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks postgres service (workspace-level + APIs), generated from the Databricks Python SDK. + qualitymonitorv2: + id: qualitymonitorv2:v00.00.00000 + name: qualitymonitorv2 + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/qualitymonitorv2.yaml + title: Databricks Qualitymonitorv2 API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks qualitymonitorv2 service + (workspace-level APIs), generated from the Databricks Python SDK. + serving: + id: serving:v00.00.00000 + name: serving + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/serving.yaml + title: Databricks Serving API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks serving service (workspace-level + APIs), generated from the Databricks Python SDK. + settings: + id: settings:v00.00.00000 + name: settings + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/settings.yaml + title: Databricks Settings API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks settings service (workspace-level + APIs), generated from the Databricks Python SDK. + settingsv2: + id: settingsv2:v00.00.00000 + name: settingsv2 + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/settingsv2.yaml + title: Databricks Settingsv2 API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks settingsv2 service + (workspace-level APIs), generated from the Databricks Python SDK. + sharing: + id: sharing:v00.00.00000 + name: sharing + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/sharing.yaml + title: Databricks Sharing API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks sharing service (workspace-level + APIs), generated from the Databricks Python SDK. + sql: + id: sql:v00.00.00000 + name: sql + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/sql.yaml + title: Databricks Sql API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks sql service (workspace-level + APIs), generated from the Databricks Python SDK. + tags: + id: tags:v00.00.00000 + name: tags + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/tags.yaml + title: Databricks Tags API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks tags service (workspace-level + APIs), generated from the Databricks Python SDK. + vectorsearch: + id: vectorsearch:v00.00.00000 + name: vectorsearch + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/vectorsearch.yaml + title: Databricks Vectorsearch API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks vectorsearch service + (workspace-level APIs), generated from the Databricks Python SDK. + workspace: + id: workspace:v00.00.00000 + name: workspace + preferred: true + service: + $ref: databricks_workspace/v00.00.00000/services/workspace.yaml + title: Databricks Workspace API (workspace) + version: v00.00.00000 + description: >- + OpenAPI specification for the Databricks workspace service + (workspace-level APIs), generated from the Databricks Python SDK. +config: + auth: + type: oauth2 + client_id_env_var: DATABRICKS_CLIENT_ID + client_secret_env_var: DATABRICKS_CLIENT_SECRET + grant_type: client_credentials + token_url: >- + https://accounts.cloud.databricks.com/oidc/accounts/{{ + .__env__DATABRICKS_ACCOUNT_ID }}/v1/token + scopes: + - all-apis diff --git a/providers/src/databricks_workspace/v00.00.00000/services/agentbricks.yaml b/providers/src/databricks_workspace/v00.00.00000/services/agentbricks.yaml new file mode 100644 index 00000000..13a00d8a --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/agentbricks.yaml @@ -0,0 +1,446 @@ +openapi: 3.0.0 +info: + title: Databricks Agentbricks API (workspace) + description: >- + OpenAPI specification for the Databricks agentbricks service + (workspace-level APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.agentbricks +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/custom-llms/{id}/optimize/cancel: + post: + operationId: agent_bricks_cancel_optimize + summary: Cancel a Custom LLM Optimization Run. + tags: + - agentbricks + - agent_bricks + description: |- + Cancel a Custom LLM Optimization Run. + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AgentBricksAPI + /api/2.0/custom-llms: + post: + operationId: agent_bricks_create_custom_llm + summary: Create a Custom LLM. + tags: + - agentbricks + - agent_bricks + description: |- + Create a Custom LLM. + + :param name: str + Name of the custom LLM. Only alphanumeric characters and dashes allowed. + :param instructions: str + Instructions for the custom LLM to follow + :param agent_artifact_path: str (optional) + This will soon be deprecated!! Optional: UC path for agent artifacts. If you are using a dataset + that you only have read permissions, please provide a destination path where you have write + permissions. Please provide this in catalog.schema format. + :param datasets: List[:class:`Dataset`] (optional) + Datasets used for training and evaluating the model, not for inference. Currently, only 1 dataset is + accepted. + :param guidelines: List[str] (optional) + Guidelines for the custom LLM to adhere to + + :returns: :class:`CustomLlm` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: >- + Name of the custom LLM. Only alphanumeric characters and + dashes allowed. + instructions: + type: string + description: Instructions for the custom LLM to follow + agent_artifact_path: + type: string + description: >- + This will soon be deprecated!! Optional: UC path for agent + artifacts. If you are using a dataset that you only have + read permissions, please provide a destination path where + you have write permissions. Please provide this in + catalog.schema format. + datasets: + type: string + description: >- + Datasets used for training and evaluating the model, not for + inference. Currently, only 1 dataset is accepted. + guidelines: + type: string + description: Guidelines for the custom LLM to adhere to + required: + - name + - instructions + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomLlm' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AgentBricksAPI + /api/2.0/custom-llms/{id}: + delete: + operationId: agent_bricks_delete_custom_llm + summary: Delete a Custom LLM. + tags: + - agentbricks + - agent_bricks + description: |- + Delete a Custom LLM. + + :param id: str + The id of the custom llm + parameters: + - name: id + in: path + required: true + schema: + type: string + description: The id of the custom llm + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AgentBricksAPI + get: + operationId: agent_bricks_get_custom_llm + summary: Get a Custom LLM. + tags: + - agentbricks + - agent_bricks + description: |- + Get a Custom LLM. + + :param id: str + The id of the custom llm + + :returns: :class:`CustomLlm` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: The id of the custom llm + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomLlm' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AgentBricksAPI + patch: + operationId: agent_bricks_update_custom_llm + summary: Update a Custom LLM. + tags: + - agentbricks + - agent_bricks + description: |- + Update a Custom LLM. + + :param id: str + The id of the custom llm + :param custom_llm: :class:`CustomLlm` + The CustomLlm containing the fields which should be updated. + :param update_mask: str + The list of the CustomLlm fields to update. These should correspond to the values (or lack thereof) + present in `custom_llm`. + + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`CustomLlm` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: The id of the custom llm + requestBody: + content: + application/json: + schema: + type: object + properties: + custom_llm: + type: string + description: The CustomLlm containing the fields which should be updated. + update_mask: + type: string + description: >- + The list of the CustomLlm fields to update. These should + correspond to the values (or lack thereof) present in + `custom_llm`. The field mask must be a single string, with + multiple fields separated by commas (no spaces). The field + path is relative to the resource object, using a dot (`.`) + to navigate sub-fields (e.g., `author.given_name`). + Specification of elements in sequence or map fields is not + allowed, as only the entire collection field can be + specified. Field names must exactly match the resource field + names. A field mask of `*` indicates full replacement. It’s + recommended to always explicitly list the fields being + updated and avoid using `*` wildcards, as it can lead to + unintended results if the API changes in the future. + required: + - custom_llm + - update_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomLlm' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AgentBricksAPI + /api/2.0/custom-llms/{id}/optimize: + post: + operationId: agent_bricks_start_optimize + summary: Start a Custom LLM Optimization Run. + tags: + - agentbricks + - agent_bricks + description: |- + Start a Custom LLM Optimization Run. + + :param id: str + The Id of the tile. + + :returns: :class:`CustomLlm` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: The Id of the tile. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomLlm' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AgentBricksAPI +components: + schemas: + CustomLlm: + type: object + properties: + name: + type: string + instructions: + type: string + description: Instructions for the custom LLM to follow + agent_artifact_path: + type: string + creation_time: + type: string + description: Creation timestamp of the custom LLM + creator: + type: string + description: Creator of the custom LLM + datasets: + type: array + items: + $ref: '#/components/schemas/Dataset' + description: >- + Datasets used for training and evaluating the model, not for + inference + endpoint_name: + type: string + description: Name of the endpoint that will be used to serve the custom LLM + guidelines: + type: array + items: + type: string + description: Guidelines for the custom LLM to adhere to + id: + type: string + optimization_state: + $ref: '#/components/schemas/State' + description: If optimization is kicked off, tracks the state of the custom LLM + required: + - name + - instructions + Dataset: + type: object + properties: + table: + $ref: '#/components/schemas/Table' + required: + - table + Table: + type: object + properties: + table_path: + type: string + request_col: + type: string + description: Name of the request column + response_col: + type: string + description: 'Optional: Name of the response column if the data is labeled' + required: + - table_path + - request_col + State: + type: string + enum: + - CANCELLED + - COMPLETED + - CREATED + - FAILED + - PENDING + - RUNNING + description: States of Custom LLM optimization lifecycle. + x-stackQL-resources: + custom_llms: + id: databricks_workspace.agentbricks.custom_llms + name: custom_llms + title: Custom Llms + methods: + cancel_optimize: + operation: + $ref: '#/paths/~1api~12.0~1custom-llms~1{id}~1optimize~1cancel/post' + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~12.0~1custom-llms/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1custom-llms~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1custom-llms~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1custom-llms~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + start_optimize: + operation: + $ref: '#/paths/~1api~12.0~1custom-llms~1{id}~1optimize/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/custom_llms/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/custom_llms/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/custom_llms/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/custom_llms/methods/delete' + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/apps.yaml b/providers/src/databricks_workspace/v00.00.00000/services/apps.yaml index 1277adac..849d1a51 100644 --- a/providers/src/databricks_workspace/v00.00.00000/services/apps.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/services/apps.yaml @@ -1,1854 +1,2777 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Apps API - description: Apps -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/apps: - post: - operationId: apps-create - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/create - x-stackQL-resource: apps - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - description: - type: string - compute_status: - type: object - properties: - state: - type: string - message: - type: string - app_status: - type: object - properties: - state: - type: string - message: - type: string - url: - type: string - active_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - updater: - type: string - pending_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - resources: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - secret: - type: object - properties: - scope: - type: string - key: - type: string - permission: - type: string - sql_warehouse: - type: object - properties: - id: - type: string - permission: - type: string - serving_endpoint: - type: object - properties: - name: - type: string - permission: - type: string - job: - type: object - properties: - id: - type: string - permission: - type: string - service_principal_id: - type: integer - service_principal_name: - type: string - default_source_code_path: - type: string - service_principal_client_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - description: - type: string - resources: - type: Array of object - example: - name: my-custom-app - description: My app description. - resources: - - name: api-key - description: API key for external service. - secret: - scope: my-scope - key: my-key - permission: READ - sql_warehouse: - id: e9ca293f79a74b5c - permission: CAN_MANAGE - serving_endpoint: - name: databricks-meta-llama-3-1-70b-instruct - permission: CAN_MANAGE - job: - id: '1234' - permission: CAN_MANAGE - description: Creates a new app. - get: - operationId: apps-list - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/list - x-stackQL-resource: apps - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - apps: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - compute_status: - type: object - properties: - state: - type: string - message: - type: string - app_status: - type: object - properties: - state: - type: string - message: - type: string - url: - type: string - active_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - updater: - type: string - pending_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - resources: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - secret: - type: object - properties: - scope: - type: string - key: - type: string - permission: - type: string - sql_warehouse: - type: object - properties: - id: - type: string - permission: - type: string - serving_endpoint: - type: object - properties: - name: - type: string - permission: - type: string - job: - type: object - properties: - id: - type: string - permission: - type: string - service_principal_id: - type: integer - service_principal_name: - type: string - default_source_code_path: - type: string - service_principal_client_id: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.apps - description: Lists all apps in the workspace. - /api/2.0/apps/{name}: - delete: - operationId: apps-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/delete - x-stackQL-resource: apps - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - description: - type: string - compute_status: - type: object - properties: - state: - type: string - message: - type: string - app_status: - type: object - properties: - state: - type: string - message: - type: string - url: - type: string - active_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - updater: - type: string - pending_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - resources: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - secret: - type: object - properties: - scope: - type: string - key: - type: string - permission: - type: string - sql_warehouse: - type: object - properties: - id: - type: string - permission: - type: string - serving_endpoint: - type: object - properties: - name: - type: string - permission: - type: string - job: - type: object - properties: - id: - type: string - permission: - type: string - service_principal_id: - type: integer - service_principal_name: - type: string - default_source_code_path: - type: string - service_principal_client_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes an app. - get: - operationId: apps-get - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/get - x-stackQL-resource: apps - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - description: - type: string - compute_status: - type: object - properties: - state: - type: string - message: - type: string - app_status: - type: object - properties: - state: - type: string - message: - type: string - url: - type: string - active_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - updater: - type: string - pending_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - resources: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - secret: - type: object - properties: - scope: - type: string - key: - type: string - permission: - type: string - sql_warehouse: - type: object - properties: - id: - type: string - permission: - type: string - serving_endpoint: - type: object - properties: - name: - type: string - permission: - type: string - job: - type: object - properties: - id: - type: string - permission: - type: string - service_principal_id: - type: integer - service_principal_name: - type: string - default_source_code_path: - type: string - service_principal_client_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves information for the app with the supplied name. - patch: - operationId: apps-update - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/update - x-stackQL-resource: apps - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - description: - type: string - compute_status: - type: object - properties: - state: - type: string - message: - type: string - app_status: - type: object - properties: - state: - type: string - message: - type: string - url: - type: string - active_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - updater: - type: string - pending_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - resources: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - secret: - type: object - properties: - scope: - type: string - key: - type: string - permission: - type: string - sql_warehouse: - type: object - properties: - id: - type: string - permission: - type: string - serving_endpoint: - type: object - properties: - name: - type: string - permission: - type: string - job: - type: object - properties: - id: - type: string - permission: - type: string - service_principal_id: - type: integer - service_principal_name: - type: string - default_source_code_path: - type: string - service_principal_client_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - description: - type: string - resources: - type: Array of object - example: - name: my-custom-app - description: My app description. - resources: - - name: api-key - description: API key for external service. - secret: - scope: my-scope - key: my-key - permission: READ - sql_warehouse: - id: e9ca293f79a74b5c - permission: CAN_MANAGE - serving_endpoint: - name: databricks-meta-llama-3-1-70b-instruct - permission: CAN_MANAGE - job: - id: '1234' - permission: CAN_MANAGE - description: Updates the app with the supplied name. - /api/2.0/apps/{app_name}/deployments: - post: - operationId: apps-deploy - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/deploy - x-stackQL-resource: apps - x-stackQL-method: deploy - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: app_name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - example: - deployment_id: 01ef0bda89f21f08a8351f41e4a9b948 - source_code_path: /Workspace/user@test.com/my_custom_app - mode: SNAPSHOT - description: Creates an app deployment for the app with the supplied name. - get: - operationId: deployments-listdeployments - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/listdeployments - x-stackQL-resource: deployments - x-stackQL-method: listdeployments - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: app_name - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - app_deployments: - type: array - items: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.app_deployments - description: Lists all app deployments for the app with the supplied name. - /api/2.0/apps/{name}/start: - post: - operationId: apps-start - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/start - x-stackQL-resource: apps - x-stackQL-method: start - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - description: - type: string - compute_status: - type: object - properties: - state: - type: string - message: - type: string - app_status: - type: object - properties: - state: - type: string - message: - type: string - url: - type: string - active_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - updater: - type: string - pending_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - resources: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - secret: - type: object - properties: - scope: - type: string - key: - type: string - permission: - type: string - sql_warehouse: - type: object - properties: - id: - type: string - permission: - type: string - serving_endpoint: - type: object - properties: - name: - type: string - permission: - type: string - job: - type: object - properties: - id: - type: string - permission: - type: string - service_principal_id: - type: integer - service_principal_name: - type: string - default_source_code_path: - type: string - service_principal_client_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Start the last active deployment of the app in the workspace. - /api/2.0/apps/{name}/stop: - post: - operationId: apps-stop - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/stop - x-stackQL-resource: apps - x-stackQL-method: stop - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - description: - type: string - compute_status: - type: object - properties: - state: - type: string - message: - type: string - app_status: - type: object - properties: - state: - type: string - message: - type: string - url: - type: string - active_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - updater: - type: string - pending_deployment: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - resources: - type: array - items: - type: object - properties: - name: - type: string - description: - type: string - secret: - type: object - properties: - scope: - type: string - key: - type: string - permission: - type: string - sql_warehouse: - type: object - properties: - id: - type: string - permission: - type: string - serving_endpoint: - type: object - properties: - name: - type: string - permission: - type: string - job: - type: object - properties: - id: - type: string - permission: - type: string - service_principal_id: - type: integer - service_principal_name: - type: string - default_source_code_path: - type: string - service_principal_client_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Stops the active deployment of the app in the workspace. - /api/2.0/apps/{app_name}/deployments/{deployment_id}: - get: - operationId: deployments-getdeployment - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/getdeployment - x-stackQL-resource: deployments - x-stackQL-method: getdeployment - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: app_name - required: true - description: string - in: path - - name: deployment_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - deployment_id: - type: string - source_code_path: - type: string - mode: - type: string - deployment_artifacts: - type: object - properties: - source_code_path: - type: string - status: - type: object - properties: - state: - type: string - message: - type: string - create_time: - type: string - creator: - type: string - update_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves information for the app deployment with the supplied - name and deployment id. - /api/2.0/permissions/apps/{app_name}: - get: - operationId: permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/getpermissions - x-stackQL-resource: permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: app_name - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of an app. Apps can inherit permissions from - their root object. - put: - operationId: permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/setpermissions - x-stackQL-resource: permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: app_name - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/updatepermissions - x-stackQL-resource: permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: app_name - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on an app. Apps can inherit permissions - from their root object. - /api/2.0/permissions/apps/{app_name}/permissionLevels: - get: - operationId: permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/apps/getpermissionlevels - x-stackQL-resource: permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: app_name - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. -components: - x-stackQL-resources: - apps: - id: databricks_workspace.apps.apps - name: apps - title: Apps - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1apps/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1apps/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.apps - delete: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deploy: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{app_name}~1deployments/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - start: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{name}~1start/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - stop: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{name}~1stop/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/apps/methods/get' - - $ref: '#/components/x-stackQL-resources/apps/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/apps/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/apps/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/apps/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/apps/methods/stop' - - $ref: '#/components/x-stackQL-resources/apps/methods/start' - - $ref: '#/components/x-stackQL-resources/apps/methods/deploy' - deployments: - id: databricks_workspace.apps.deployments - name: deployments - title: Deployments - methods: - listdeployments: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{app_name}~1deployments/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.app_deployments - getdeployment: - operation: - $ref: '#/paths/~1api~12.0~1apps~1{app_name}~1deployments~1{deployment_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/deployments/methods/getdeployment' - - $ref: '#/components/x-stackQL-resources/deployments/methods/listdeployments' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - permissions: - id: databricks_workspace.apps.permissions - name: permissions - title: Permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1apps~1{app_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1apps~1{app_name}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1apps~1{app_name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/permissions/methods/setpermissions' - delete: [] - exec: [] - permission_levels: - id: databricks_workspace.apps.permission_levels - name: permission_levels - title: Permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1apps~1{app_name}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] +openapi: 3.0.0 +info: + title: Databricks Apps API (workspace) + description: >- + OpenAPI specification for the Databricks apps service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.apps +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/apps: + post: + operationId: apps_create + summary: Creates a new app. + tags: + - apps + description: |- + Creates a new app. + + :param app: :class:`App` + :param no_compute: bool (optional) + If true, the app will not be started after creation. + + :returns: + Long-running operation waiter for :class:`App`. + See :method:wait_get_app_active for more details. + parameters: + - name: no_compute + in: query + required: false + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + app: + type: string + description: >- + :param no_compute: bool (optional) If true, the app will not + be started after creation. + required: + - app + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/App' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + get: + operationId: apps_list + summary: Lists all apps in the workspace. + tags: + - apps + description: |- + Lists all apps in the workspace. + + :param page_size: int (optional) + Upper bound for items returned. + :param page_token: str (optional) + Pagination token to go to the next page of apps. Requests first page if absent. + + :returns: Iterator over :class:`App` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of apps. Requests first page + if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAppsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/apps/{app_name}/update: + post: + operationId: apps_create_update + summary: >- + Creates an app update and starts the update process. The update process + is asynchronous and the status + tags: + - apps + description: >- + Creates an app update and starts the update process. The update process + is asynchronous and the status + + of the update can be checked with the GetAppUpdate method. + + + :param app_name: str + + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + :param app: :class:`App` (optional) + + + :returns: + Long-running operation waiter for :class:`AppUpdate`. + See :method:wait_get_update_app_succeeded for more details. + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: >- + :param update_mask: str The field mask must be a single string, with + multiple fields separated by commas (no spaces). The field path is + relative to the resource object, using a dot (`.`) to navigate + sub-fields (e.g., `author.given_name`). Specification of elements in + sequence or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the resource + field names. A field mask of `*` indicates full replacement. It’s + recommended to always explicitly list the fields being updated and + avoid using `*` wildcards, as it can lead to unintended results if + the API changes in the future. + requestBody: + content: + application/json: + schema: + type: object + properties: + update_mask: + type: string + app: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`AppUpdate`. See + :method:wait_get_update_app_succeeded for more details. + required: + - update_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AppUpdate' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + get: + operationId: apps_get_update + summary: Gets the status of an app update. + tags: + - apps + description: |- + Gets the status of an app update. + + :param app_name: str + The name of the app. + + :returns: :class:`AppUpdate` + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The name of the app. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AppUpdate' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/apps/{name}: + delete: + operationId: apps_delete + summary: Deletes an app. + tags: + - apps + description: |- + Deletes an app. + + :param name: str + The name of the app. + + :returns: :class:`App` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the app. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/App' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + get: + operationId: apps_get + summary: Retrieves information for the app with the supplied name. + tags: + - apps + description: |- + Retrieves information for the app with the supplied name. + + :param name: str + The name of the app. + + :returns: :class:`App` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the app. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/App' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + patch: + operationId: apps_update + summary: Updates the app with the supplied name. + tags: + - apps + description: |- + Updates the app with the supplied name. + + :param name: str + The name of the app. The name must contain only lowercase alphanumeric characters and hyphens. It + must be unique within the workspace. + :param app: :class:`App` + + :returns: :class:`App` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The name of the app. The name must contain only lowercase + alphanumeric characters and hyphens. It must be unique within the + workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + app: + type: string + description: ':returns: :class:`App`' + required: + - app + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/App' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/apps/{app_name}/deployments: + post: + operationId: apps_deploy + summary: Creates an app deployment for the app with the supplied name. + tags: + - apps + description: |- + Creates an app deployment for the app with the supplied name. + + :param app_name: str + The name of the app. + :param app_deployment: :class:`AppDeployment` + The app deployment configuration. + + :returns: + Long-running operation waiter for :class:`AppDeployment`. + See :method:wait_get_deployment_app_succeeded for more details. + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The name of the app. + requestBody: + content: + application/json: + schema: + type: object + properties: + app_deployment: + type: string + description: The app deployment configuration. + required: + - app_deployment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AppDeployment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + get: + operationId: apps_list_deployments + summary: Lists all app deployments for the app with the supplied name. + tags: + - apps + description: |- + Lists all app deployments for the app with the supplied name. + + :param app_name: str + The name of the app. + :param page_size: int (optional) + Upper bound for items returned. + :param page_token: str (optional) + Pagination token to go to the next page of apps. Requests first page if absent. + + :returns: Iterator over :class:`AppDeployment` + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The name of the app. + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of apps. Requests first page + if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAppDeploymentsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/apps/{app_name}/deployments/{deployment_id}: + get: + operationId: apps_get_deployment + summary: >- + Retrieves information for the app deployment with the supplied name and + deployment id. + tags: + - apps + description: >- + Retrieves information for the app deployment with the supplied name and + deployment id. + + + :param app_name: str + The name of the app. + :param deployment_id: str + The unique id of the deployment. + + :returns: :class:`AppDeployment` + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The name of the app. + - name: deployment_id + in: path + required: true + schema: + type: string + description: The unique id of the deployment. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AppDeployment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/permissions/apps/{app_name}/permissionLevels: + get: + operationId: apps_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - apps + description: |- + Gets the permission levels that a user can have on an object. + + :param app_name: str + The app for which to get or manage permissions. + + :returns: :class:`GetAppPermissionLevelsResponse` + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The app for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetAppPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/permissions/apps/{app_name}: + get: + operationId: apps_get_permissions + summary: >- + Gets the permissions of an app. Apps can inherit permissions from their + root object. + tags: + - apps + description: >- + Gets the permissions of an app. Apps can inherit permissions from their + root object. + + + :param app_name: str + The app for which to get or manage permissions. + + :returns: :class:`AppPermissions` + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The app for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AppPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + put: + operationId: apps_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - apps + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param app_name: str + The app for which to get or manage permissions. + :param access_control_list: List[:class:`AppAccessControlRequest`] + (optional) + + + :returns: :class:`AppPermissions` + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The app for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`AppPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AppPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + patch: + operationId: apps_update_permissions + summary: >- + Updates the permissions on an app. Apps can inherit permissions from + their root object. + tags: + - apps + description: >- + Updates the permissions on an app. Apps can inherit permissions from + their root object. + + + :param app_name: str + The app for which to get or manage permissions. + :param access_control_list: List[:class:`AppAccessControlRequest`] + (optional) + + + :returns: :class:`AppPermissions` + parameters: + - name: app_name + in: path + required: true + schema: + type: string + description: The app for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`AppPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AppPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/apps/{name}/start: + post: + operationId: apps_start + summary: Start the last active deployment of the app in the workspace. + tags: + - apps + description: |- + Start the last active deployment of the app in the workspace. + + :param name: str + The name of the app. + + :returns: + Long-running operation waiter for :class:`App`. + See :method:wait_get_app_active for more details. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the app. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/App' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/apps/{name}/stop: + post: + operationId: apps_stop + summary: Stops the active deployment of the app in the workspace. + tags: + - apps + description: |- + Stops the active deployment of the app in the workspace. + + :param name: str + The name of the app. + + :returns: + Long-running operation waiter for :class:`App`. + See :method:wait_get_app_stopped for more details. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the app. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/App' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsAPI + /api/2.0/apps-settings/templates: + post: + operationId: apps_settings_create_custom_template + summary: Creates a custom template. + tags: + - apps + - apps_settings + description: |- + Creates a custom template. + + :param template: :class:`CustomTemplate` + + :returns: :class:`CustomTemplate` + requestBody: + content: + application/json: + schema: + type: object + properties: + template: + type: string + description: ':returns: :class:`CustomTemplate`' + required: + - template + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomTemplate' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsSettingsAPI + get: + operationId: apps_settings_list_custom_templates + summary: Lists all custom templates in the workspace. + tags: + - apps + - apps_settings + description: |- + Lists all custom templates in the workspace. + + :param page_size: int (optional) + Upper bound for items returned. + :param page_token: str (optional) + Pagination token to go to the next page of custom templates. Requests first page if absent. + + :returns: Iterator over :class:`CustomTemplate` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of custom templates. + Requests first page if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCustomTemplatesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsSettingsAPI + /api/2.0/apps-settings/templates/{name}: + delete: + operationId: apps_settings_delete_custom_template + summary: Deletes the custom template with the specified name. + tags: + - apps + - apps_settings + description: |- + Deletes the custom template with the specified name. + + :param name: str + The name of the custom template. + + :returns: :class:`CustomTemplate` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the custom template. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomTemplate' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsSettingsAPI + get: + operationId: apps_settings_get_custom_template + summary: Gets the custom template with the specified name. + tags: + - apps + - apps_settings + description: |- + Gets the custom template with the specified name. + + :param name: str + The name of the custom template. + + :returns: :class:`CustomTemplate` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the custom template. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomTemplate' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsSettingsAPI + put: + operationId: apps_settings_update_custom_template + summary: >- + Updates the custom template with the specified name. Note that the + template name cannot be updated. + tags: + - apps + - apps_settings + description: >- + Updates the custom template with the specified name. Note that the + template name cannot be updated. + + + :param name: str + The name of the template. It must contain only alphanumeric characters, hyphens, underscores, and + whitespaces. It must be unique within the workspace. + :param template: :class:`CustomTemplate` + + + :returns: :class:`CustomTemplate` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The name of the template. It must contain only alphanumeric + characters, hyphens, underscores, and whitespaces. It must be unique + within the workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + template: + type: string + description: ':returns: :class:`CustomTemplate`' + required: + - template + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CustomTemplate' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AppsSettingsAPI +components: + schemas: + App: + type: object + properties: + name: + type: string + active_deployment: + $ref: '#/components/schemas/AppDeployment' + description: >- + The active deployment of the app. A deployment is considered active + when it has been deployed to the app compute. + app_status: + $ref: '#/components/schemas/ApplicationStatus' + budget_policy_id: + type: string + compute_size: + $ref: '#/components/schemas/ComputeSize' + compute_status: + $ref: '#/components/schemas/ComputeStatus' + create_time: + type: string + description: The creation time of the app. Formatted timestamp in ISO 6801. + creator: + type: string + description: The email of the user that created the app. + default_source_code_path: + type: string + description: >- + The default workspace file system path of the source code from which + app deployment are created. This field tracks the workspace source + code path of the last active deployment. + description: + type: string + description: The description of the app. + effective_budget_policy_id: + type: string + effective_usage_policy_id: + type: string + effective_user_api_scopes: + type: array + items: + type: string + description: The effective api scopes granted to the user access token. + git_repository: + $ref: '#/components/schemas/GitRepository' + description: >- + Git repository configuration for app deployments. When specified, + deployments can reference code from this repository by providing + only the git reference (branch, tag, or commit). + id: + type: string + description: The unique identifier of the app. + oauth2_app_client_id: + type: string + oauth2_app_integration_id: + type: string + pending_deployment: + $ref: '#/components/schemas/AppDeployment' + description: >- + The pending deployment of the app. A deployment is considered + pending when it is being prepared for deployment to the app compute. + resources: + type: array + items: + $ref: '#/components/schemas/AppResource' + description: Resources for the app. + service_principal_client_id: + type: string + service_principal_id: + type: integer + service_principal_name: + type: string + update_time: + type: string + description: The update time of the app. Formatted timestamp in ISO 6801. + updater: + type: string + description: The email of the user that last updated the app. + url: + type: string + description: The URL of the app once it is deployed. + usage_policy_id: + type: string + user_api_scopes: + type: array + items: + type: string + required: + - name + AppAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/AppPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + AppAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/AppPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + AppDeployment: + type: object + properties: + command: + type: array + items: + type: string + create_time: + type: string + description: >- + The creation time of the deployment. Formatted timestamp in ISO + 6801. + creator: + type: string + description: The email of the user creates the deployment. + deployment_artifacts: + $ref: '#/components/schemas/AppDeploymentArtifacts' + description: The deployment artifacts for an app. + deployment_id: + type: string + description: The unique id of the deployment. + env_vars: + type: array + items: + $ref: '#/components/schemas/EnvVar' + description: >- + The environment variables to set in the app runtime environment. + This will override the environment variables specified in the + app.yaml file. + git_source: + $ref: '#/components/schemas/GitSource' + description: Git repository to use as the source for the app deployment. + mode: + $ref: '#/components/schemas/AppDeploymentMode' + description: The mode of which the deployment will manage the source code. + source_code_path: + type: string + description: >- + The workspace file system path of the source code used to create the + app deployment. This is different from + `deployment_artifacts.source_code_path`, which is the path used by + the deployed app. The former refers to the original source code + location of the app in the workspace during deployment creation, + whereas the latter provides a system generated stable snapshotted + source code path used by the deployment. + status: + $ref: '#/components/schemas/AppDeploymentStatus' + description: Status and status message of the deployment + update_time: + type: string + description: The update time of the deployment. Formatted timestamp in ISO 6801. + AppDeploymentArtifacts: + type: object + properties: + source_code_path: + type: string + AppDeploymentStatus: + type: object + properties: + message: + type: string + state: + $ref: '#/components/schemas/AppDeploymentState' + description: State of the deployment. + AppManifest: + type: object + properties: + version: + type: integer + description: The manifest schema version, for now only 1 is allowed + name: + type: string + description: Name of the app defined by manifest author / publisher + description: + type: string + description: Description of the app defined by manifest author / publisher + resource_specs: + type: array + items: + $ref: '#/components/schemas/AppManifestAppResourceSpec' + required: + - version + - name + description: App manifest definition + AppManifestAppResourceExperimentSpec: + type: object + properties: + permission: + $ref: >- + #/components/schemas/AppManifestAppResourceExperimentSpecExperimentPermission + required: + - permission + AppManifestAppResourceJobSpec: + type: object + properties: + permission: + $ref: '#/components/schemas/AppManifestAppResourceJobSpecJobPermission' + required: + - permission + AppManifestAppResourceSecretSpec: + type: object + properties: + permission: + $ref: >- + #/components/schemas/AppManifestAppResourceSecretSpecSecretPermission + required: + - permission + AppManifestAppResourceServingEndpointSpec: + type: object + properties: + permission: + $ref: >- + #/components/schemas/AppManifestAppResourceServingEndpointSpecServingEndpointPermission + required: + - permission + AppManifestAppResourceSpec: + type: object + properties: + name: + type: string + description: Name of the App Resource. + description: + type: string + description: Description of the App Resource. + experiment_spec: + $ref: '#/components/schemas/AppManifestAppResourceExperimentSpec' + job_spec: + $ref: '#/components/schemas/AppManifestAppResourceJobSpec' + secret_spec: + $ref: '#/components/schemas/AppManifestAppResourceSecretSpec' + serving_endpoint_spec: + $ref: '#/components/schemas/AppManifestAppResourceServingEndpointSpec' + sql_warehouse_spec: + $ref: '#/components/schemas/AppManifestAppResourceSqlWarehouseSpec' + uc_securable_spec: + $ref: '#/components/schemas/AppManifestAppResourceUcSecurableSpec' + required: + - name + description: >- + AppResource related fields are copied from app.proto but excludes + resource identifiers (e.g. + name, id, key, scope, etc.) + AppManifestAppResourceSqlWarehouseSpec: + type: object + properties: + permission: + $ref: >- + #/components/schemas/AppManifestAppResourceSqlWarehouseSpecSqlWarehousePermission + required: + - permission + AppManifestAppResourceUcSecurableSpec: + type: object + properties: + securable_type: + $ref: >- + #/components/schemas/AppManifestAppResourceUcSecurableSpecUcSecurableType + permission: + $ref: >- + #/components/schemas/AppManifestAppResourceUcSecurableSpecUcSecurablePermission + required: + - securable_type + - permission + AppPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/AppPermissionLevel' + AppPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/AppAccessControlResponse' + object_id: + type: string + object_type: + type: string + AppPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/AppPermissionLevel' + AppResource: + type: object + properties: + name: + type: string + database: + $ref: '#/components/schemas/AppResourceDatabase' + description: + type: string + description: Description of the App Resource. + experiment: + $ref: '#/components/schemas/AppResourceExperiment' + genie_space: + $ref: '#/components/schemas/AppResourceGenieSpace' + job: + $ref: '#/components/schemas/AppResourceJob' + secret: + $ref: '#/components/schemas/AppResourceSecret' + serving_endpoint: + $ref: '#/components/schemas/AppResourceServingEndpoint' + sql_warehouse: + $ref: '#/components/schemas/AppResourceSqlWarehouse' + uc_securable: + $ref: '#/components/schemas/AppResourceUcSecurable' + required: + - name + AppResourceDatabase: + type: object + properties: + instance_name: + type: string + database_name: + type: string + permission: + $ref: '#/components/schemas/AppResourceDatabaseDatabasePermission' + required: + - instance_name + - database_name + - permission + AppResourceExperiment: + type: object + properties: + experiment_id: + type: string + permission: + $ref: '#/components/schemas/AppResourceExperimentExperimentPermission' + required: + - experiment_id + - permission + AppResourceGenieSpace: + type: object + properties: + name: + type: string + space_id: + type: string + permission: + $ref: '#/components/schemas/AppResourceGenieSpaceGenieSpacePermission' + required: + - name + - space_id + - permission + AppResourceJob: + type: object + properties: + id: + type: string + permission: + $ref: '#/components/schemas/AppResourceJobJobPermission' + description: >- + Permissions to grant on the Job. Supported permissions are: + "CAN_MANAGE", "IS_OWNER", "CAN_MANAGE_RUN", "CAN_VIEW". + required: + - id + - permission + AppResourceSecret: + type: object + properties: + scope: + type: string + key: + type: string + description: Key of the secret to grant permission on. + permission: + $ref: '#/components/schemas/AppResourceSecretSecretPermission' + description: >- + Permission to grant on the secret scope. For secrets, only one + permission is allowed. Permission must be one of: "READ", "WRITE", + "MANAGE". + required: + - scope + - key + - permission + AppResourceServingEndpoint: + type: object + properties: + name: + type: string + permission: + $ref: >- + #/components/schemas/AppResourceServingEndpointServingEndpointPermission + description: >- + Permission to grant on the serving endpoint. Supported permissions + are: "CAN_MANAGE", "CAN_QUERY", "CAN_VIEW". + required: + - name + - permission + AppResourceSqlWarehouse: + type: object + properties: + id: + type: string + permission: + $ref: '#/components/schemas/AppResourceSqlWarehouseSqlWarehousePermission' + description: >- + Permission to grant on the SQL warehouse. Supported permissions are: + "CAN_MANAGE", "CAN_USE", "IS_OWNER". + required: + - id + - permission + AppResourceUcSecurable: + type: object + properties: + securable_full_name: + type: string + securable_type: + $ref: '#/components/schemas/AppResourceUcSecurableUcSecurableType' + permission: + $ref: '#/components/schemas/AppResourceUcSecurableUcSecurablePermission' + required: + - securable_full_name + - securable_type + - permission + AppUpdate: + type: object + properties: + budget_policy_id: + type: string + compute_size: + $ref: '#/components/schemas/ComputeSize' + description: + type: string + git_repository: + $ref: '#/components/schemas/GitRepository' + resources: + type: array + items: + $ref: '#/components/schemas/AppResource' + status: + $ref: '#/components/schemas/AppUpdateUpdateStatus' + usage_policy_id: + type: string + user_api_scopes: + type: array + items: + type: string + AppUpdateUpdateStatus: + type: object + properties: + message: + type: string + state: + $ref: '#/components/schemas/AppUpdateUpdateStatusUpdateState' + ApplicationStatus: + type: object + properties: + message: + type: string + state: + $ref: '#/components/schemas/ApplicationState' + description: State of the application. + ComputeStatus: + type: object + properties: + active_instances: + type: integer + message: + type: string + description: Compute status message + state: + $ref: '#/components/schemas/ComputeState' + description: State of the app compute. + CustomTemplate: + type: object + properties: + name: + type: string + git_repo: + type: string + description: The Git repository URL that the template resides in. + path: + type: string + description: The path to the template within the Git repository. + manifest: + $ref: '#/components/schemas/AppManifest' + description: >- + The manifest of the template. It defines fields and default values + when installing the template. + git_provider: + type: string + description: The Git provider of the template. + creator: + type: string + description: + type: string + description: The description of the template. + required: + - name + - git_repo + - path + - manifest + - git_provider + EnvVar: + type: object + properties: + name: + type: string + value: + type: string + description: The value for the environment variable. + value_from: + type: string + description: >- + The name of an external Databricks resource that contains the value, + such as a secret or a database table. + GetAppPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/AppPermissionsDescription' + GitRepository: + type: object + properties: + url: + type: string + description: URL of the Git repository. + provider: + type: string + description: >- + Git provider. Case insensitive. Supported values: gitHub, + gitHubEnterprise, bitbucketCloud, bitbucketServer, + azureDevOpsServices, gitLab, gitLabEnterpriseEdition, awsCodeCommit. + required: + - url + - provider + description: Git repository configuration specifying the location of the repository. + GitSource: + type: object + properties: + branch: + type: string + description: Git branch to checkout. + commit: + type: string + description: Git commit SHA to checkout. + git_repository: + $ref: '#/components/schemas/GitRepository' + description: >- + Git repository configuration. Populated from the app's + git_repository configuration. + resolved_commit: + type: string + description: >- + The resolved commit SHA that was actually used for the deployment. + This is populated by the system after resolving the reference + (branch, tag, or commit). If commit is specified directly, this will + match commit. If a branch or tag is specified, this contains the + commit SHA that the branch or tag pointed to at deployment time. + source_code_path: + type: string + description: >- + Relative path to the app source code within the Git repository. If + not specified, the root of the repository is used. + tag: + type: string + description: Git tag to checkout. + description: >- + Complete git source specification including repository location and + reference. + ListAppDeploymentsResponse: + type: object + properties: + app_deployments: + type: array + items: + $ref: '#/components/schemas/AppDeployment' + next_page_token: + type: string + description: Pagination token to request the next page of apps. + ListAppsResponse: + type: object + properties: + apps: + type: array + items: + $ref: '#/components/schemas/App' + next_page_token: + type: string + description: Pagination token to request the next page of apps. + ListCustomTemplatesResponse: + type: object + properties: + next_page_token: + type: string + templates: + type: array + items: + $ref: '#/components/schemas/CustomTemplate' + AppDeploymentMode: + type: string + enum: + - AUTO_SYNC + - SNAPSHOT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppDeploymentState: + type: string + enum: + - CANCELLED + - FAILED + - IN_PROGRESS + - SUCCEEDED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppManifestAppResourceExperimentSpecExperimentPermission: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_READ + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppManifestAppResourceJobSpecJobPermission: + type: string + enum: + - CAN_MANAGE + - CAN_MANAGE_RUN + - CAN_VIEW + - IS_OWNER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppManifestAppResourceSecretSpecSecretPermission: + type: string + enum: + - MANAGE + - READ + - WRITE + description: >- + Permission to grant on the secret scope. Supported permissions are: + "READ", "WRITE", "MANAGE". + AppManifestAppResourceServingEndpointSpecServingEndpointPermission: + type: string + enum: + - CAN_MANAGE + - CAN_QUERY + - CAN_VIEW + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppManifestAppResourceSqlWarehouseSpecSqlWarehousePermission: + type: string + enum: + - CAN_MANAGE + - CAN_USE + - IS_OWNER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppManifestAppResourceUcSecurableSpecUcSecurablePermission: + type: string + enum: + - EXECUTE + - MANAGE + - READ_VOLUME + - SELECT + - USE_CONNECTION + - WRITE_VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppManifestAppResourceUcSecurableSpecUcSecurableType: + type: string + enum: + - CONNECTION + - FUNCTION + - TABLE + - VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppPermissionLevel: + type: string + enum: + - CAN_MANAGE + - CAN_USE + description: Permission level + AppResourceDatabaseDatabasePermission: + type: string + enum: + - CAN_CONNECT_AND_CREATE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppResourceExperimentExperimentPermission: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_READ + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppResourceGenieSpaceGenieSpacePermission: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_RUN + - CAN_VIEW + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppResourceJobJobPermission: + type: string + enum: + - CAN_MANAGE + - CAN_MANAGE_RUN + - CAN_VIEW + - IS_OWNER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppResourceSecretSecretPermission: + type: string + enum: + - MANAGE + - READ + - WRITE + description: >- + Permission to grant on the secret scope. Supported permissions are: + "READ", "WRITE", "MANAGE". + AppResourceServingEndpointServingEndpointPermission: + type: string + enum: + - CAN_MANAGE + - CAN_QUERY + - CAN_VIEW + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppResourceSqlWarehouseSqlWarehousePermission: + type: string + enum: + - CAN_MANAGE + - CAN_USE + - IS_OWNER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppResourceUcSecurableUcSecurablePermission: + type: string + enum: + - EXECUTE + - READ_VOLUME + - SELECT + - USE_CONNECTION + - WRITE_VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppResourceUcSecurableUcSecurableType: + type: string + enum: + - CONNECTION + - FUNCTION + - TABLE + - VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AppUpdateUpdateStatusUpdateState: + type: string + enum: + - FAILED + - IN_PROGRESS + - NOT_UPDATED + - SUCCEEDED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ApplicationState: + type: string + enum: + - CRASHED + - DEPLOYING + - RUNNING + - UNAVAILABLE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ComputeSize: + type: string + enum: + - LARGE + - MEDIUM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ComputeState: + type: string + enum: + - ACTIVE + - DELETING + - ERROR + - STARTING + - STOPPED + - STOPPING + - UPDATING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + apps: + id: databricks_workspace.apps.apps + name: apps + title: Apps + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1apps/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1apps/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.apps + delete: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + start: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{name}~1start/post' + response: + mediaType: application/json + openAPIDocKey: '200' + stop: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{name}~1stop/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/apps/methods/get' + - $ref: '#/components/x-stackQL-resources/apps/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/apps/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/apps/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/apps/methods/delete' + replace: [] + app_updates: + id: databricks_workspace.apps.app_updates + name: app_updates + title: App Updates + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{app_name}~1update/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{app_name}~1update/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/app_updates/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/app_updates/methods/create' + update: [] + delete: [] + replace: [] + app_deployments: + id: databricks_workspace.apps.app_deployments + name: app_deployments + title: App Deployments + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{app_name}~1deployments/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1apps~1{app_name}~1deployments/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.app_deployments + get: + operation: + $ref: >- + #/paths/~1api~12.0~1apps~1{app_name}~1deployments~1{deployment_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/app_deployments/methods/get' + - $ref: '#/components/x-stackQL-resources/app_deployments/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/app_deployments/methods/create' + update: [] + delete: [] + replace: [] + app_permission_levels: + id: databricks_workspace.apps.app_permission_levels + name: app_permission_levels + title: App Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1apps~1{app_name}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/app_permission_levels/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + app_permissions: + id: databricks_workspace.apps.app_permissions + name: app_permissions + title: App Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1apps~1{app_name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1apps~1{app_name}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1apps~1{app_name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/app_permissions/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/app_permissions/methods/update' + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/app_permissions/methods/set' + apps_settings: + id: databricks_workspace.apps.apps_settings + name: apps_settings + title: Apps Settings + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1apps-settings~1templates/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1apps-settings~1templates/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.templates + delete: + operation: + $ref: '#/paths/~1api~12.0~1apps-settings~1templates~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1apps-settings~1templates~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1apps-settings~1templates~1{name}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/apps_settings/methods/get' + - $ref: '#/components/x-stackQL-resources/apps_settings/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/apps_settings/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/apps_settings/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/apps_settings/methods/replace' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/catalog.yaml b/providers/src/databricks_workspace/v00.00.00000/services/catalog.yaml new file mode 100644 index 00000000..7e17980c --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/catalog.yaml @@ -0,0 +1,15911 @@ +openapi: 3.0.0 +info: + title: Databricks Catalog API (workspace) + description: >- + OpenAPI specification for the Databricks catalog service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.catalog +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.1/unity-catalog/artifact-allowlists/{artifact_type.value}: + get: + operationId: artifact_allowlists_get + summary: >- + Get the artifact allowlist of a certain artifact type. The caller must + be a metastore admin or have + tags: + - catalog + - artifact_allowlists + description: >- + Get the artifact allowlist of a certain artifact type. The caller must + be a metastore admin or have + + the **MANAGE ALLOWLIST** privilege on the metastore. + + + :param artifact_type: :class:`ArtifactType` + The artifact type of the allowlist. + + :returns: :class:`ArtifactAllowlistInfo` + parameters: + - name: artifact_type.value + in: path + required: true + schema: + type: string + - name: artifact_type + in: query + required: true + schema: + type: string + description: The artifact type of the allowlist. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ArtifactAllowlistInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ArtifactAllowlistsAPI + put: + operationId: artifact_allowlists_update + summary: >- + Set the artifact allowlist of a certain artifact type. The whole + artifact allowlist is replaced with + tags: + - catalog + - artifact_allowlists + description: >- + Set the artifact allowlist of a certain artifact type. The whole + artifact allowlist is replaced with + + the new allowlist. The caller must be a metastore admin or have the + **MANAGE ALLOWLIST** privilege on + + the metastore. + + + :param artifact_type: :class:`ArtifactType` + The artifact type of the allowlist. + :param artifact_matchers: List[:class:`ArtifactMatcher`] + A list of allowed artifact match patterns. + :param created_at: int (optional) + Time at which this artifact allowlist was set, in epoch milliseconds. + :param created_by: str (optional) + Username of the user who set the artifact allowlist. + :param metastore_id: str (optional) + Unique identifier of parent metastore. + + :returns: :class:`ArtifactAllowlistInfo` + parameters: + - name: artifact_type.value + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + artifact_type: + type: string + description: The artifact type of the allowlist. + artifact_matchers: + type: string + description: A list of allowed artifact match patterns. + created_at: + type: string + description: >- + Time at which this artifact allowlist was set, in epoch + milliseconds. + created_by: + type: string + description: Username of the user who set the artifact allowlist. + metastore_id: + type: string + description: Unique identifier of parent metastore. + required: + - artifact_type + - artifact_matchers + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ArtifactAllowlistInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ArtifactAllowlistsAPI + /api/2.1/unity-catalog/catalogs: + post: + operationId: catalogs_create + summary: >- + Creates a new catalog instance in the parent metastore if the caller is + a metastore admin or has the + tags: + - catalog + - catalogs + description: >- + Creates a new catalog instance in the parent metastore if the caller is + a metastore admin or has the + + **CREATE_CATALOG** privilege. + + + :param name: str + Name of catalog. + :param comment: str (optional) + User-provided free-form text description. + :param connection_name: str (optional) + The name of the connection to an external data source. + :param options: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + :param properties: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + :param provider_name: str (optional) + The name of delta sharing provider. + + A Delta Sharing catalog is a catalog that is based on a Delta share on a remote sharing server. + :param share_name: str (optional) + The name of the share under the share provider. + :param storage_root: str (optional) + Storage root URL for managed tables within catalog. + + :returns: :class:`CatalogInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of catalog. + comment: + type: string + description: User-provided free-form text description. + connection_name: + type: string + description: The name of the connection to an external data source. + options: + type: string + description: A map of key-value properties attached to the securable. + properties: + type: string + description: A map of key-value properties attached to the securable. + provider_name: + type: string + description: >- + The name of delta sharing provider. A Delta Sharing catalog + is a catalog that is based on a Delta share on a remote + sharing server. + share_name: + type: string + description: The name of the share under the share provider. + storage_root: + type: string + description: Storage root URL for managed tables within catalog. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CatalogInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CatalogsAPI + get: + operationId: catalogs_list + summary: >- + Gets an array of catalogs in the metastore. If the caller is the + metastore admin, all catalogs will be + tags: + - catalog + - catalogs + description: >- + Gets an array of catalogs in the metastore. If the caller is the + metastore admin, all catalogs will be + + retrieved. Otherwise, only catalogs owned by the caller (or for which + the caller has the + + **USE_CATALOG** privilege) will be retrieved. There is no guarantee of a + specific ordering of the + + elements in the array. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param include_browse: bool (optional) + Whether to include catalogs in the response for which the principal can only access selective + metadata for + :param include_unbound: bool (optional) + Whether to include catalogs not bound to the workspace. Effective only if the user has permission to + update the catalog–workspace binding. + :param max_results: int (optional) + Maximum number of catalogs to return. - when set to 0, the page length is set to a server configured + value (recommended); - when set to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than 0, an invalid parameter error + is returned; - If not set, all valid catalogs are returned (not recommended). - Note: The number of + returned catalogs might be less than the specified max_results size, even zero. The only definitive + indication that no further catalogs can be fetched is when the next_page_token is unset from the + response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`CatalogInfo` + parameters: + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include catalogs in the response for which the principal + can only access selective metadata for + - name: include_unbound + in: query + required: false + schema: + type: string + description: >- + Whether to include catalogs not bound to the workspace. Effective + only if the user has permission to update the catalog–workspace + binding. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of catalogs to return. - when set to 0, the page + length is set to a server configured value (recommended); - when set + to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than + 0, an invalid parameter error is returned; - If not set, all valid + catalogs are returned (not recommended). - Note: The number of + returned catalogs might be less than the specified max_results size, + even zero. The only definitive indication that no further catalogs + can be fetched is when the next_page_token is unset from the + response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCatalogsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CatalogsAPI + /api/2.1/unity-catalog/catalogs/{name}: + delete: + operationId: catalogs_delete + summary: >- + Deletes the catalog that matches the supplied name. The caller must be a + metastore admin or the owner + tags: + - catalog + - catalogs + description: >- + Deletes the catalog that matches the supplied name. The caller must be a + metastore admin or the owner + + of the catalog. + + + :param name: str + The name of the catalog. + :param force: bool (optional) + Force deletion even if the catalog is not empty. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the catalog. + - name: force + in: query + required: false + schema: + type: string + description: Force deletion even if the catalog is not empty. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CatalogsAPI + get: + operationId: catalogs_get + summary: >- + Gets the specified catalog in a metastore. The caller must be a + metastore admin, the owner of the + tags: + - catalog + - catalogs + description: >- + Gets the specified catalog in a metastore. The caller must be a + metastore admin, the owner of the + + catalog, or a user that has the **USE_CATALOG** privilege set for their + account. + + + :param name: str + The name of the catalog. + :param include_browse: bool (optional) + Whether to include catalogs in the response for which the principal can only access selective + metadata for + + :returns: :class:`CatalogInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the catalog. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include catalogs in the response for which the principal + can only access selective metadata for + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CatalogInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CatalogsAPI + patch: + operationId: catalogs_update + summary: >- + Updates the catalog that matches the supplied name. The caller must be + either the owner of the + tags: + - catalog + - catalogs + description: >- + Updates the catalog that matches the supplied name. The caller must be + either the owner of the + + catalog, or a metastore admin (when changing the owner field of the + catalog). + + + :param name: str + The name of the catalog. + :param comment: str (optional) + User-provided free-form text description. + :param enable_predictive_optimization: + :class:`EnablePredictiveOptimization` (optional) + Whether predictive optimization should be enabled for this object and objects under it. + :param isolation_mode: :class:`CatalogIsolationMode` (optional) + Whether the current securable is accessible from all workspaces or a specific set of workspaces. + :param new_name: str (optional) + New name for the catalog. + :param options: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + :param owner: str (optional) + Username of current owner of catalog. + :param properties: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + + :returns: :class:`CatalogInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the catalog. + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: User-provided free-form text description. + enable_predictive_optimization: + type: string + description: >- + Whether predictive optimization should be enabled for this + object and objects under it. + isolation_mode: + type: string + description: >- + Whether the current securable is accessible from all + workspaces or a specific set of workspaces. + new_name: + type: string + description: New name for the catalog. + options: + type: string + description: A map of key-value properties attached to the securable. + owner: + type: string + description: Username of current owner of catalog. + properties: + type: string + description: A map of key-value properties attached to the securable. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CatalogInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CatalogsAPI + /api/2.1/unity-catalog/connections: + post: + operationId: connections_create + summary: Creates a new connection + tags: + - catalog + - connections + description: >- + Creates a new connection + + + Creates a new connection to an external data source. It allows users to + specify connection details and + + configurations for interaction with the external server. + + + :param name: str + Name of the connection. + :param connection_type: :class:`ConnectionType` + The type of connection. + :param options: Dict[str,str] + A map of key-value properties attached to the securable. + :param comment: str (optional) + User-provided free-form text description. + :param properties: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + :param read_only: bool (optional) + If the connection is read only. + + :returns: :class:`ConnectionInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the connection. + connection_type: + type: string + description: The type of connection. + options: + type: string + description: A map of key-value properties attached to the securable. + comment: + type: string + description: User-provided free-form text description. + properties: + type: string + description: A map of key-value properties attached to the securable. + read_only: + type: string + description: If the connection is read only. + required: + - name + - connection_type + - options + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ConnectionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConnectionsAPI + get: + operationId: connections_list + summary: List all connections. + tags: + - catalog + - connections + description: >- + List all connections. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param max_results: int (optional) + Maximum number of connections to return. - If not set, all connections are returned (not + recommended). - when set to a value greater than 0, the page length is the minimum of this value and + a server configured value; - when set to 0, the page length is set to a server configured value + (recommended); - when set to a value less than 0, an invalid parameter error is returned; + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ConnectionInfo` + parameters: + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of connections to return. - If not set, all + connections are returned (not recommended). - when set to a value + greater than 0, the page length is the minimum of this value and a + server configured value; - when set to 0, the page length is set to + a server configured value (recommended); - when set to a value less + than 0, an invalid parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListConnectionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConnectionsAPI + /api/2.1/unity-catalog/connections/{name}: + delete: + operationId: connections_delete + summary: Deletes the connection that matches the supplied name. + tags: + - catalog + - connections + description: |- + Deletes the connection that matches the supplied name. + + :param name: str + The name of the connection to be deleted. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the connection to be deleted. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConnectionsAPI + get: + operationId: connections_get + summary: Gets a connection from it's name. + tags: + - catalog + - connections + description: |- + Gets a connection from it's name. + + :param name: str + Name of the connection. + + :returns: :class:`ConnectionInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the connection. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ConnectionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConnectionsAPI + patch: + operationId: connections_update + summary: Updates the connection that matches the supplied name. + tags: + - catalog + - connections + description: |- + Updates the connection that matches the supplied name. + + :param name: str + Name of the connection. + :param options: Dict[str,str] + A map of key-value properties attached to the securable. + :param new_name: str (optional) + New name for the connection. + :param owner: str (optional) + Username of current owner of the connection. + + :returns: :class:`ConnectionInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the connection. + requestBody: + content: + application/json: + schema: + type: object + properties: + options: + type: string + description: A map of key-value properties attached to the securable. + new_name: + type: string + description: New name for the connection. + owner: + type: string + description: Username of current owner of the connection. + required: + - options + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ConnectionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConnectionsAPI + /api/2.1/unity-catalog/entity-tag-assignments: + post: + operationId: entity_tag_assignments_create + summary: Creates a tag assignment for an Unity Catalog entity. + tags: + - catalog + - entity_tag_assignments + description: >- + Creates a tag assignment for an Unity Catalog entity. + + + To add tags to Unity Catalog entities, you must own the entity or have + the following privileges: - + + **APPLY TAG** on the entity - **USE SCHEMA** on the entity's parent + schema - **USE CATALOG** on the + + entity's parent catalog + + + To add a governed tag to Unity Catalog entities, you must also have the + **ASSIGN** or **MANAGE** + + permission on the tag policy. See [Manage tag policy permissions]. + + + [Manage tag policy permissions]: + https://docs.databricks.com/aws/en/admin/tag-policies/manage-permissions + + + :param tag_assignment: :class:`EntityTagAssignment` + + + :returns: :class:`EntityTagAssignment` + requestBody: + content: + application/json: + schema: + type: object + properties: + tag_assignment: + type: string + description: ':returns: :class:`EntityTagAssignment`' + required: + - tag_assignment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EntityTagAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EntityTagAssignmentsAPI + /api/2.1/unity-catalog/entity-tag-assignments/{entity_type}/{entity_name}/tags/{tag_key}: + delete: + operationId: entity_tag_assignments_delete + summary: Deletes a tag assignment for an Unity Catalog entity by its key. + tags: + - catalog + - entity_tag_assignments + description: >- + Deletes a tag assignment for an Unity Catalog entity by its key. + + + To delete tags from Unity Catalog entities, you must own the entity or + have the following privileges: + + - **APPLY TAG** on the entity - **USE_SCHEMA** on the entity's parent + schema - **USE_CATALOG** on the + + entity's parent catalog + + + To delete a governed tag from Unity Catalog entities, you must also have + the **ASSIGN** or **MANAGE** + + permission on the tag policy. See [Manage tag policy permissions]. + + + [Manage tag policy permissions]: + https://docs.databricks.com/aws/en/admin/tag-policies/manage-permissions + + + :param entity_type: str + The type of the entity to which the tag is assigned. Allowed values are: catalogs, schemas, tables, + columns, volumes. + :param entity_name: str + The fully qualified name of the entity to which the tag is assigned + :param tag_key: str + Required. The key of the tag to delete + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of the entity to which the tag is assigned. Allowed values + are: catalogs, schemas, tables, columns, volumes. + - name: entity_name + in: path + required: true + schema: + type: string + description: The fully qualified name of the entity to which the tag is assigned + - name: tag_key + in: path + required: true + schema: + type: string + description: Required. The key of the tag to delete + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EntityTagAssignmentsAPI + get: + operationId: entity_tag_assignments_get + summary: Gets a tag assignment for an Unity Catalog entity by tag key. + tags: + - catalog + - entity_tag_assignments + description: |- + Gets a tag assignment for an Unity Catalog entity by tag key. + + :param entity_type: str + The type of the entity to which the tag is assigned. Allowed values are: catalogs, schemas, tables, + columns, volumes. + :param entity_name: str + The fully qualified name of the entity to which the tag is assigned + :param tag_key: str + Required. The key of the tag + + :returns: :class:`EntityTagAssignment` + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of the entity to which the tag is assigned. Allowed values + are: catalogs, schemas, tables, columns, volumes. + - name: entity_name + in: path + required: true + schema: + type: string + description: The fully qualified name of the entity to which the tag is assigned + - name: tag_key + in: path + required: true + schema: + type: string + description: Required. The key of the tag + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EntityTagAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EntityTagAssignmentsAPI + patch: + operationId: entity_tag_assignments_update + summary: Updates an existing tag assignment for an Unity Catalog entity. + tags: + - catalog + - entity_tag_assignments + description: >- + Updates an existing tag assignment for an Unity Catalog entity. + + + To update tags to Unity Catalog entities, you must own the entity or + have the following privileges: - + + **APPLY TAG** on the entity - **USE SCHEMA** on the entity's parent + schema - **USE CATALOG** on the + + entity's parent catalog + + + To update a governed tag to Unity Catalog entities, you must also have + the **ASSIGN** or **MANAGE** + + permission on the tag policy. See [Manage tag policy permissions]. + + + [Manage tag policy permissions]: + https://docs.databricks.com/aws/en/admin/tag-policies/manage-permissions + + + :param entity_type: str + The type of the entity to which the tag is assigned. Allowed values are: catalogs, schemas, tables, + columns, volumes. + :param entity_name: str + The fully qualified name of the entity to which the tag is assigned + :param tag_key: str + The key of the tag + :param tag_assignment: :class:`EntityTagAssignment` + + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`EntityTagAssignment` + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of the entity to which the tag is assigned. Allowed values + are: catalogs, schemas, tables, columns, volumes. + - name: entity_name + in: path + required: true + schema: + type: string + description: The fully qualified name of the entity to which the tag is assigned + - name: tag_key + in: path + required: true + schema: + type: string + description: The key of the tag + - name: update_mask + in: query + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + tag_assignment: + type: string + description: >- + :param update_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + required: + - tag_assignment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EntityTagAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EntityTagAssignmentsAPI + /api/2.1/unity-catalog/entity-tag-assignments/{entity_type}/{entity_name}/tags: + get: + operationId: entity_tag_assignments_list + summary: List tag assignments for an Unity Catalog entity + tags: + - catalog + - entity_tag_assignments + description: >- + List tag assignments for an Unity Catalog entity + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param entity_type: str + The type of the entity to which the tag is assigned. Allowed values are: catalogs, schemas, tables, + columns, volumes. + :param entity_name: str + The fully qualified name of the entity to which the tag is assigned + :param max_results: int (optional) + Optional. Maximum number of tag assignments to return in a single page + :param page_token: str (optional) + Optional. Pagination token to retrieve the next page of results + + :returns: Iterator over :class:`EntityTagAssignment` + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of the entity to which the tag is assigned. Allowed values + are: catalogs, schemas, tables, columns, volumes. + - name: entity_name + in: path + required: true + schema: + type: string + description: The fully qualified name of the entity to which the tag is assigned + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Optional. Maximum number of tag assignments to return in a single + page + - name: page_token + in: query + required: false + schema: + type: string + description: Optional. Pagination token to retrieve the next page of results + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListEntityTagAssignmentsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EntityTagAssignmentsAPI + /api/2.0/lineage-tracking/external-lineage: + post: + operationId: external_lineage_create_external_lineage_relationship + summary: >- + Creates an external lineage relationship between a Databricks or + external metadata object and another + tags: + - catalog + - external_lineage + description: >- + Creates an external lineage relationship between a Databricks or + external metadata object and another + + external metadata object. + + + :param external_lineage_relationship: + :class:`CreateRequestExternalLineage` + + + :returns: :class:`ExternalLineageRelationship` + requestBody: + content: + application/json: + schema: + type: object + properties: + external_lineage_relationship: + type: string + description: ':returns: :class:`ExternalLineageRelationship`' + required: + - external_lineage_relationship + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalLineageRelationship' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLineageAPI + delete: + operationId: external_lineage_delete_external_lineage_relationship + summary: >- + Deletes an external lineage relationship between a Databricks or + external metadata object and another + tags: + - catalog + - external_lineage + description: >- + Deletes an external lineage relationship between a Databricks or + external metadata object and another + + external metadata object. + + + :param external_lineage_relationship: + :class:`DeleteRequestExternalLineage` + parameters: + - name: external_lineage_relationship + in: query + required: true + schema: + type: string + description: ':class:`DeleteRequestExternalLineage`' + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLineageAPI + get: + operationId: external_lineage_list_external_lineage_relationships + summary: >- + Lists external lineage relationships of a Databricks object or external + metadata given a supplied + tags: + - catalog + - external_lineage + description: >- + Lists external lineage relationships of a Databricks object or external + metadata given a supplied + + direction. + + + :param object_info: :class:`ExternalLineageObject` + The object to query external lineage relationships for. Since this field is a query parameter, + please flatten the nested fields. For example, if the object is a table, the query parameter should + look like: `object_info.table.name=main.sales.customers` + :param lineage_direction: :class:`LineageDirection` + The lineage direction to filter on. + :param page_size: int (optional) + Specifies the maximum number of external lineage relationships to return in a single response. The + value must be less than or equal to 1000. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ExternalLineageInfo` + parameters: + - name: object_info + in: query + required: true + schema: + type: string + description: >- + The object to query external lineage relationships for. Since this + field is a query parameter, please flatten the nested fields. For + example, if the object is a table, the query parameter should look + like: `object_info.table.name=main.sales.customers` + - name: lineage_direction + in: query + required: true + schema: + type: string + description: The lineage direction to filter on. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + Specifies the maximum number of external lineage relationships to + return in a single response. The value must be less than or equal to + 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExternalLineageRelationshipsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLineageAPI + patch: + operationId: external_lineage_update_external_lineage_relationship + summary: >- + Updates an external lineage relationship between a Databricks or + external metadata object and another + tags: + - catalog + - external_lineage + description: >- + Updates an external lineage relationship between a Databricks or + external metadata object and another + + external metadata object. + + + :param external_lineage_relationship: + :class:`UpdateRequestExternalLineage` + + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`ExternalLineageRelationship` + parameters: + - name: update_mask + in: query + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + external_lineage_relationship: + type: string + description: >- + :param update_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + required: + - external_lineage_relationship + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalLineageRelationship' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLineageAPI + /api/2.1/unity-catalog/external-locations: + post: + operationId: external_locations_create + summary: >- + Creates a new external location entry in the metastore. The caller must + be a metastore admin or have + tags: + - catalog + - external_locations + description: >- + Creates a new external location entry in the metastore. The caller must + be a metastore admin or have + + the **CREATE_EXTERNAL_LOCATION** privilege on both the metastore and the + associated storage + + credential. + + + :param name: str + Name of the external location. + :param url: str + Path URL of the external location. + :param credential_name: str + Name of the storage credential used with this location. + :param comment: str (optional) + User-provided free-form text description. + :param enable_file_events: bool (optional) + Whether to enable file events on this external location. Default to `true`. Set to `false` to + disable file events. + :param encryption_details: :class:`EncryptionDetails` (optional) + + :param fallback: bool (optional) + Indicates whether fallback mode is enabled for this external location. When fallback mode is + enabled, the access to the location falls back to cluster credentials if UC credentials are not + sufficient. + :param file_event_queue: :class:`FileEventQueue` (optional) + File event queue settings. If `enable_file_events` is not `false`, must be defined and have exactly + one of the documented properties. + :param read_only: bool (optional) + Indicates whether the external location is read-only. + :param skip_validation: bool (optional) + Skips validation of the storage credential associated with the external location. + + :returns: :class:`ExternalLocationInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the external location. + url: + type: string + description: Path URL of the external location. + credential_name: + type: string + description: Name of the storage credential used with this location. + comment: + type: string + description: User-provided free-form text description. + enable_file_events: + type: string + description: >- + Whether to enable file events on this external location. + Default to `true`. Set to `false` to disable file events. + encryption_details: + type: string + description: >- + :param fallback: bool (optional) Indicates whether fallback + mode is enabled for this external location. When fallback + mode is enabled, the access to the location falls back to + cluster credentials if UC credentials are not sufficient. + fallback: + type: string + file_event_queue: + type: string + description: >- + File event queue settings. If `enable_file_events` is not + `false`, must be defined and have exactly one of the + documented properties. + read_only: + type: string + description: Indicates whether the external location is read-only. + skip_validation: + type: string + description: >- + Skips validation of the storage credential associated with + the external location. + required: + - name + - url + - credential_name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalLocationInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLocationsAPI + get: + operationId: external_locations_list + summary: >- + Gets an array of external locations (__ExternalLocationInfo__ objects) + from the metastore. The caller + tags: + - catalog + - external_locations + description: >- + Gets an array of external locations (__ExternalLocationInfo__ objects) + from the metastore. The caller + + must be a metastore admin, the owner of the external location, or a user + that has some privilege on + + the external location. There is no guarantee of a specific ordering of + the elements in the array. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param include_browse: bool (optional) + Whether to include external locations in the response for which the principal can only access + selective metadata for + :param include_unbound: bool (optional) + Whether to include external locations not bound to the workspace. Effective only if the user has + permission to update the location–workspace binding. + :param max_results: int (optional) + Maximum number of external locations to return. If not set, all the external locations are returned + (not recommended). - when set to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to 0, the page length is set to a server configured + value (recommended); - when set to a value less than 0, an invalid parameter error is returned; + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ExternalLocationInfo` + parameters: + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include external locations in the response for which the + principal can only access selective metadata for + - name: include_unbound + in: query + required: false + schema: + type: string + description: >- + Whether to include external locations not bound to the workspace. + Effective only if the user has permission to update the + location–workspace binding. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of external locations to return. If not set, all the + external locations are returned (not recommended). - when set to a + value greater than 0, the page length is the minimum of this value + and a server configured value; - when set to 0, the page length is + set to a server configured value (recommended); - when set to a + value less than 0, an invalid parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExternalLocationsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLocationsAPI + /api/2.1/unity-catalog/external-locations/{name}: + delete: + operationId: external_locations_delete + summary: >- + Deletes the specified external location from the metastore. The caller + must be the owner of the + tags: + - catalog + - external_locations + description: >- + Deletes the specified external location from the metastore. The caller + must be the owner of the + + external location. + + + :param name: str + Name of the external location. + :param force: bool (optional) + Force deletion even if there are dependent external tables or mounts. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the external location. + - name: force + in: query + required: false + schema: + type: string + description: >- + Force deletion even if there are dependent external tables or + mounts. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLocationsAPI + get: + operationId: external_locations_get + summary: >- + Gets an external location from the metastore. The caller must be either + a metastore admin, the owner + tags: + - catalog + - external_locations + description: >- + Gets an external location from the metastore. The caller must be either + a metastore admin, the owner + + of the external location, or a user that has some privilege on the + external location. + + + :param name: str + Name of the external location. + :param include_browse: bool (optional) + Whether to include external locations in the response for which the principal can only access + selective metadata for + + :returns: :class:`ExternalLocationInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the external location. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include external locations in the response for which the + principal can only access selective metadata for + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalLocationInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLocationsAPI + patch: + operationId: external_locations_update + summary: >- + Updates an external location in the metastore. The caller must be the + owner of the external location, + tags: + - catalog + - external_locations + description: >- + Updates an external location in the metastore. The caller must be the + owner of the external location, + + or be a metastore admin. In the second case, the admin can only update + the name of the external + + location. + + + :param name: str + Name of the external location. + :param comment: str (optional) + User-provided free-form text description. + :param credential_name: str (optional) + Name of the storage credential used with this location. + :param enable_file_events: bool (optional) + Whether to enable file events on this external location. Default to `true`. Set to `false` to + disable file events. + :param encryption_details: :class:`EncryptionDetails` (optional) + + :param fallback: bool (optional) + Indicates whether fallback mode is enabled for this external location. When fallback mode is + enabled, the access to the location falls back to cluster credentials if UC credentials are not + sufficient. + :param file_event_queue: :class:`FileEventQueue` (optional) + File event queue settings. If `enable_file_events` is not `false`, must be defined and have exactly + one of the documented properties. + :param force: bool (optional) + Force update even if changing url invalidates dependent external tables or mounts. + :param isolation_mode: :class:`IsolationMode` (optional) + + :param new_name: str (optional) + New name for the external location. + :param owner: str (optional) + The owner of the external location. + :param read_only: bool (optional) + Indicates whether the external location is read-only. + :param skip_validation: bool (optional) + Skips validation of the storage credential associated with the external location. + :param url: str (optional) + Path URL of the external location. + + :returns: :class:`ExternalLocationInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the external location. + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: User-provided free-form text description. + credential_name: + type: string + description: Name of the storage credential used with this location. + enable_file_events: + type: string + description: >- + Whether to enable file events on this external location. + Default to `true`. Set to `false` to disable file events. + encryption_details: + type: string + description: >- + :param fallback: bool (optional) Indicates whether fallback + mode is enabled for this external location. When fallback + mode is enabled, the access to the location falls back to + cluster credentials if UC credentials are not sufficient. + fallback: + type: string + file_event_queue: + type: string + description: >- + File event queue settings. If `enable_file_events` is not + `false`, must be defined and have exactly one of the + documented properties. + force: + type: string + description: >- + Force update even if changing url invalidates dependent + external tables or mounts. + isolation_mode: + type: string + description: >- + :param new_name: str (optional) New name for the external + location. + new_name: + type: string + owner: + type: string + description: The owner of the external location. + read_only: + type: string + description: Indicates whether the external location is read-only. + skip_validation: + type: string + description: >- + Skips validation of the storage credential associated with + the external location. + url: + type: string + description: Path URL of the external location. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalLocationInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalLocationsAPI + /api/2.0/lineage-tracking/external-metadata: + post: + operationId: external_metadata_create_external_metadata + summary: >- + Creates a new external metadata object in the parent metastore if the + caller is a metastore admin or + tags: + - catalog + - external_metadata + description: >- + Creates a new external metadata object in the parent metastore if the + caller is a metastore admin or + + has the **CREATE_EXTERNAL_METADATA** privilege. Grants **BROWSE** to all + account users upon creation + + by default. + + + :param external_metadata: :class:`ExternalMetadata` + + + :returns: :class:`ExternalMetadata` + requestBody: + content: + application/json: + schema: + type: object + properties: + external_metadata: + type: string + description: ':returns: :class:`ExternalMetadata`' + required: + - external_metadata + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalMetadata' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalMetadataAPI + get: + operationId: external_metadata_list_external_metadata + summary: >- + Gets an array of external metadata objects in the metastore. If the + caller is the metastore admin, all + tags: + - catalog + - external_metadata + description: >- + Gets an array of external metadata objects in the metastore. If the + caller is the metastore admin, all + + external metadata objects will be retrieved. Otherwise, only external + metadata objects that the caller + + has **BROWSE** on will be retrieved. There is no guarantee of a specific + ordering of the elements in + + the array. + + + :param page_size: int (optional) + Specifies the maximum number of external metadata objects to return in a single response. The value + must be less than or equal to 1000. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ExternalMetadata` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: >- + Specifies the maximum number of external metadata objects to return + in a single response. The value must be less than or equal to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExternalMetadataResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalMetadataAPI + /api/2.0/lineage-tracking/external-metadata/{name}: + delete: + operationId: external_metadata_delete_external_metadata + summary: >- + Deletes the external metadata object that matches the supplied name. The + caller must be a metastore + tags: + - catalog + - external_metadata + description: >- + Deletes the external metadata object that matches the supplied name. The + caller must be a metastore + + admin, the owner of the external metadata object, or a user that has the + **MANAGE** privilege. + + + :param name: str + parameters: + - name: name + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalMetadataAPI + get: + operationId: external_metadata_get_external_metadata + summary: >- + Gets the specified external metadata object in a metastore. The caller + must be a metastore admin, the + tags: + - catalog + - external_metadata + description: >- + Gets the specified external metadata object in a metastore. The caller + must be a metastore admin, the + + owner of the external metadata object, or a user that has the **BROWSE** + privilege. + + + :param name: str + + + :returns: :class:`ExternalMetadata` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: ':returns: :class:`ExternalMetadata`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalMetadata' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalMetadataAPI + patch: + operationId: external_metadata_update_external_metadata + summary: >- + Updates the external metadata object that matches the supplied name. The + caller can only update either + tags: + - catalog + - external_metadata + description: >- + Updates the external metadata object that matches the supplied name. The + caller can only update either + + the owner or other metadata fields in one request. The caller must be a + metastore admin, the owner of + + the external metadata object, or a user that has the **MODIFY** + privilege. If the caller is updating + + the owner, they must also have the **MANAGE** privilege. + + + :param name: str + Name of the external metadata object. + :param external_metadata: :class:`ExternalMetadata` + + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`ExternalMetadata` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the external metadata object. + - name: update_mask + in: query + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + external_metadata: + type: string + description: >- + :param update_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + required: + - external_metadata + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExternalMetadata' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExternalMetadataAPI + /api/2.1/unity-catalog/functions: + post: + operationId: functions_create + summary: '**WARNING: This API is experimental and will change in future versions**' + tags: + - catalog + - functions + description: >- + **WARNING: This API is experimental and will change in future versions** + + + Creates a new function + + + The user must have the following permissions in order for the function + to be created: - + + **USE_CATALOG** on the function's parent catalog - **USE_SCHEMA** and + **CREATE_FUNCTION** on the + + function's parent schema + + + :param function_info: :class:`CreateFunction` + Partial __FunctionInfo__ specifying the function to be created. + + :returns: :class:`FunctionInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + function_info: + type: string + description: >- + Partial __FunctionInfo__ specifying the function to be + created. + required: + - function_info + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FunctionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FunctionsAPI + get: + operationId: functions_list + summary: >- + List functions within the specified parent catalog and schema. If the + user is a metastore admin, all + tags: + - catalog + - functions + description: >- + List functions within the specified parent catalog and schema. If the + user is a metastore admin, all + + functions are returned in the output list. Otherwise, the user must have + the **USE_CATALOG** privilege + + on the catalog and the **USE_SCHEMA** privilege on the schema, and the + output list contains only + + functions for which either the user has the **EXECUTE** privilege or the + user is the owner. There is + + no guarantee of a specific ordering of the elements in the array. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param catalog_name: str + Name of parent catalog for functions of interest. + :param schema_name: str + Parent schema of functions. + :param include_browse: bool (optional) + Whether to include functions in the response for which the principal can only access selective + metadata for + :param max_results: int (optional) + Maximum number of functions to return. If not set, all the functions are returned (not recommended). + - when set to a value greater than 0, the page length is the minimum of this value and a server + configured value; - when set to 0, the page length is set to a server configured value + (recommended); - when set to a value less than 0, an invalid parameter error is returned; + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`FunctionInfo` + parameters: + - name: catalog_name + in: query + required: true + schema: + type: string + description: Name of parent catalog for functions of interest. + - name: schema_name + in: query + required: true + schema: + type: string + description: Parent schema of functions. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include functions in the response for which the principal + can only access selective metadata for + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of functions to return. If not set, all the functions + are returned (not recommended). - when set to a value greater than + 0, the page length is the minimum of this value and a server + configured value; - when set to 0, the page length is set to a + server configured value (recommended); - when set to a value less + than 0, an invalid parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFunctionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FunctionsAPI + /api/2.1/unity-catalog/functions/{name}: + delete: + operationId: functions_delete + summary: >- + Deletes the function that matches the supplied name. For the deletion to + succeed, the user must + tags: + - catalog + - functions + description: >- + Deletes the function that matches the supplied name. For the deletion to + succeed, the user must + + satisfy one of the following conditions: - Is the owner of the + function's parent catalog - Is the + + owner of the function's parent schema and have the **USE_CATALOG** + privilege on its parent catalog - + + Is the owner of the function itself and have both the **USE_CATALOG** + privilege on its parent catalog + + and the **USE_SCHEMA** privilege on its parent schema + + + :param name: str + The fully-qualified name of the function (of the form + __catalog_name__.__schema_name__.__function__name__) . + :param force: bool (optional) + Force deletion even if the function is notempty. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The fully-qualified name of the function (of the form + __catalog_name__.__schema_name__.__function__name__) . + - name: force + in: query + required: false + schema: + type: string + description: Force deletion even if the function is notempty. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FunctionsAPI + get: + operationId: functions_get + summary: >- + Gets a function from within a parent catalog and schema. For the fetch + to succeed, the user must + tags: + - catalog + - functions + description: >- + Gets a function from within a parent catalog and schema. For the fetch + to succeed, the user must + + satisfy one of the following requirements: - Is a metastore admin - Is + an owner of the function's + + parent catalog - Have the **USE_CATALOG** privilege on the function's + parent catalog and be the owner + + of the function - Have the **USE_CATALOG** privilege on the function's + parent catalog, the + + **USE_SCHEMA** privilege on the function's parent schema, and the + **EXECUTE** privilege on the + + function itself + + + :param name: str + The fully-qualified name of the function (of the form + __catalog_name__.__schema_name__.__function__name__). + :param include_browse: bool (optional) + Whether to include functions in the response for which the principal can only access selective + metadata for + + :returns: :class:`FunctionInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The fully-qualified name of the function (of the form + __catalog_name__.__schema_name__.__function__name__). + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include functions in the response for which the principal + can only access selective metadata for + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FunctionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FunctionsAPI + patch: + operationId: functions_update + summary: >- + Updates the function that matches the supplied name. Only the owner of + the function can be updated. If + tags: + - catalog + - functions + description: >- + Updates the function that matches the supplied name. Only the owner of + the function can be updated. If + + the user is not a metastore admin, the user must be a member of the + group that is the new function + + owner. - Is a metastore admin - Is the owner of the function's parent + catalog - Is the owner of the + + function's parent schema and has the **USE_CATALOG** privilege on its + parent catalog - Is the owner of + + the function itself and has the **USE_CATALOG** privilege on its parent + catalog as well as the + + **USE_SCHEMA** privilege on the function's parent schema. + + + :param name: str + The fully-qualified name of the function (of the form + __catalog_name__.__schema_name__.__function__name__). + :param owner: str (optional) + Username of current owner of the function. + + :returns: :class:`FunctionInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The fully-qualified name of the function (of the form + __catalog_name__.__schema_name__.__function__name__). + requestBody: + content: + application/json: + schema: + type: object + properties: + owner: + type: string + description: Username of current owner of the function. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FunctionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FunctionsAPI + /api/2.1/unity-catalog/permissions/{securable_type}/{full_name}: + get: + operationId: grants_get + summary: >- + Gets the permissions for a securable. Does not include inherited + permissions. + tags: + - catalog + - grants + description: >- + Gets the permissions for a securable. Does not include inherited + permissions. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param securable_type: str + Type of securable. + :param full_name: str + Full name of securable. + :param max_results: int (optional) + Specifies the maximum number of privileges to return (page length). Every PrivilegeAssignment + present in a single page response is guaranteed to contain all the privileges granted on the + requested Securable for the respective principal. + + If not set, all the permissions are returned. If set to - lesser than 0: invalid parameter error - + 0: page length is set to a server configured value - lesser than 150 but greater than 0: invalid + parameter error (this is to ensure that server is able to return at least one complete + PrivilegeAssignment in a single page response) - greater than (or equal to) 150: page length is the + minimum of this value and a server configured value + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + :param principal: str (optional) + If provided, only the permissions for the specified principal (user or group) are returned. + + :returns: :class:`GetPermissionsResponse` + parameters: + - name: securable_type + in: path + required: true + schema: + type: string + description: Type of securable. + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of securable. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Specifies the maximum number of privileges to return (page length). + Every PrivilegeAssignment present in a single page response is + guaranteed to contain all the privileges granted on the requested + Securable for the respective principal. If not set, all the + permissions are returned. If set to - lesser than 0: invalid + parameter error - 0: page length is set to a server configured value + - lesser than 150 but greater than 0: invalid parameter error (this + is to ensure that server is able to return at least one complete + PrivilegeAssignment in a single page response) - greater than (or + equal to) 150: page length is the minimum of this value and a server + configured value + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + - name: principal + in: query + required: false + schema: + type: string + description: >- + If provided, only the permissions for the specified principal (user + or group) are returned. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPermissionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GrantsAPI + patch: + operationId: grants_update + summary: Updates the permissions for a securable. + tags: + - catalog + - grants + description: |- + Updates the permissions for a securable. + + :param securable_type: str + Type of securable. + :param full_name: str + Full name of securable. + :param changes: List[:class:`PermissionsChange`] (optional) + Array of permissions change objects. + + :returns: :class:`UpdatePermissionsResponse` + parameters: + - name: securable_type + in: path + required: true + schema: + type: string + description: Type of securable. + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of securable. + requestBody: + content: + application/json: + schema: + type: object + properties: + changes: + type: string + description: Array of permissions change objects. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdatePermissionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GrantsAPI + /api/2.1/unity-catalog/effective-permissions/{securable_type}/{full_name}: + get: + operationId: grants_get_effective + summary: >- + Gets the effective permissions for a securable. Includes inherited + permissions from any parent + tags: + - catalog + - grants + description: >- + Gets the effective permissions for a securable. Includes inherited + permissions from any parent + + securables. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param securable_type: str + Type of securable. + :param full_name: str + Full name of securable. + :param max_results: int (optional) + Specifies the maximum number of privileges to return (page length). Every + EffectivePrivilegeAssignment present in a single page response is guaranteed to contain all the + effective privileges granted on (or inherited by) the requested Securable for the respective + principal. + + If not set, all the effective permissions are returned. If set to - lesser than 0: invalid parameter + error - 0: page length is set to a server configured value - lesser than 150 but greater than 0: + invalid parameter error (this is to ensure that server is able to return at least one complete + EffectivePrivilegeAssignment in a single page response) - greater than (or equal to) 150: page + length is the minimum of this value and a server configured value + :param page_token: str (optional) + Opaque token for the next page of results (pagination). + :param principal: str (optional) + If provided, only the effective permissions for the specified principal (user or group) are + returned. + + :returns: :class:`EffectivePermissionsList` + parameters: + - name: securable_type + in: path + required: true + schema: + type: string + description: Type of securable. + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of securable. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Specifies the maximum number of privileges to return (page length). + Every EffectivePrivilegeAssignment present in a single page response + is guaranteed to contain all the effective privileges granted on (or + inherited by) the requested Securable for the respective principal. + If not set, all the effective permissions are returned. If set to - + lesser than 0: invalid parameter error - 0: page length is set to a + server configured value - lesser than 150 but greater than 0: + invalid parameter error (this is to ensure that server is able to + return at least one complete EffectivePrivilegeAssignment in a + single page response) - greater than (or equal to) 150: page length + is the minimum of this value and a server configured value + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque token for the next page of results (pagination). + - name: principal + in: query + required: false + schema: + type: string + description: >- + If provided, only the effective permissions for the specified + principal (user or group) are returned. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EffectivePermissionsList' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GrantsAPI + /api/2.1/unity-catalog/workspaces/{workspace_id}/metastore: + put: + operationId: metastores_assign + summary: >- + Creates a new metastore assignment. If an assignment for the same + __workspace_id__ exists, it will be + tags: + - catalog + - metastores + description: >- + Creates a new metastore assignment. If an assignment for the same + __workspace_id__ exists, it will be + + overwritten by the new __metastore_id__ and __default_catalog_name__. + The caller must be an account + + admin. + + + :param workspace_id: int + A workspace ID. + :param metastore_id: str + The unique ID of the metastore. + :param default_catalog_name: str + The name of the default catalog in the metastore. This field is deprecated. Please use "Default + Namespace API" to configure the default catalog for a Databricks workspace. + parameters: + - name: workspace_id + in: path + required: true + schema: + type: integer + description: A workspace ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + metastore_id: + type: string + description: The unique ID of the metastore. + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field + is deprecated. Please use "Default Namespace API" to + configure the default catalog for a Databricks workspace. + required: + - metastore_id + - default_catalog_name + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + delete: + operationId: metastores_unassign + summary: >- + Deletes a metastore assignment. The caller must be an account + administrator. + tags: + - catalog + - metastores + description: >- + Deletes a metastore assignment. The caller must be an account + administrator. + + + :param workspace_id: int + A workspace ID. + :param metastore_id: str + Query for the ID of the metastore to delete. + parameters: + - name: workspace_id + in: path + required: true + schema: + type: integer + description: A workspace ID. + - name: metastore_id + in: query + required: true + schema: + type: string + description: Query for the ID of the metastore to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + patch: + operationId: metastores_update_assignment + summary: >- + Updates a metastore assignment. This operation can be used to update + __metastore_id__ or + tags: + - catalog + - metastores + description: >- + Updates a metastore assignment. This operation can be used to update + __metastore_id__ or + + __default_catalog_name__ for a specified Workspace, if the Workspace is + already assigned a metastore. + + The caller must be an account admin to update __metastore_id__; + otherwise, the caller can be a + + Workspace admin. + + + :param workspace_id: int + A workspace ID. + :param default_catalog_name: str (optional) + The name of the default catalog in the metastore. This field is deprecated. Please use "Default + Namespace API" to configure the default catalog for a Databricks workspace. + :param metastore_id: str (optional) + The unique ID of the metastore. + parameters: + - name: workspace_id + in: path + required: true + schema: + type: integer + description: A workspace ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field + is deprecated. Please use "Default Namespace API" to + configure the default catalog for a Databricks workspace. + metastore_id: + type: string + description: The unique ID of the metastore. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + /api/2.1/unity-catalog/metastores: + post: + operationId: metastores_create + summary: >- + Creates a new metastore based on a provided name and optional storage + root path. By default (if the + tags: + - catalog + - metastores + description: >- + Creates a new metastore based on a provided name and optional storage + root path. By default (if the + + __owner__ field is not set), the owner of the new metastore is the user + calling the + + __createMetastore__ API. If the __owner__ field is set to the empty + string (**""**), the ownership is + + assigned to the System User instead. + + + :param name: str + The user-specified name of the metastore. + :param external_access_enabled: bool (optional) + Whether to allow non-DBR clients to directly access entities under the metastore. + :param region: str (optional) + Cloud region which the metastore serves (e.g., `us-west-2`, `westus`). + :param storage_root: str (optional) + The storage root URL for metastore + + :returns: :class:`MetastoreInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: The user-specified name of the metastore. + external_access_enabled: + type: string + description: >- + Whether to allow non-DBR clients to directly access entities + under the metastore. + region: + type: string + description: >- + Cloud region which the metastore serves (e.g., `us-west-2`, + `westus`). + storage_root: + type: string + description: The storage root URL for metastore + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MetastoreInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + get: + operationId: metastores_list + summary: >- + Gets an array of the available metastores (as __MetastoreInfo__ + objects). The caller must be an admin + tags: + - catalog + - metastores + description: >- + Gets an array of the available metastores (as __MetastoreInfo__ + objects). The caller must be an admin + + to retrieve this info. There is no guarantee of a specific ordering of + the elements in the array. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param max_results: int (optional) + Maximum number of metastores to return. - when set to a value greater than 0, the page length is the + minimum of this value and a server configured value; - when set to 0, the page length is set to a + server configured value (recommended); - when set to a value less than 0, an invalid parameter error + is returned; - If not set, all the metastores are returned (not recommended). - Note: The number of + returned metastores might be less than the specified max_results size, even zero. The only + definitive indication that no further metastores can be fetched is when the next_page_token is unset + from the response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`MetastoreInfo` + parameters: + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of metastores to return. - when set to a value + greater than 0, the page length is the minimum of this value and a + server configured value; - when set to 0, the page length is set to + a server configured value (recommended); - when set to a value less + than 0, an invalid parameter error is returned; - If not set, all + the metastores are returned (not recommended). - Note: The number of + returned metastores might be less than the specified max_results + size, even zero. The only definitive indication that no further + metastores can be fetched is when the next_page_token is unset from + the response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountsListMetastoresResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + /api/2.1/unity-catalog/current-metastore-assignment: + get: + operationId: metastores_current + summary: Gets the metastore assignment for the workspace being accessed. + tags: + - catalog + - metastores + description: |- + Gets the metastore assignment for the workspace being accessed. + + + :returns: :class:`MetastoreAssignment` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MetastoreAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + /api/2.1/unity-catalog/metastores/{id}: + delete: + operationId: metastores_delete + summary: Deletes a metastore. The caller must be a metastore admin. + tags: + - catalog + - metastores + description: |- + Deletes a metastore. The caller must be a metastore admin. + + :param id: str + Unique ID of the metastore. + :param force: bool (optional) + Force deletion even if the metastore is not empty. Default is false. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID of the metastore. + - name: force + in: query + required: false + schema: + type: string + description: Force deletion even if the metastore is not empty. Default is false. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + get: + operationId: metastores_get + summary: >- + Gets a metastore that matches the supplied ID. The caller must be a + metastore admin to retrieve this + tags: + - catalog + - metastores + description: >- + Gets a metastore that matches the supplied ID. The caller must be a + metastore admin to retrieve this + + info. + + + :param id: str + Unique ID of the metastore. + + :returns: :class:`MetastoreInfo` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID of the metastore. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MetastoreInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + patch: + operationId: metastores_update + summary: >- + Updates information for a specific metastore. The caller must be a + metastore admin. If the __owner__ + tags: + - catalog + - metastores + description: >- + Updates information for a specific metastore. The caller must be a + metastore admin. If the __owner__ + + field is set to the empty string (**""**), the ownership is updated to + the System User. + + + :param id: str + Unique ID of the metastore. + :param delta_sharing_organization_name: str (optional) + The organization name of a Delta Sharing entity, to be used in Databricks-to-Databricks Delta + Sharing as the official name. + :param delta_sharing_recipient_token_lifetime_in_seconds: int (optional) + The lifetime of delta sharing recipient token in seconds. + :param delta_sharing_scope: :class:`DeltaSharingScopeEnum` (optional) + The scope of Delta Sharing enabled for the metastore. + :param external_access_enabled: bool (optional) + Whether to allow non-DBR clients to directly access entities under the metastore. + :param new_name: str (optional) + New name for the metastore. + :param owner: str (optional) + The owner of the metastore. + :param privilege_model_version: str (optional) + Privilege model version of the metastore, of the form `major.minor` (e.g., `1.0`). + :param storage_root_credential_id: str (optional) + UUID of storage credential to access the metastore storage_root. + + :returns: :class:`MetastoreInfo` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID of the metastore. + requestBody: + content: + application/json: + schema: + type: object + properties: + delta_sharing_organization_name: + type: string + description: >- + The organization name of a Delta Sharing entity, to be used + in Databricks-to-Databricks Delta Sharing as the official + name. + delta_sharing_recipient_token_lifetime_in_seconds: + type: string + description: The lifetime of delta sharing recipient token in seconds. + delta_sharing_scope: + type: string + description: The scope of Delta Sharing enabled for the metastore. + external_access_enabled: + type: string + description: >- + Whether to allow non-DBR clients to directly access entities + under the metastore. + new_name: + type: string + description: New name for the metastore. + owner: + type: string + description: The owner of the metastore. + privilege_model_version: + type: string + description: >- + Privilege model version of the metastore, of the form + `major.minor` (e.g., `1.0`). + storage_root_credential_id: + type: string + description: >- + UUID of storage credential to access the metastore + storage_root. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MetastoreInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + /api/2.1/unity-catalog/metastore_summary: + get: + operationId: metastores_summary + summary: >- + Gets information about a metastore. This summary includes the storage + credential, the cloud vendor, + tags: + - catalog + - metastores + description: >- + Gets information about a metastore. This summary includes the storage + credential, the cloud vendor, + + the cloud region, and the global metastore ID. + + + + :returns: :class:`GetMetastoreSummaryResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetMetastoreSummaryResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MetastoresAPI + /api/2.1/unity-catalog/models/{full_name}/versions/{version}: + delete: + operationId: model_versions_delete + summary: >- + Deletes a model version from the specified registered model. Any aliases + assigned to the model version + tags: + - catalog + - model_versions + description: >- + Deletes a model version from the specified registered model. Any aliases + assigned to the model version + + will also be deleted. + + + The caller must be a metastore admin or an owner of the parent + registered model. For the latter case, + + the caller must also be the owner or have the **USE_CATALOG** privilege + on the parent catalog and the + + **USE_SCHEMA** privilege on the parent schema. + + + :param full_name: str + The three-level (fully qualified) name of the model version + :param version: int + The integer version number of the model version + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the model version + - name: version + in: path + required: true + schema: + type: integer + description: The integer version number of the model version + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelVersionsAPI + get: + operationId: model_versions_get + summary: Get a model version. + tags: + - catalog + - model_versions + description: >- + Get a model version. + + + The caller must be a metastore admin or an owner of (or have the + **EXECUTE** privilege on) the parent + + registered model. For the latter case, the caller must also be the owner + or have the **USE_CATALOG** + + privilege on the parent catalog and the **USE_SCHEMA** privilege on the + parent schema. + + + :param full_name: str + The three-level (fully qualified) name of the model version + :param version: int + The integer version number of the model version + :param include_aliases: bool (optional) + Whether to include aliases associated with the model version in the response + :param include_browse: bool (optional) + Whether to include model versions in the response for which the principal can only access selective + metadata for + + :returns: :class:`ModelVersionInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the model version + - name: version + in: path + required: true + schema: + type: integer + description: The integer version number of the model version + - name: include_aliases + in: query + required: false + schema: + type: string + description: >- + Whether to include aliases associated with the model version in the + response + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include model versions in the response for which the + principal can only access selective metadata for + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ModelVersionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelVersionsAPI + patch: + operationId: model_versions_update + summary: Updates the specified model version. + tags: + - catalog + - model_versions + description: >- + Updates the specified model version. + + + The caller must be a metastore admin or an owner of the parent + registered model. For the latter case, + + the caller must also be the owner or have the **USE_CATALOG** privilege + on the parent catalog and the + + **USE_SCHEMA** privilege on the parent schema. + + + Currently only the comment of the model version can be updated. + + + :param full_name: str + The three-level (fully qualified) name of the model version + :param version: int + The integer version number of the model version + :param aliases: List[:class:`RegisteredModelAlias`] (optional) + List of aliases associated with the model version + :param catalog_name: str (optional) + The name of the catalog containing the model version + :param comment: str (optional) + The comment attached to the model version + :param created_at: int (optional) + + :param created_by: str (optional) + The identifier of the user who created the model version + :param id: str (optional) + The unique identifier of the model version + :param metastore_id: str (optional) + The unique identifier of the metastore containing the model version + :param model_name: str (optional) + The name of the parent registered model of the model version, relative to parent schema + :param model_version_dependencies: :class:`DependencyList` (optional) + Model version dependencies, for feature-store packaged models + :param run_id: str (optional) + MLflow run ID used when creating the model version, if ``source`` was generated by an experiment run + stored in an MLflow tracking server + :param run_workspace_id: int (optional) + ID of the Databricks workspace containing the MLflow run that generated this model version, if + applicable + :param schema_name: str (optional) + The name of the schema containing the model version, relative to parent catalog + :param source: str (optional) + URI indicating the location of the source artifacts (files) for the model version + :param status: :class:`ModelVersionInfoStatus` (optional) + Current status of the model version. Newly created model versions start in PENDING_REGISTRATION + status, then move to READY status once the model version files are uploaded and the model version is + finalized. Only model versions in READY status can be loaded for inference or served. + :param storage_location: str (optional) + The storage location on the cloud under which model version data files are stored + :param updated_at: int (optional) + + :param updated_by: str (optional) + The identifier of the user who updated the model version last time + + :returns: :class:`ModelVersionInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the model version + - name: version + in: path + required: true + schema: + type: integer + description: The integer version number of the model version + requestBody: + content: + application/json: + schema: + type: object + properties: + aliases: + type: string + description: List of aliases associated with the model version + catalog_name: + type: string + description: The name of the catalog containing the model version + comment: + type: string + description: The comment attached to the model version + created_at: + type: string + description: >- + :param created_by: str (optional) The identifier of the user + who created the model version + created_by: + type: string + id: + type: string + description: The unique identifier of the model version + metastore_id: + type: string + description: >- + The unique identifier of the metastore containing the model + version + model_name: + type: string + description: >- + The name of the parent registered model of the model + version, relative to parent schema + model_version_dependencies: + type: string + description: >- + Model version dependencies, for feature-store packaged + models + run_id: + type: string + description: >- + MLflow run ID used when creating the model version, if + ``source`` was generated by an experiment run stored in an + MLflow tracking server + run_workspace_id: + type: string + description: >- + ID of the Databricks workspace containing the MLflow run + that generated this model version, if applicable + schema_name: + type: string + description: >- + The name of the schema containing the model version, + relative to parent catalog + source: + type: string + description: >- + URI indicating the location of the source artifacts (files) + for the model version + status: + type: string + description: >- + Current status of the model version. Newly created model + versions start in PENDING_REGISTRATION status, then move to + READY status once the model version files are uploaded and + the model version is finalized. Only model versions in READY + status can be loaded for inference or served. + storage_location: + type: string + description: >- + The storage location on the cloud under which model version + data files are stored + updated_at: + type: string + description: >- + :param updated_by: str (optional) The identifier of the user + who updated the model version last time + updated_by: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ModelVersionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelVersionsAPI + /api/2.1/unity-catalog/models/{full_name}/aliases/{alias}: + get: + operationId: model_versions_get_by_alias + summary: Get a model version by alias. + tags: + - catalog + - model_versions + description: >- + Get a model version by alias. + + + The caller must be a metastore admin or an owner of (or have the + **EXECUTE** privilege on) the + + registered model. For the latter case, the caller must also be the owner + or have the **USE_CATALOG** + + privilege on the parent catalog and the **USE_SCHEMA** privilege on the + parent schema. + + + :param full_name: str + The three-level (fully qualified) name of the registered model + :param alias: str + The name of the alias + :param include_aliases: bool (optional) + Whether to include aliases associated with the model version in the response + + :returns: :class:`ModelVersionInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the registered model + - name: alias + in: path + required: true + schema: + type: string + description: The name of the alias + - name: include_aliases + in: query + required: false + schema: + type: string + description: >- + Whether to include aliases associated with the model version in the + response + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ModelVersionInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelVersionsAPI + delete: + operationId: registered_models_delete_alias + summary: Deletes a registered model alias. + tags: + - catalog + - registered_models + description: >- + Deletes a registered model alias. + + + The caller must be a metastore admin or an owner of the registered + model. For the latter case, the + + caller must also be the owner or have the **USE_CATALOG** privilege on + the parent catalog and the + + **USE_SCHEMA** privilege on the parent schema. + + + :param full_name: str + The three-level (fully qualified) name of the registered model + :param alias: str + The name of the alias + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the registered model + - name: alias + in: path + required: true + schema: + type: string + description: The name of the alias + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RegisteredModelsAPI + put: + operationId: registered_models_set_alias + summary: Set an alias on the specified registered model. + tags: + - catalog + - registered_models + description: >- + Set an alias on the specified registered model. + + + The caller must be a metastore admin or an owner of the registered + model. For the latter case, the + + caller must also be the owner or have the **USE_CATALOG** privilege on + the parent catalog and the + + **USE_SCHEMA** privilege on the parent schema. + + + :param full_name: str + The three-level (fully qualified) name of the registered model + :param alias: str + The name of the alias + :param version_num: int + The version number of the model version to which the alias points + + :returns: :class:`RegisteredModelAlias` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the registered model + - name: alias + in: path + required: true + schema: + type: string + description: The name of the alias + requestBody: + content: + application/json: + schema: + type: object + properties: + version_num: + type: integer + description: >- + The version number of the model version to which the alias + points + required: + - version_num + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegisteredModelAlias' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RegisteredModelsAPI + /api/2.1/unity-catalog/models/{full_name}/versions: + get: + operationId: model_versions_list + summary: >- + List model versions. You can list model versions under a particular + schema, or list all model versions + tags: + - catalog + - model_versions + description: >- + List model versions. You can list model versions under a particular + schema, or list all model versions + + in the current metastore. + + + The returned models are filtered based on the privileges of the calling + user. For example, the + + metastore admin is able to list all the model versions. A regular user + needs to be the owner or have + + the **EXECUTE** privilege on the parent registered model to recieve the + model versions in the + + response. For the latter case, the caller must also be the owner or have + the **USE_CATALOG** privilege + + on the parent catalog and the **USE_SCHEMA** privilege on the parent + schema. + + + There is no guarantee of a specific ordering of the elements in the + response. The elements in the + + response will not contain any aliases or tags. + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param full_name: str + The full three-level name of the registered model under which to list model versions + :param include_browse: bool (optional) + Whether to include model versions in the response for which the principal can only access selective + metadata for + :param max_results: int (optional) + Maximum number of model versions to return. If not set, the page length is set to a server + configured value (100, as of 1/3/2024). - when set to a value greater than 0, the page length is the + minimum of this value and a server configured value(1000, as of 1/3/2024); - when set to 0, the page + length is set to a server configured value (100, as of 1/3/2024) (recommended); - when set to a + value less than 0, an invalid parameter error is returned; + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ModelVersionInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: >- + The full three-level name of the registered model under which to + list model versions + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include model versions in the response for which the + principal can only access selective metadata for + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of model versions to return. If not set, the page + length is set to a server configured value (100, as of 1/3/2024). - + when set to a value greater than 0, the page length is the minimum + of this value and a server configured value(1000, as of 1/3/2024); - + when set to 0, the page length is set to a server configured value + (100, as of 1/3/2024) (recommended); - when set to a value less than + 0, an invalid parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListModelVersionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelVersionsAPI + /api/2.0/online-tables: + post: + operationId: online_tables_create + summary: Create a new Online Table. + tags: + - catalog + - online_tables + description: |- + Create a new Online Table. + + :param table: :class:`OnlineTable` + Specification of the online table to be created. + + :returns: + Long-running operation waiter for :class:`OnlineTable`. + See :method:wait_get_online_table_active for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + table: + type: string + description: Specification of the online table to be created. + required: + - table + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/OnlineTable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: OnlineTablesAPI + /api/2.0/online-tables/{name}: + delete: + operationId: online_tables_delete + summary: >- + Delete an online table. Warning: This will delete all the data in the + online table. If the source + tags: + - catalog + - online_tables + description: >- + Delete an online table. Warning: This will delete all the data in the + online table. If the source + + Delta table was deleted or modified since this Online Table was created, + this will lose the data + + forever! + + + :param name: str + Full three-part (catalog, schema, table) name of the table. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Full three-part (catalog, schema, table) name of the table. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: OnlineTablesAPI + get: + operationId: online_tables_get + summary: Get information about an existing online table and its status. + tags: + - catalog + - online_tables + description: |- + Get information about an existing online table and its status. + + :param name: str + Full three-part (catalog, schema, table) name of the table. + + :returns: :class:`OnlineTable` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Full three-part (catalog, schema, table) name of the table. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/OnlineTable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: OnlineTablesAPI + /api/2.1/unity-catalog/policies: + post: + operationId: policies_create_policy + summary: >- + Creates a new policy on a securable. The new policy applies to the + securable and all its descendants. + tags: + - catalog + - policies + description: >- + Creates a new policy on a securable. The new policy applies to the + securable and all its descendants. + + + :param policy_info: :class:`PolicyInfo` + Required. The policy to create. + + :returns: :class:`PolicyInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + policy_info: + type: string + description: Required. The policy to create. + required: + - policy_info + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PolicyInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PoliciesAPI + /api/2.1/unity-catalog/policies/{on_securable_type}/{on_securable_fullname}/{name}: + delete: + operationId: policies_delete_policy + summary: Delete an ABAC policy defined on a securable. + tags: + - catalog + - policies + description: |- + Delete an ABAC policy defined on a securable. + + :param on_securable_type: str + Required. The type of the securable to delete the policy from. + :param on_securable_fullname: str + Required. The fully qualified name of the securable to delete the policy from. + :param name: str + Required. The name of the policy to delete + + :returns: :class:`DeletePolicyResponse` + parameters: + - name: on_securable_type + in: path + required: true + schema: + type: string + description: Required. The type of the securable to delete the policy from. + - name: on_securable_fullname + in: path + required: true + schema: + type: string + description: >- + Required. The fully qualified name of the securable to delete the + policy from. + - name: name + in: path + required: true + schema: + type: string + description: Required. The name of the policy to delete + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeletePolicyResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PoliciesAPI + get: + operationId: policies_get_policy + summary: Get the policy definition on a securable + tags: + - catalog + - policies + description: |- + Get the policy definition on a securable + + :param on_securable_type: str + Required. The type of the securable to retrieve the policy for. + :param on_securable_fullname: str + Required. The fully qualified name of securable to retrieve policy for. + :param name: str + Required. The name of the policy to retrieve. + + :returns: :class:`PolicyInfo` + parameters: + - name: on_securable_type + in: path + required: true + schema: + type: string + description: Required. The type of the securable to retrieve the policy for. + - name: on_securable_fullname + in: path + required: true + schema: + type: string + description: >- + Required. The fully qualified name of securable to retrieve policy + for. + - name: name + in: path + required: true + schema: + type: string + description: Required. The name of the policy to retrieve. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PolicyInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PoliciesAPI + patch: + operationId: policies_update_policy + summary: Update an ABAC policy on a securable. + tags: + - catalog + - policies + description: |- + Update an ABAC policy on a securable. + + :param on_securable_type: str + Required. The type of the securable to update the policy for. + :param on_securable_fullname: str + Required. The fully qualified name of the securable to update the policy for. + :param name: str + Required. The name of the policy to update. + :param policy_info: :class:`PolicyInfo` + Optional fields to update. This is the request body for updating a policy. Use `update_mask` field + to specify which fields in the request is to be updated. - If `update_mask` is empty or "*", all + specified fields will be updated. - If `update_mask` is specified, only the fields specified in the + `update_mask` will be updated. If a field is specified in `update_mask` and not set in the request, + the field will be cleared. Users can use the update mask to explicitly unset optional fields such as + `exception_principals` and `when_condition`. + :param update_mask: str (optional) + Optional. The update mask field for specifying user intentions on which fields to update in the + request. + + :returns: :class:`PolicyInfo` + parameters: + - name: on_securable_type + in: path + required: true + schema: + type: string + description: Required. The type of the securable to update the policy for. + - name: on_securable_fullname + in: path + required: true + schema: + type: string + description: >- + Required. The fully qualified name of the securable to update the + policy for. + - name: name + in: path + required: true + schema: + type: string + description: Required. The name of the policy to update. + - name: update_mask + in: query + required: false + schema: + type: string + description: >- + Optional. The update mask field for specifying user intentions on + which fields to update in the request. + requestBody: + content: + application/json: + schema: + type: object + properties: + policy_info: + type: string + description: >- + Optional fields to update. This is the request body for + updating a policy. Use `update_mask` field to specify which + fields in the request is to be updated. - If `update_mask` + is empty or "*", all specified fields will be updated. - If + `update_mask` is specified, only the fields specified in the + `update_mask` will be updated. If a field is specified in + `update_mask` and not set in the request, the field will be + cleared. Users can use the update mask to explicitly unset + optional fields such as `exception_principals` and + `when_condition`. + required: + - policy_info + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PolicyInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PoliciesAPI + /api/2.1/unity-catalog/policies/{on_securable_type}/{on_securable_fullname}: + get: + operationId: policies_list_policies + summary: >- + List all policies defined on a securable. Optionally, the list can + include inherited policies defined + tags: + - catalog + - policies + description: >- + List all policies defined on a securable. Optionally, the list can + include inherited policies defined + + on the securable's parent schema or catalog. + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param on_securable_type: str + Required. The type of the securable to list policies for. + :param on_securable_fullname: str + Required. The fully qualified name of securable to list policies for. + :param include_inherited: bool (optional) + Optional. Whether to include policies defined on parent securables. By default, the inherited + policies are not included. + :param max_results: int (optional) + Optional. Maximum number of policies to return on a single page (page length). - When not set or set + to 0, the page length is set to a server configured value (recommended); - When set to a value + greater than 0, the page length is the minimum of this value and a server configured value; + :param page_token: str (optional) + Optional. Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`PolicyInfo` + parameters: + - name: on_securable_type + in: path + required: true + schema: + type: string + description: Required. The type of the securable to list policies for. + - name: on_securable_fullname + in: path + required: true + schema: + type: string + description: >- + Required. The fully qualified name of securable to list policies + for. + - name: include_inherited + in: query + required: false + schema: + type: string + description: >- + Optional. Whether to include policies defined on parent securables. + By default, the inherited policies are not included. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Optional. Maximum number of policies to return on a single page + (page length). - When not set or set to 0, the page length is set to + a server configured value (recommended); - When set to a value + greater than 0, the page length is the minimum of this value and a + server configured value; + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Optional. Opaque pagination token to go to next page based on + previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PoliciesAPI + /api/2.1/unity-catalog/tables/{table_name}/monitor/refreshes/{refresh_id}/cancel: + post: + operationId: quality_monitors_cancel_refresh + summary: >- + [DEPRECATED] Cancels an already-initiated refresh job. Use Data Quality + Monitors API instead + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Cancels an already-initiated refresh job. Use Data Quality + Monitors API instead + + (/api/data-quality/v1/monitors). + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. table_name is case insensitive and spaces are + disallowed. + :param refresh_id: int + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. table_name is + case insensitive and spaces are disallowed. + - name: refresh_id + in: path + required: true + schema: + type: integer + description: int + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + /api/2.1/unity-catalog/tables/{table_name}/monitor: + post: + operationId: quality_monitors_create + summary: >- + [DEPRECATED] Creates a new monitor for the specified table. Use Data + Quality Monitors API instead + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Creates a new monitor for the specified table. Use Data + Quality Monitors API instead + + (/api/data-quality/v1/monitors). + + + The caller must either: 1. be an owner of the table's parent catalog, + have **USE_SCHEMA** on the + + table's parent schema, and have **SELECT** access on the table 2. have + **USE_CATALOG** on the table's + + parent catalog, be an owner of the table's parent schema, and have + **SELECT** access on the table. 3. + + have the following permissions: - **USE_CATALOG** on the table's parent + catalog - **USE_SCHEMA** on + + the table's parent schema - be an owner of the table. + + + Workspace assets, such as the dashboard, will be created in the + workspace where this call was made. + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. This field corresponds to the + {full_table_name_arg} arg in the endpoint path. + :param output_schema_name: str + [Create:REQ Update:REQ] Schema where output tables are created. Needs to be in 2-level format + {catalog}.{schema} + :param assets_dir: str + [Create:REQ Update:IGN] Field for specifying the absolute path to a custom directory to store + data-monitoring assets. Normally prepopulated to a default user location via UI and Python APIs. + :param baseline_table_name: str (optional) + [Create:OPT Update:OPT] Baseline table name. Baseline data is used to compute drift from the data in + the monitored `table_name`. The baseline table and the monitored table shall have the same schema. + :param custom_metrics: List[:class:`MonitorMetric`] (optional) + [Create:OPT Update:OPT] Custom metrics. + :param data_classification_config: + :class:`MonitorDataClassificationConfig` (optional) + [Create:OPT Update:OPT] Data classification related config. + :param inference_log: :class:`MonitorInferenceLog` (optional) + + :param latest_monitor_failure_msg: str (optional) + [Create:ERR Update:IGN] The latest error message for a monitor failure. + :param notifications: :class:`MonitorNotifications` (optional) + [Create:OPT Update:OPT] Field for specifying notification settings. + :param schedule: :class:`MonitorCronSchedule` (optional) + [Create:OPT Update:OPT] The monitor schedule. + :param skip_builtin_dashboard: bool (optional) + Whether to skip creating a default dashboard summarizing data quality metrics. + :param slicing_exprs: List[str] (optional) + [Create:OPT Update:OPT] List of column expressions to slice data with for targeted analysis. The + data is grouped by each expression independently, resulting in a separate slice for each predicate + and its complements. For example `slicing_exprs=[“col_1”, “col_2 > 10”]` will generate the + following slices: two slices for `col_2 > 10` (True and False), and one slice per unique value in + `col1`. For high-cardinality columns, only the top 100 unique values by frequency will generate + slices. + :param snapshot: :class:`MonitorSnapshot` (optional) + Configuration for monitoring snapshot tables. + :param time_series: :class:`MonitorTimeSeries` (optional) + Configuration for monitoring time series tables. + :param warehouse_id: str (optional) + Optional argument to specify the warehouse for dashboard creation. If not specified, the first + running warehouse will be used. + + :returns: :class:`MonitorInfo` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. This field + corresponds to the {full_table_name_arg} arg in the endpoint path. + requestBody: + content: + application/json: + schema: + type: object + properties: + output_schema_name: + type: string + description: >- + [Create:REQ Update:REQ] Schema where output tables are + created. Needs to be in 2-level format {catalog}.{schema} + assets_dir: + type: string + description: >- + [Create:REQ Update:IGN] Field for specifying the absolute + path to a custom directory to store data-monitoring assets. + Normally prepopulated to a default user location via UI and + Python APIs. + baseline_table_name: + type: string + description: >- + [Create:OPT Update:OPT] Baseline table name. Baseline data + is used to compute drift from the data in the monitored + `table_name`. The baseline table and the monitored table + shall have the same schema. + custom_metrics: + type: string + description: '[Create:OPT Update:OPT] Custom metrics.' + data_classification_config: + type: string + description: '[Create:OPT Update:OPT] Data classification related config.' + inference_log: + type: string + description: >- + :param latest_monitor_failure_msg: str (optional) + [Create:ERR Update:IGN] The latest error message for a + monitor failure. + latest_monitor_failure_msg: + type: string + notifications: + type: string + description: >- + [Create:OPT Update:OPT] Field for specifying notification + settings. + schedule: + type: string + description: '[Create:OPT Update:OPT] The monitor schedule.' + skip_builtin_dashboard: + type: string + description: >- + Whether to skip creating a default dashboard summarizing + data quality metrics. + slicing_exprs: + type: string + description: >- + [Create:OPT Update:OPT] List of column expressions to slice + data with for targeted analysis. The data is grouped by each + expression independently, resulting in a separate slice for + each predicate and its complements. For example + `slicing_exprs=[“col_1”, “col_2 > 10”]` will generate the + following slices: two slices for `col_2 > 10` (True and + False), and one slice per unique value in `col1`. For + high-cardinality columns, only the top 100 unique values by + frequency will generate slices. + snapshot: + type: string + description: Configuration for monitoring snapshot tables. + time_series: + type: string + description: Configuration for monitoring time series tables. + warehouse_id: + type: string + description: >- + Optional argument to specify the warehouse for dashboard + creation. If not specified, the first running warehouse will + be used. + required: + - output_schema_name + - assets_dir + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MonitorInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + delete: + operationId: quality_monitors_delete + summary: >- + [DEPRECATED] Deletes a monitor for the specified table. Use Data Quality + Monitors API instead + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Deletes a monitor for the specified table. Use Data Quality + Monitors API instead + + (/api/data-quality/v1/monitors). + + + The caller must either: 1. be an owner of the table's parent catalog 2. + have **USE_CATALOG** on the + + table's parent catalog and be an owner of the table's parent schema 3. + have the following permissions: + + - **USE_CATALOG** on the table's parent catalog - **USE_SCHEMA** on the + table's parent schema - be an + + owner of the table. + + + Additionally, the call must be made from the workspace where the monitor + was created. + + + Note that the metric tables and dashboard will not be deleted as part of + this call; those assets must + + be manually cleaned up (if desired). + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. This field corresponds to the + {full_table_name_arg} arg in the endpoint path. + + :returns: :class:`DeleteMonitorResponse` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. This field + corresponds to the {full_table_name_arg} arg in the endpoint path. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteMonitorResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + get: + operationId: quality_monitors_get + summary: >- + [DEPRECATED] Gets a monitor for the specified table. Use Data Quality + Monitors API instead + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Gets a monitor for the specified table. Use Data Quality + Monitors API instead + + (/api/data-quality/v1/monitors). + + + The caller must either: 1. be an owner of the table's parent catalog 2. + have **USE_CATALOG** on the + + table's parent catalog and be an owner of the table's parent schema. 3. + have the following + + permissions: - **USE_CATALOG** on the table's parent catalog - + **USE_SCHEMA** on the table's parent + + schema - **SELECT** privilege on the table. + + + The returned information includes configuration values, as well as + information on assets created by + + the monitor. Some information (e.g., dashboard) may be filtered out if + the caller is in a different + + workspace than where the monitor was created. + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. This field corresponds to the + {full_table_name_arg} arg in the endpoint path. + + :returns: :class:`MonitorInfo` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. This field + corresponds to the {full_table_name_arg} arg in the endpoint path. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MonitorInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + put: + operationId: quality_monitors_update + summary: >- + [DEPRECATED] Updates a monitor for the specified table. Use Data Quality + Monitors API instead + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Updates a monitor for the specified table. Use Data Quality + Monitors API instead + + (/api/data-quality/v1/monitors). + + + The caller must either: 1. be an owner of the table's parent catalog 2. + have **USE_CATALOG** on the + + table's parent catalog and be an owner of the table's parent schema 3. + have the following permissions: + + - **USE_CATALOG** on the table's parent catalog - **USE_SCHEMA** on the + table's parent schema - be an + + owner of the table. + + + Additionally, the call must be made from the workspace where the monitor + was created, and the caller + + must be the original creator of the monitor. + + + Certain configuration fields, such as output asset identifiers, cannot + be updated. + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. This field corresponds to the + {full_table_name_arg} arg in the endpoint path. + :param output_schema_name: str + [Create:REQ Update:REQ] Schema where output tables are created. Needs to be in 2-level format + {catalog}.{schema} + :param baseline_table_name: str (optional) + [Create:OPT Update:OPT] Baseline table name. Baseline data is used to compute drift from the data in + the monitored `table_name`. The baseline table and the monitored table shall have the same schema. + :param custom_metrics: List[:class:`MonitorMetric`] (optional) + [Create:OPT Update:OPT] Custom metrics. + :param dashboard_id: str (optional) + [Create:ERR Update:OPT] Id of dashboard that visualizes the computed metrics. This can be empty if + the monitor is in PENDING state. + :param data_classification_config: + :class:`MonitorDataClassificationConfig` (optional) + [Create:OPT Update:OPT] Data classification related config. + :param inference_log: :class:`MonitorInferenceLog` (optional) + + :param latest_monitor_failure_msg: str (optional) + [Create:ERR Update:IGN] The latest error message for a monitor failure. + :param notifications: :class:`MonitorNotifications` (optional) + [Create:OPT Update:OPT] Field for specifying notification settings. + :param schedule: :class:`MonitorCronSchedule` (optional) + [Create:OPT Update:OPT] The monitor schedule. + :param slicing_exprs: List[str] (optional) + [Create:OPT Update:OPT] List of column expressions to slice data with for targeted analysis. The + data is grouped by each expression independently, resulting in a separate slice for each predicate + and its complements. For example `slicing_exprs=[“col_1”, “col_2 > 10”]` will generate the + following slices: two slices for `col_2 > 10` (True and False), and one slice per unique value in + `col1`. For high-cardinality columns, only the top 100 unique values by frequency will generate + slices. + :param snapshot: :class:`MonitorSnapshot` (optional) + Configuration for monitoring snapshot tables. + :param time_series: :class:`MonitorTimeSeries` (optional) + Configuration for monitoring time series tables. + + :returns: :class:`MonitorInfo` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. This field + corresponds to the {full_table_name_arg} arg in the endpoint path. + requestBody: + content: + application/json: + schema: + type: object + properties: + output_schema_name: + type: string + description: >- + [Create:REQ Update:REQ] Schema where output tables are + created. Needs to be in 2-level format {catalog}.{schema} + baseline_table_name: + type: string + description: >- + [Create:OPT Update:OPT] Baseline table name. Baseline data + is used to compute drift from the data in the monitored + `table_name`. The baseline table and the monitored table + shall have the same schema. + custom_metrics: + type: string + description: '[Create:OPT Update:OPT] Custom metrics.' + dashboard_id: + type: string + description: >- + [Create:ERR Update:OPT] Id of dashboard that visualizes the + computed metrics. This can be empty if the monitor is in + PENDING state. + data_classification_config: + type: string + description: '[Create:OPT Update:OPT] Data classification related config.' + inference_log: + type: string + description: >- + :param latest_monitor_failure_msg: str (optional) + [Create:ERR Update:IGN] The latest error message for a + monitor failure. + latest_monitor_failure_msg: + type: string + notifications: + type: string + description: >- + [Create:OPT Update:OPT] Field for specifying notification + settings. + schedule: + type: string + description: '[Create:OPT Update:OPT] The monitor schedule.' + slicing_exprs: + type: string + description: >- + [Create:OPT Update:OPT] List of column expressions to slice + data with for targeted analysis. The data is grouped by each + expression independently, resulting in a separate slice for + each predicate and its complements. For example + `slicing_exprs=[“col_1”, “col_2 > 10”]` will generate the + following slices: two slices for `col_2 > 10` (True and + False), and one slice per unique value in `col1`. For + high-cardinality columns, only the top 100 unique values by + frequency will generate slices. + snapshot: + type: string + description: Configuration for monitoring snapshot tables. + time_series: + type: string + description: Configuration for monitoring time series tables. + required: + - output_schema_name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MonitorInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + /api/2.1/unity-catalog/tables/{table_name}/monitor/refreshes/{refresh_id}: + get: + operationId: quality_monitors_get_refresh + summary: >- + [DEPRECATED] Gets info about a specific monitor refresh using the given + refresh ID. Use Data Quality + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Gets info about a specific monitor refresh using the given + refresh ID. Use Data Quality + + Monitors API instead (/api/data-quality/v1/monitors). + + + The caller must either: 1. be an owner of the table's parent catalog 2. + have **USE_CATALOG** on the + + table's parent catalog and be an owner of the table's parent schema 3. + have the following permissions: + + - **USE_CATALOG** on the table's parent catalog - **USE_SCHEMA** on the + table's parent schema - + + **SELECT** privilege on the table. + + + Additionally, the call must be made from the workspace where the monitor + was created. + + + :param table_name: str + Full name of the table. + :param refresh_id: int + ID of the refresh. + + :returns: :class:`MonitorRefreshInfo` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: Full name of the table. + - name: refresh_id + in: path + required: true + schema: + type: integer + description: ID of the refresh. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MonitorRefreshInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + /api/2.1/unity-catalog/tables/{table_name}/monitor/refreshes: + get: + operationId: quality_monitors_list_refreshes + summary: >- + [DEPRECATED] Gets an array containing the history of the most recent + refreshes (up to 25) for this + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Gets an array containing the history of the most recent + refreshes (up to 25) for this + + table. Use Data Quality Monitors API instead + (/api/data-quality/v1/monitors). + + + The caller must either: 1. be an owner of the table's parent catalog 2. + have **USE_CATALOG** on the + + table's parent catalog and be an owner of the table's parent schema 3. + have the following permissions: + + - **USE_CATALOG** on the table's parent catalog - **USE_SCHEMA** on the + table's parent schema - + + **SELECT** privilege on the table. + + + Additionally, the call must be made from the workspace where the monitor + was created. + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. table_name is case insensitive and spaces are + disallowed. + + :returns: :class:`MonitorRefreshListResponse` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. table_name is + case insensitive and spaces are disallowed. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MonitorRefreshListResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + post: + operationId: quality_monitors_run_refresh + summary: >- + [DEPRECATED] Queues a metric refresh on the monitor for the specified + table. Use Data Quality Monitors + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Queues a metric refresh on the monitor for the specified + table. Use Data Quality Monitors + + API instead (/api/data-quality/v1/monitors). The refresh will execute in + the background. + + + The caller must either: 1. be an owner of the table's parent catalog 2. + have **USE_CATALOG** on the + + table's parent catalog and be an owner of the table's parent schema 3. + have the following permissions: + + - **USE_CATALOG** on the table's parent catalog - **USE_SCHEMA** on the + table's parent schema - be an + + owner of the table + + + Additionally, the call must be made from the workspace where the monitor + was created. + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. table_name is case insensitive and spaces are + disallowed. + + :returns: :class:`MonitorRefreshInfo` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. table_name is + case insensitive and spaces are disallowed. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MonitorRefreshInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + /api/2.1/quality-monitoring/tables/{table_name}/monitor/dashboard: + post: + operationId: quality_monitors_regenerate_dashboard + summary: >- + [DEPRECATED] Regenerates the monitoring dashboard for the specified + table. Use Data Quality Monitors + tags: + - catalog + - quality_monitors + description: >- + [DEPRECATED] Regenerates the monitoring dashboard for the specified + table. Use Data Quality Monitors + + API instead (/api/data-quality/v1/monitors). + + + The caller must either: 1. be an owner of the table's parent catalog 2. + have **USE_CATALOG** on the + + table's parent catalog and be an owner of the table's parent schema 3. + have the following permissions: + + - **USE_CATALOG** on the table's parent catalog - **USE_SCHEMA** on the + table's parent schema - be an + + owner of the table + + + The call must be made from the workspace where the monitor was created. + The dashboard will be + + regenerated in the assets directory that was specified when the monitor + was created. + + + :param table_name: str + UC table name in format `catalog.schema.table_name`. This field corresponds to the + {full_table_name_arg} arg in the endpoint path. + :param warehouse_id: str (optional) + Optional argument to specify the warehouse for dashboard regeneration. If not specified, the first + running warehouse will be used. + + :returns: :class:`RegenerateDashboardResponse` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: >- + UC table name in format `catalog.schema.table_name`. This field + corresponds to the {full_table_name_arg} arg in the endpoint path. + requestBody: + content: + application/json: + schema: + type: object + properties: + warehouse_id: + type: string + description: >- + Optional argument to specify the warehouse for dashboard + regeneration. If not specified, the first running warehouse + will be used. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegenerateDashboardResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorsAPI + /api/2.1/unity-catalog/models: + post: + operationId: registered_models_create + summary: Creates a new registered model in Unity Catalog. + tags: + - catalog + - registered_models + description: >- + Creates a new registered model in Unity Catalog. + + + File storage for model versions in the registered model will be located + in the default location which + + is specified by the parent schema, or the parent catalog, or the + Metastore. + + + For registered model creation to succeed, the user must satisfy the + following conditions: - The caller + + must be a metastore admin, or be the owner of the parent catalog and + schema, or have the + + **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** + privilege on the parent schema. + + - The caller must have the **CREATE MODEL** or **CREATE FUNCTION** + privilege on the parent schema. + + + :param aliases: List[:class:`RegisteredModelAlias`] (optional) + List of aliases associated with the registered model + :param browse_only: bool (optional) + Indicates whether the principal is limited to retrieving metadata for the associated object through + the BROWSE privilege when include_browse is enabled in the request. + :param catalog_name: str (optional) + The name of the catalog where the schema and the registered model reside + :param comment: str (optional) + The comment attached to the registered model + :param created_at: int (optional) + Creation timestamp of the registered model in milliseconds since the Unix epoch + :param created_by: str (optional) + The identifier of the user who created the registered model + :param full_name: str (optional) + The three-level (fully qualified) name of the registered model + :param metastore_id: str (optional) + The unique identifier of the metastore + :param name: str (optional) + The name of the registered model + :param owner: str (optional) + The identifier of the user who owns the registered model + :param schema_name: str (optional) + The name of the schema where the registered model resides + :param storage_location: str (optional) + The storage location on the cloud under which model version data files are stored + :param updated_at: int (optional) + Last-update timestamp of the registered model in milliseconds since the Unix epoch + :param updated_by: str (optional) + The identifier of the user who updated the registered model last time + + :returns: :class:`RegisteredModelInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + aliases: + type: string + description: List of aliases associated with the registered model + browse_only: + type: string + description: >- + Indicates whether the principal is limited to retrieving + metadata for the associated object through the BROWSE + privilege when include_browse is enabled in the request. + catalog_name: + type: string + description: >- + The name of the catalog where the schema and the registered + model reside + comment: + type: string + description: The comment attached to the registered model + created_at: + type: string + description: >- + Creation timestamp of the registered model in milliseconds + since the Unix epoch + created_by: + type: string + description: The identifier of the user who created the registered model + full_name: + type: string + description: >- + The three-level (fully qualified) name of the registered + model + metastore_id: + type: string + description: The unique identifier of the metastore + name: + type: string + description: The name of the registered model + owner: + type: string + description: The identifier of the user who owns the registered model + schema_name: + type: string + description: The name of the schema where the registered model resides + storage_location: + type: string + description: >- + The storage location on the cloud under which model version + data files are stored + updated_at: + type: string + description: >- + Last-update timestamp of the registered model in + milliseconds since the Unix epoch + updated_by: + type: string + description: >- + The identifier of the user who updated the registered model + last time + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegisteredModelInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RegisteredModelsAPI + get: + operationId: registered_models_list + summary: >- + List registered models. You can list registered models under a + particular schema, or list all + tags: + - catalog + - registered_models + description: >- + List registered models. You can list registered models under a + particular schema, or list all + + registered models in the current metastore. + + + The returned models are filtered based on the privileges of the calling + user. For example, the + + metastore admin is able to list all the registered models. A regular + user needs to be the owner or + + have the **EXECUTE** privilege on the registered model to recieve the + registered models in the + + response. For the latter case, the caller must also be the owner or have + the **USE_CATALOG** privilege + + on the parent catalog and the **USE_SCHEMA** privilege on the parent + schema. + + + There is no guarantee of a specific ordering of the elements in the + response. + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param catalog_name: str (optional) + The identifier of the catalog under which to list registered models. If specified, schema_name must + be specified. + :param include_browse: bool (optional) + Whether to include registered models in the response for which the principal can only access + selective metadata for + :param max_results: int (optional) + Max number of registered models to return. + + If both catalog and schema are specified: - when max_results is not specified, the page length is + set to a server configured value (10000, as of 4/2/2024). - when set to a value greater than 0, the + page length is the minimum of this value and a server configured value (10000, as of 4/2/2024); - + when set to 0, the page length is set to a server configured value (10000, as of 4/2/2024); - when + set to a value less than 0, an invalid parameter error is returned; + + If neither schema nor catalog is specified: - when max_results is not specified, the page length is + set to a server configured value (100, as of 4/2/2024). - when set to a value greater than 0, the + page length is the minimum of this value and a server configured value (1000, as of 4/2/2024); - + when set to 0, the page length is set to a server configured value (100, as of 4/2/2024); - when set + to a value less than 0, an invalid parameter error is returned; + :param page_token: str (optional) + Opaque token to send for the next page of results (pagination). + :param schema_name: str (optional) + The identifier of the schema under which to list registered models. If specified, catalog_name must + be specified. + + :returns: Iterator over :class:`RegisteredModelInfo` + parameters: + - name: catalog_name + in: query + required: false + schema: + type: string + description: >- + The identifier of the catalog under which to list registered models. + If specified, schema_name must be specified. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include registered models in the response for which the + principal can only access selective metadata for + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Max number of registered models to return. If both catalog and + schema are specified: - when max_results is not specified, the page + length is set to a server configured value (10000, as of 4/2/2024). + - when set to a value greater than 0, the page length is the minimum + of this value and a server configured value (10000, as of 4/2/2024); + - when set to 0, the page length is set to a server configured value + (10000, as of 4/2/2024); - when set to a value less than 0, an + invalid parameter error is returned; If neither schema nor catalog + is specified: - when max_results is not specified, the page length + is set to a server configured value (100, as of 4/2/2024). - when + set to a value greater than 0, the page length is the minimum of + this value and a server configured value (1000, as of 4/2/2024); - + when set to 0, the page length is set to a server configured value + (100, as of 4/2/2024); - when set to a value less than 0, an invalid + parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque token to send for the next page of results (pagination). + - name: schema_name + in: query + required: false + schema: + type: string + description: >- + The identifier of the schema under which to list registered models. + If specified, catalog_name must be specified. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListRegisteredModelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RegisteredModelsAPI + /api/2.1/unity-catalog/models/{full_name}: + delete: + operationId: registered_models_delete + summary: >- + Deletes a registered model and all its model versions from the specified + parent catalog and schema. + tags: + - catalog + - registered_models + description: >- + Deletes a registered model and all its model versions from the specified + parent catalog and schema. + + + The caller must be a metastore admin or an owner of the registered + model. For the latter case, the + + caller must also be the owner or have the **USE_CATALOG** privilege on + the parent catalog and the + + **USE_SCHEMA** privilege on the parent schema. + + + :param full_name: str + The three-level (fully qualified) name of the registered model + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the registered model + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RegisteredModelsAPI + get: + operationId: registered_models_get + summary: Get a registered model. + tags: + - catalog + - registered_models + description: >- + Get a registered model. + + + The caller must be a metastore admin or an owner of (or have the + **EXECUTE** privilege on) the + + registered model. For the latter case, the caller must also be the owner + or have the **USE_CATALOG** + + privilege on the parent catalog and the **USE_SCHEMA** privilege on the + parent schema. + + + :param full_name: str + The three-level (fully qualified) name of the registered model + :param include_aliases: bool (optional) + Whether to include registered model aliases in the response + :param include_browse: bool (optional) + Whether to include registered models in the response for which the principal can only access + selective metadata for + + :returns: :class:`RegisteredModelInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the registered model + - name: include_aliases + in: query + required: false + schema: + type: string + description: Whether to include registered model aliases in the response + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include registered models in the response for which the + principal can only access selective metadata for + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegisteredModelInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RegisteredModelsAPI + patch: + operationId: registered_models_update + summary: Updates the specified registered model. + tags: + - catalog + - registered_models + description: >- + Updates the specified registered model. + + + The caller must be a metastore admin or an owner of the registered + model. For the latter case, the + + caller must also be the owner or have the **USE_CATALOG** privilege on + the parent catalog and the + + **USE_SCHEMA** privilege on the parent schema. + + + Currently only the name, the owner or the comment of the registered + model can be updated. + + + :param full_name: str + The three-level (fully qualified) name of the registered model + :param aliases: List[:class:`RegisteredModelAlias`] (optional) + List of aliases associated with the registered model + :param browse_only: bool (optional) + Indicates whether the principal is limited to retrieving metadata for the associated object through + the BROWSE privilege when include_browse is enabled in the request. + :param catalog_name: str (optional) + The name of the catalog where the schema and the registered model reside + :param comment: str (optional) + The comment attached to the registered model + :param created_at: int (optional) + Creation timestamp of the registered model in milliseconds since the Unix epoch + :param created_by: str (optional) + The identifier of the user who created the registered model + :param metastore_id: str (optional) + The unique identifier of the metastore + :param name: str (optional) + The name of the registered model + :param new_name: str (optional) + New name for the registered model. + :param owner: str (optional) + The identifier of the user who owns the registered model + :param schema_name: str (optional) + The name of the schema where the registered model resides + :param storage_location: str (optional) + The storage location on the cloud under which model version data files are stored + :param updated_at: int (optional) + Last-update timestamp of the registered model in milliseconds since the Unix epoch + :param updated_by: str (optional) + The identifier of the user who updated the registered model last time + + :returns: :class:`RegisteredModelInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the registered model + requestBody: + content: + application/json: + schema: + type: object + properties: + aliases: + type: string + description: List of aliases associated with the registered model + browse_only: + type: string + description: >- + Indicates whether the principal is limited to retrieving + metadata for the associated object through the BROWSE + privilege when include_browse is enabled in the request. + catalog_name: + type: string + description: >- + The name of the catalog where the schema and the registered + model reside + comment: + type: string + description: The comment attached to the registered model + created_at: + type: string + description: >- + Creation timestamp of the registered model in milliseconds + since the Unix epoch + created_by: + type: string + description: The identifier of the user who created the registered model + metastore_id: + type: string + description: The unique identifier of the metastore + name: + type: string + description: The name of the registered model + new_name: + type: string + description: New name for the registered model. + owner: + type: string + description: The identifier of the user who owns the registered model + schema_name: + type: string + description: The name of the schema where the registered model resides + storage_location: + type: string + description: >- + The storage location on the cloud under which model version + data files are stored + updated_at: + type: string + description: >- + Last-update timestamp of the registered model in + milliseconds since the Unix epoch + updated_by: + type: string + description: >- + The identifier of the user who updated the registered model + last time + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegisteredModelInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RegisteredModelsAPI + /api/2.1/unity-catalog/resource-quotas/{parent_securable_type}/{parent_full_name}/{quota_name}: + get: + operationId: resource_quotas_get_quota + summary: >- + The GetQuota API returns usage information for a single resource quota, + defined as a child-parent + tags: + - catalog + - resource_quotas + description: >- + The GetQuota API returns usage information for a single resource quota, + defined as a child-parent + + pair. This API also refreshes the quota count if it is out of date. + Refreshes are triggered + + asynchronously. The updated count might not be returned in the first + call. + + + :param parent_securable_type: str + Securable type of the quota parent. + :param parent_full_name: str + Full name of the parent resource. Provide the metastore ID if the parent is a metastore. + :param quota_name: str + Name of the quota. Follows the pattern of the quota type, with "-quota" added as a suffix. + + :returns: :class:`GetQuotaResponse` + parameters: + - name: parent_securable_type + in: path + required: true + schema: + type: string + description: Securable type of the quota parent. + - name: parent_full_name + in: path + required: true + schema: + type: string + description: >- + Full name of the parent resource. Provide the metastore ID if the + parent is a metastore. + - name: quota_name + in: path + required: true + schema: + type: string + description: >- + Name of the quota. Follows the pattern of the quota type, with + "-quota" added as a suffix. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetQuotaResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ResourceQuotasAPI + /api/2.1/unity-catalog/resource-quotas/all-resource-quotas: + get: + operationId: resource_quotas_list_quotas + summary: >- + ListQuotas returns all quota values under the metastore. There are no + SLAs on the freshness of the + tags: + - catalog + - resource_quotas + description: >- + ListQuotas returns all quota values under the metastore. There are no + SLAs on the freshness of the + + counts returned. This API does not trigger a refresh of quota counts. + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param max_results: int (optional) + The number of quotas to return. + :param page_token: str (optional) + Opaque token for the next page of results. + + :returns: Iterator over :class:`QuotaInfo` + parameters: + - name: max_results + in: query + required: false + schema: + type: string + description: The number of quotas to return. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque token for the next page of results. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListQuotasResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ResourceQuotasAPI + /api/3.0/rfa/requests: + post: + operationId: rfa_batch_create_access_requests + summary: >- + Creates access requests for Unity Catalog permissions for a specified + principal on a securable object. + tags: + - catalog + - rfa + description: >- + Creates access requests for Unity Catalog permissions for a specified + principal on a securable object. + + This Batch API can take in multiple principals, securable objects, and + permissions as the input and + + returns the access request destinations for each. Principals must be + unique across the API call. + + + The supported securable types are: "metastore", "catalog", "schema", + "table", "external_location", + + "connection", "credential", "function", "registered_model", and + "volume". + + + :param requests: List[:class:`CreateAccessRequest`] (optional) + A list of individual access requests, where each request corresponds to a set of permissions being + requested on a list of securables for a specified principal. + + At most 30 requests per API call. + + :returns: :class:`BatchCreateAccessRequestsResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + requests: + type: string + description: >- + A list of individual access requests, where each request + corresponds to a set of permissions being requested on a + list of securables for a specified principal. At most 30 + requests per API call. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchCreateAccessRequestsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RfaAPI + /api/3.0/rfa/destinations/{securable_type}/{full_name}: + get: + operationId: rfa_get_access_request_destinations + summary: >- + Gets an array of access request destinations for the specified + securable. Any caller can see URL + tags: + - catalog + - rfa + description: >- + Gets an array of access request destinations for the specified + securable. Any caller can see URL + + destinations or the destinations on the metastore. Otherwise, only those + with **BROWSE** permissions + + on the securable can see destinations. + + + The supported securable types are: "metastore", "catalog", "schema", + "table", "external_location", + + "connection", "credential", "function", "registered_model", and + "volume". + + + :param securable_type: str + The type of the securable. + :param full_name: str + The full name of the securable. + + :returns: :class:`AccessRequestDestinations` + parameters: + - name: securable_type + in: path + required: true + schema: + type: string + description: The type of the securable. + - name: full_name + in: path + required: true + schema: + type: string + description: The full name of the securable. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccessRequestDestinations' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RfaAPI + /api/3.0/rfa/destinations: + patch: + operationId: rfa_update_access_request_destinations + summary: >- + Updates the access request destinations for the given securable. The + caller must be a metastore admin, + tags: + - catalog + - rfa + description: >- + Updates the access request destinations for the given securable. The + caller must be a metastore admin, + + the owner of the securable, or a user that has the **MANAGE** privilege + on the securable in order to + + assign destinations. Destinations cannot be updated for securables + underneath schemas (tables, + + volumes, functions, and models). For these securable types, destinations + are inherited from the parent + + securable. A maximum of 5 emails and 5 external notification + destinations (Slack, Microsoft Teams, and + + Generic Webhook destinations) can be assigned to a securable. If a URL + destination is assigned, no + + other destinations can be set. + + + The supported securable types are: "metastore", "catalog", "schema", + "table", "external_location", + + "connection", "credential", "function", "registered_model", and + "volume". + + + :param access_request_destinations: :class:`AccessRequestDestinations` + The access request destinations to assign to the securable. For each destination, a + **destination_id** and **destination_type** must be defined. + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`AccessRequestDestinations` + parameters: + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + The field mask must be a single string, with multiple fields + separated by commas (no spaces). The field path is relative to the + resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map + fields is not allowed, as only the entire collection field can be + specified. Field names must exactly match the resource field names. + A field mask of `*` indicates full replacement. It’s recommended to + always explicitly list the fields being updated and avoid using `*` + wildcards, as it can lead to unintended results if the API changes + in the future. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_request_destinations: + type: string + description: >- + The access request destinations to assign to the securable. + For each destination, a **destination_id** and + **destination_type** must be defined. + required: + - access_request_destinations + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccessRequestDestinations' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RfaAPI + /api/2.1/unity-catalog/schemas: + post: + operationId: schemas_create + summary: >- + Creates a new schema for catalog in the Metastore. The caller must be a + metastore admin, or have the + tags: + - catalog + - schemas + description: >- + Creates a new schema for catalog in the Metastore. The caller must be a + metastore admin, or have the + + **CREATE_SCHEMA** privilege in the parent catalog. + + + :param name: str + Name of schema, relative to parent catalog. + :param catalog_name: str + Name of parent catalog. + :param comment: str (optional) + User-provided free-form text description. + :param properties: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + :param storage_root: str (optional) + Storage root URL for managed tables within schema. + + :returns: :class:`SchemaInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of schema, relative to parent catalog. + catalog_name: + type: string + description: Name of parent catalog. + comment: + type: string + description: User-provided free-form text description. + properties: + type: string + description: A map of key-value properties attached to the securable. + storage_root: + type: string + description: Storage root URL for managed tables within schema. + required: + - name + - catalog_name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SchemaInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SchemasAPI + get: + operationId: schemas_list + summary: >- + Gets an array of schemas for a catalog in the metastore. If the caller + is the metastore admin or the + tags: + - catalog + - schemas + description: >- + Gets an array of schemas for a catalog in the metastore. If the caller + is the metastore admin or the + + owner of the parent catalog, all schemas for the catalog will be + retrieved. Otherwise, only schemas + + owned by the caller (or for which the caller has the **USE_SCHEMA** + privilege) will be retrieved. + + There is no guarantee of a specific ordering of the elements in the + array. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param catalog_name: str + Parent catalog for schemas of interest. + :param include_browse: bool (optional) + Whether to include schemas in the response for which the principal can only access selective + metadata for + :param max_results: int (optional) + Maximum number of schemas to return. If not set, all the schemas are returned (not recommended). - + when set to a value greater than 0, the page length is the minimum of this value and a server + configured value; - when set to 0, the page length is set to a server configured value + (recommended); - when set to a value less than 0, an invalid parameter error is returned; + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`SchemaInfo` + parameters: + - name: catalog_name + in: query + required: true + schema: + type: string + description: Parent catalog for schemas of interest. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include schemas in the response for which the principal + can only access selective metadata for + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of schemas to return. If not set, all the schemas are + returned (not recommended). - when set to a value greater than 0, + the page length is the minimum of this value and a server configured + value; - when set to 0, the page length is set to a server + configured value (recommended); - when set to a value less than 0, + an invalid parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListSchemasResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SchemasAPI + /api/2.1/unity-catalog/schemas/{full_name}: + delete: + operationId: schemas_delete + summary: >- + Deletes the specified schema from the parent catalog. The caller must be + the owner of the schema or an + tags: + - catalog + - schemas + description: >- + Deletes the specified schema from the parent catalog. The caller must be + the owner of the schema or an + + owner of the parent catalog. + + + :param full_name: str + Full name of the schema. + :param force: bool (optional) + Force deletion even if the schema is not empty. + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the schema. + - name: force + in: query + required: false + schema: + type: string + description: Force deletion even if the schema is not empty. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SchemasAPI + get: + operationId: schemas_get + summary: >- + Gets the specified schema within the metastore. The caller must be a + metastore admin, the owner of the + tags: + - catalog + - schemas + description: >- + Gets the specified schema within the metastore. The caller must be a + metastore admin, the owner of the + + schema, or a user that has the **USE_SCHEMA** privilege on the schema. + + + :param full_name: str + Full name of the schema. + :param include_browse: bool (optional) + Whether to include schemas in the response for which the principal can only access selective + metadata for + + :returns: :class:`SchemaInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the schema. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include schemas in the response for which the principal + can only access selective metadata for + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SchemaInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SchemasAPI + patch: + operationId: schemas_update + summary: >- + Updates a schema for a catalog. The caller must be the owner of the + schema or a metastore admin. If + tags: + - catalog + - schemas + description: >- + Updates a schema for a catalog. The caller must be the owner of the + schema or a metastore admin. If + + the caller is a metastore admin, only the __owner__ field can be changed + in the update. If the + + __name__ field must be updated, the caller must be a metastore admin or + have the **CREATE_SCHEMA** + + privilege on the parent catalog. + + + :param full_name: str + Full name of the schema. + :param comment: str (optional) + User-provided free-form text description. + :param enable_predictive_optimization: + :class:`EnablePredictiveOptimization` (optional) + Whether predictive optimization should be enabled for this object and objects under it. + :param new_name: str (optional) + New name for the schema. + :param owner: str (optional) + Username of current owner of schema. + :param properties: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + + :returns: :class:`SchemaInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the schema. + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: User-provided free-form text description. + enable_predictive_optimization: + type: string + description: >- + Whether predictive optimization should be enabled for this + object and objects under it. + new_name: + type: string + description: New name for the schema. + owner: + type: string + description: Username of current owner of schema. + properties: + type: string + description: A map of key-value properties attached to the securable. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SchemaInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SchemasAPI + /api/2.1/unity-catalog/storage-credentials: + post: + operationId: storage_credentials_create + summary: Creates a new storage credential. + tags: + - catalog + - storage_credentials + description: >- + Creates a new storage credential. + + + The caller must be a metastore admin or have the + **CREATE_STORAGE_CREDENTIAL** privilege on the + + metastore. + + + :param name: str + The credential name. The name must be unique among storage and service credentials within the + metastore. + :param aws_iam_role: :class:`AwsIamRoleRequest` (optional) + The AWS IAM role configuration. + :param azure_managed_identity: :class:`AzureManagedIdentityRequest` + (optional) + The Azure managed identity configuration. + :param azure_service_principal: :class:`AzureServicePrincipal` + (optional) + The Azure service principal configuration. + :param cloudflare_api_token: :class:`CloudflareApiToken` (optional) + The Cloudflare API token configuration. + :param comment: str (optional) + Comment associated with the credential. + :param databricks_gcp_service_account: + :class:`DatabricksGcpServiceAccountRequest` (optional) + The Databricks managed GCP service account configuration. + :param read_only: bool (optional) + Whether the credential is usable only for read operations. Only applicable when purpose is + **STORAGE**. + :param skip_validation: bool (optional) + Supplying true to this argument skips validation of the created credential. + + :returns: :class:`StorageCredentialInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: >- + The credential name. The name must be unique among storage + and service credentials within the metastore. + aws_iam_role: + type: string + description: The AWS IAM role configuration. + azure_managed_identity: + type: string + description: The Azure managed identity configuration. + azure_service_principal: + type: string + description: The Azure service principal configuration. + cloudflare_api_token: + type: string + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + type: string + description: The Databricks managed GCP service account configuration. + read_only: + type: string + description: >- + Whether the credential is usable only for read operations. + Only applicable when purpose is **STORAGE**. + skip_validation: + type: string + description: >- + Supplying true to this argument skips validation of the + created credential. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StorageCredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageCredentialsAPI + get: + operationId: storage_credentials_list + summary: >- + Gets an array of storage credentials (as __StorageCredentialInfo__ + objects). The array is limited to + tags: + - catalog + - storage_credentials + description: >- + Gets an array of storage credentials (as __StorageCredentialInfo__ + objects). The array is limited to + + only those storage credentials the caller has permission to access. If + the caller is a metastore + + admin, retrieval of credentials is unrestricted. There is no guarantee + of a specific ordering of the + + elements in the array. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param include_unbound: bool (optional) + Whether to include credentials not bound to the workspace. Effective only if the user has permission + to update the credential–workspace binding. + :param max_results: int (optional) + Maximum number of storage credentials to return. If not set, all the storage credentials are + returned (not recommended). - when set to a value greater than 0, the page length is the minimum of + this value and a server configured value; - when set to 0, the page length is set to a server + configured value (recommended); - when set to a value less than 0, an invalid parameter error is + returned; + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`StorageCredentialInfo` + parameters: + - name: include_unbound + in: query + required: false + schema: + type: string + description: >- + Whether to include credentials not bound to the workspace. Effective + only if the user has permission to update the credential–workspace + binding. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of storage credentials to return. If not set, all the + storage credentials are returned (not recommended). - when set to a + value greater than 0, the page length is the minimum of this value + and a server configured value; - when set to 0, the page length is + set to a server configured value (recommended); - when set to a + value less than 0, an invalid parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountStorageCredentialsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageCredentialsAPI + /api/2.1/unity-catalog/storage-credentials/{name}: + delete: + operationId: storage_credentials_delete + summary: >- + Deletes a storage credential from the metastore. The caller must be an + owner of the storage + tags: + - catalog + - storage_credentials + description: >- + Deletes a storage credential from the metastore. The caller must be an + owner of the storage + + credential. + + + :param name: str + Name of the storage credential. + :param force: bool (optional) + Force an update even if there are dependent external locations or external tables (when purpose is + **STORAGE**) or dependent services (when purpose is **SERVICE**). + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the storage credential. + - name: force + in: query + required: false + schema: + type: string + description: >- + Force an update even if there are dependent external locations or + external tables (when purpose is **STORAGE**) or dependent services + (when purpose is **SERVICE**). + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageCredentialsAPI + get: + operationId: storage_credentials_get + summary: >- + Gets a storage credential from the metastore. The caller must be a + metastore admin, the owner of the + tags: + - catalog + - storage_credentials + description: >- + Gets a storage credential from the metastore. The caller must be a + metastore admin, the owner of the + + storage credential, or have some permission on the storage credential. + + + :param name: str + Name of the storage credential. + + :returns: :class:`StorageCredentialInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the storage credential. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StorageCredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageCredentialsAPI + patch: + operationId: storage_credentials_update + summary: Updates a storage credential on the metastore. + tags: + - catalog + - storage_credentials + description: >- + Updates a storage credential on the metastore. + + + The caller must be the owner of the storage credential or a metastore + admin. If the caller is a + + metastore admin, only the **owner** field can be changed. + + + :param name: str + Name of the storage credential. + :param aws_iam_role: :class:`AwsIamRoleRequest` (optional) + The AWS IAM role configuration. + :param azure_managed_identity: :class:`AzureManagedIdentityResponse` + (optional) + The Azure managed identity configuration. + :param azure_service_principal: :class:`AzureServicePrincipal` + (optional) + The Azure service principal configuration. + :param cloudflare_api_token: :class:`CloudflareApiToken` (optional) + The Cloudflare API token configuration. + :param comment: str (optional) + Comment associated with the credential. + :param databricks_gcp_service_account: + :class:`DatabricksGcpServiceAccountRequest` (optional) + The Databricks managed GCP service account configuration. + :param force: bool (optional) + Force update even if there are dependent external locations or external tables. + :param isolation_mode: :class:`IsolationMode` (optional) + Whether the current securable is accessible from all workspaces or a specific set of workspaces. + :param new_name: str (optional) + New name for the storage credential. + :param owner: str (optional) + Username of current owner of credential. + :param read_only: bool (optional) + Whether the credential is usable only for read operations. Only applicable when purpose is + **STORAGE**. + :param skip_validation: bool (optional) + Supplying true to this argument skips validation of the updated credential. + + :returns: :class:`StorageCredentialInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the storage credential. + requestBody: + content: + application/json: + schema: + type: object + properties: + aws_iam_role: + type: string + description: The AWS IAM role configuration. + azure_managed_identity: + type: string + description: The Azure managed identity configuration. + azure_service_principal: + type: string + description: The Azure service principal configuration. + cloudflare_api_token: + type: string + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + type: string + description: The Databricks managed GCP service account configuration. + force: + type: string + description: >- + Force update even if there are dependent external locations + or external tables. + isolation_mode: + type: string + description: >- + Whether the current securable is accessible from all + workspaces or a specific set of workspaces. + new_name: + type: string + description: New name for the storage credential. + owner: + type: string + description: Username of current owner of credential. + read_only: + type: string + description: >- + Whether the credential is usable only for read operations. + Only applicable when purpose is **STORAGE**. + skip_validation: + type: string + description: >- + Supplying true to this argument skips validation of the + updated credential. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StorageCredentialInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageCredentialsAPI + /api/2.1/unity-catalog/validate-storage-credentials: + post: + operationId: storage_credentials_validate + summary: >- + Validates a storage credential. At least one of + __external_location_name__ and __url__ need to be + tags: + - catalog + - storage_credentials + description: >- + Validates a storage credential. At least one of + __external_location_name__ and __url__ need to be + + provided. If only one of them is provided, it will be used for + validation. And if both are provided, + + the __url__ will be used for validation, and __external_location_name__ + will be ignored when checking + + overlapping urls. + + + Either the __storage_credential_name__ or the cloud-specific credential + must be provided. + + + The caller must be a metastore admin or the storage credential owner or + have the + + **CREATE_EXTERNAL_LOCATION** privilege on the metastore and the storage + credential. + + + :param aws_iam_role: :class:`AwsIamRoleRequest` (optional) + The AWS IAM role configuration. + :param azure_managed_identity: :class:`AzureManagedIdentityRequest` + (optional) + The Azure managed identity configuration. + :param azure_service_principal: :class:`AzureServicePrincipal` + (optional) + The Azure service principal configuration. + :param cloudflare_api_token: :class:`CloudflareApiToken` (optional) + The Cloudflare API token configuration. + :param databricks_gcp_service_account: + :class:`DatabricksGcpServiceAccountRequest` (optional) + The Databricks created GCP service account configuration. + :param external_location_name: str (optional) + The name of an existing external location to validate. + :param read_only: bool (optional) + Whether the storage credential is only usable for read operations. + :param storage_credential_name: str (optional) + Required. The name of an existing credential or long-lived cloud credential to validate. + :param url: str (optional) + The external location url to validate. + + :returns: :class:`ValidateStorageCredentialResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + aws_iam_role: + type: string + description: The AWS IAM role configuration. + azure_managed_identity: + type: string + description: The Azure managed identity configuration. + azure_service_principal: + type: string + description: The Azure service principal configuration. + cloudflare_api_token: + type: string + description: The Cloudflare API token configuration. + databricks_gcp_service_account: + type: string + description: The Databricks created GCP service account configuration. + external_location_name: + type: string + description: The name of an existing external location to validate. + read_only: + type: string + description: >- + Whether the storage credential is only usable for read + operations. + storage_credential_name: + type: string + description: >- + Required. The name of an existing credential or long-lived + cloud credential to validate. + url: + type: string + description: The external location url to validate. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ValidateStorageCredentialResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StorageCredentialsAPI + /api/2.1/unity-catalog/metastores/{metastore_id}/systemschemas/{schema_name}: + delete: + operationId: system_schemas_disable + summary: >- + Disables the system schema and removes it from the system catalog. The + caller must be an account admin + tags: + - catalog + - system_schemas + description: >- + Disables the system schema and removes it from the system catalog. The + caller must be an account admin + + or a metastore admin. + + + :param metastore_id: str + The metastore ID under which the system schema lives. + :param schema_name: str + Full name of the system schema. + parameters: + - name: metastore_id + in: path + required: true + schema: + type: string + description: The metastore ID under which the system schema lives. + - name: schema_name + in: path + required: true + schema: + type: string + description: Full name of the system schema. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SystemSchemasAPI + put: + operationId: system_schemas_enable + summary: >- + Enables the system schema and adds it to the system catalog. The caller + must be an account admin or a + tags: + - catalog + - system_schemas + description: >- + Enables the system schema and adds it to the system catalog. The caller + must be an account admin or a + + metastore admin. + + + :param metastore_id: str + The metastore ID under which the system schema lives. + :param schema_name: str + Full name of the system schema. + :param catalog_name: str (optional) + the catalog for which the system schema is to enabled in + parameters: + - name: metastore_id + in: path + required: true + schema: + type: string + description: The metastore ID under which the system schema lives. + - name: schema_name + in: path + required: true + schema: + type: string + description: Full name of the system schema. + requestBody: + content: + application/json: + schema: + type: object + properties: + catalog_name: + type: string + description: the catalog for which the system schema is to enabled in + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SystemSchemasAPI + /api/2.1/unity-catalog/metastores/{metastore_id}/systemschemas: + get: + operationId: system_schemas_list + summary: >- + Gets an array of system schemas for a metastore. The caller must be an + account admin or a metastore + tags: + - catalog + - system_schemas + description: >- + Gets an array of system schemas for a metastore. The caller must be an + account admin or a metastore + + admin. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param metastore_id: str + The ID for the metastore in which the system schema resides. + :param max_results: int (optional) + Maximum number of schemas to return. - When set to 0, the page length is set to a server configured + value (recommended); - When set to a value greater than 0, the page length is the minimum of this + value and a server configured value; - When set to a value less than 0, an invalid parameter error + is returned; - If not set, all the schemas are returned (not recommended). + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`SystemSchemaInfo` + parameters: + - name: metastore_id + in: path + required: true + schema: + type: string + description: The ID for the metastore in which the system schema resides. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of schemas to return. - When set to 0, the page + length is set to a server configured value (recommended); - When set + to a value greater than 0, the page length is the minimum of this + value and a server configured value; - When set to a value less than + 0, an invalid parameter error is returned; - If not set, all the + schemas are returned (not recommended). + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListSystemSchemasResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SystemSchemasAPI + /api/2.1/unity-catalog/constraints: + post: + operationId: table_constraints_create + summary: Creates a new table constraint. + tags: + - catalog + - table_constraints + description: >- + Creates a new table constraint. + + + For the table constraint creation to succeed, the user must satisfy both + of these conditions: - the + + user must have the **USE_CATALOG** privilege on the table's parent + catalog, the **USE_SCHEMA** + + privilege on the table's parent schema, and be the owner of the table. - + if the new constraint is a + + __ForeignKeyConstraint__, the user must have the **USE_CATALOG** + privilege on the referenced parent + + table's catalog, the **USE_SCHEMA** privilege on the referenced parent + table's schema, and be the + + owner of the referenced parent table. + + + :param full_name_arg: str + The full name of the table referenced by the constraint. + :param constraint: :class:`TableConstraint` + + + :returns: :class:`TableConstraint` + requestBody: + content: + application/json: + schema: + type: object + properties: + full_name_arg: + type: string + description: The full name of the table referenced by the constraint. + constraint: + type: string + description: ':returns: :class:`TableConstraint`' + required: + - full_name_arg + - constraint + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TableConstraint' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TableConstraintsAPI + /api/2.1/unity-catalog/constraints/{full_name}: + delete: + operationId: table_constraints_delete + summary: Deletes a table constraint. + tags: + - catalog + - table_constraints + description: >- + Deletes a table constraint. + + + For the table constraint deletion to succeed, the user must satisfy both + of these conditions: - the + + user must have the **USE_CATALOG** privilege on the table's parent + catalog, the **USE_SCHEMA** + + privilege on the table's parent schema, and be the owner of the table. - + if __cascade__ argument is + + **true**, the user must have the following permissions on all of the + child tables: the **USE_CATALOG** + + privilege on the table's catalog, the **USE_SCHEMA** privilege on the + table's schema, and be the owner + + of the table. + + + :param full_name: str + Full name of the table referenced by the constraint. + :param constraint_name: str + The name of the constraint to delete. + :param cascade: bool + If true, try deleting all child constraints of the current constraint. If false, reject this + operation if the current constraint has any child constraints. + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the table referenced by the constraint. + - name: constraint_name + in: query + required: true + schema: + type: string + description: The name of the constraint to delete. + - name: cascade + in: query + required: true + schema: + type: boolean + description: >- + If true, try deleting all child constraints of the current + constraint. If false, reject this operation if the current + constraint has any child constraints. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TableConstraintsAPI + /api/2.1/unity-catalog/tables: + post: + operationId: tables_create + summary: Creates a new table in the specified catalog and schema. + tags: + - catalog + - tables + description: >- + Creates a new table in the specified catalog and schema. + + + To create an external delta table, the caller must have the + **EXTERNAL_USE_SCHEMA** privilege on the + + parent schema and the **EXTERNAL_USE_LOCATION** privilege on the + external location. These privileges + + must always be granted explicitly, and cannot be inherited through + ownership or **ALL_PRIVILEGES**. + + + Standard UC permissions needed to create tables still apply: + **USE_CATALOG** on the parent catalog (or + + ownership of the parent catalog), **CREATE_TABLE** and **USE_SCHEMA** on + the parent schema (or + + ownership of the parent schema), and **CREATE_EXTERNAL_TABLE** on + external location. + + + The **columns** field needs to be in a Spark compatible format, so we + recommend you use Spark to + + create these tables. The API itself does not validate the correctness of + the column spec. If the spec + + is not Spark compatible, the tables may not be readable by Databricks + Runtime. + + + NOTE: The Create Table API for external clients only supports creating + **external delta tables**. The + + values shown in the respective enums are all values supported by + Databricks, however for this specific + + Create Table API, only **table_type** **EXTERNAL** and + **data_source_format** **DELTA** are supported. + + Additionally, column masks are not supported when creating tables + through this API. + + + :param name: str + Name of table, relative to parent schema. + :param catalog_name: str + Name of parent catalog. + :param schema_name: str + Name of parent schema relative to its parent catalog. + :param table_type: :class:`TableType` + + :param data_source_format: :class:`DataSourceFormat` + + :param storage_location: str + Storage root URL for table (for **MANAGED**, **EXTERNAL** tables). + :param columns: List[:class:`ColumnInfo`] (optional) + The array of __ColumnInfo__ definitions of the table's columns. + :param properties: Dict[str,str] (optional) + A map of key-value properties attached to the securable. + + :returns: :class:`TableInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of table, relative to parent schema. + catalog_name: + type: string + description: Name of parent catalog. + schema_name: + type: string + description: Name of parent schema relative to its parent catalog. + table_type: + type: string + description: ':param data_source_format: :class:`DataSourceFormat`' + data_source_format: + type: string + storage_location: + type: string + description: >- + Storage root URL for table (for **MANAGED**, **EXTERNAL** + tables). + columns: + type: string + description: >- + The array of __ColumnInfo__ definitions of the table's + columns. + properties: + type: string + description: A map of key-value properties attached to the securable. + required: + - name + - catalog_name + - schema_name + - table_type + - data_source_format + - storage_location + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TableInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TablesAPI + get: + operationId: tables_list + summary: >- + Gets an array of all tables for the current metastore under the parent + catalog and schema. The caller + tags: + - catalog + - tables + description: >- + Gets an array of all tables for the current metastore under the parent + catalog and schema. The caller + + must be a metastore admin or an owner of (or have the **SELECT** + privilege on) the table. For the + + latter case, the caller must also be the owner or have the + **USE_CATALOG** privilege on the parent + + catalog and the **USE_SCHEMA** privilege on the parent schema. There is + no guarantee of a specific + + ordering of the elements in the array. + + + NOTE: **view_dependencies** and **table_constraints** are not returned + by ListTables queries. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param catalog_name: str + Name of parent catalog for tables of interest. + :param schema_name: str + Parent schema of tables. + :param include_browse: bool (optional) + Whether to include tables in the response for which the principal can only access selective metadata + for. + :param include_manifest_capabilities: bool (optional) + Whether to include a manifest containing table capabilities in the response. + :param max_results: int (optional) + Maximum number of tables to return. If not set, all the tables are returned (not recommended). - + when set to a value greater than 0, the page length is the minimum of this value and a server + configured value; - when set to 0, the page length is set to a server configured value + (recommended); - when set to a value less than 0, an invalid parameter error is returned; + :param omit_columns: bool (optional) + Whether to omit the columns of the table from the response or not. + :param omit_properties: bool (optional) + Whether to omit the properties of the table from the response or not. + :param omit_username: bool (optional) + Whether to omit the username of the table (e.g. owner, updated_by, created_by) from the response or + not. + :param page_token: str (optional) + Opaque token to send for the next page of results (pagination). + + :returns: Iterator over :class:`TableInfo` + parameters: + - name: catalog_name + in: query + required: true + schema: + type: string + description: Name of parent catalog for tables of interest. + - name: schema_name + in: query + required: true + schema: + type: string + description: Parent schema of tables. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include tables in the response for which the principal + can only access selective metadata for. + - name: include_manifest_capabilities + in: query + required: false + schema: + type: string + description: >- + Whether to include a manifest containing table capabilities in the + response. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of tables to return. If not set, all the tables are + returned (not recommended). - when set to a value greater than 0, + the page length is the minimum of this value and a server configured + value; - when set to 0, the page length is set to a server + configured value (recommended); - when set to a value less than 0, + an invalid parameter error is returned; + - name: omit_columns + in: query + required: false + schema: + type: string + description: Whether to omit the columns of the table from the response or not. + - name: omit_properties + in: query + required: false + schema: + type: string + description: >- + Whether to omit the properties of the table from the response or + not. + - name: omit_username + in: query + required: false + schema: + type: string + description: >- + Whether to omit the username of the table (e.g. owner, updated_by, + created_by) from the response or not. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque token to send for the next page of results (pagination). + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListTablesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TablesAPI + /api/2.1/unity-catalog/tables/{full_name}: + delete: + operationId: tables_delete + summary: >- + Deletes a table from the specified parent catalog and schema. The caller + must be the owner of the + tags: + - catalog + - tables + description: >- + Deletes a table from the specified parent catalog and schema. The caller + must be the owner of the + + parent catalog, have the **USE_CATALOG** privilege on the parent catalog + and be the owner of the + + parent schema, or be the owner of the table and have the **USE_CATALOG** + privilege on the parent + + catalog and the **USE_SCHEMA** privilege on the parent schema. + + + :param full_name: str + Full name of the table. + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the table. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TablesAPI + get: + operationId: tables_get + summary: >- + Gets a table from the metastore for a specific catalog and schema. The + caller must satisfy one of the + tags: + - catalog + - tables + description: >- + Gets a table from the metastore for a specific catalog and schema. The + caller must satisfy one of the + + following requirements: * Be a metastore admin * Be the owner of the + parent catalog * Be the owner of + + the parent schema and have the **USE_CATALOG** privilege on the parent + catalog * Have the + + **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** + privilege on the parent schema, + + and either be the table owner or have the **SELECT** privilege on the + table. + + + :param full_name: str + Full name of the table. + :param include_browse: bool (optional) + Whether to include tables in the response for which the principal can only access selective metadata + for. + :param include_delta_metadata: bool (optional) + Whether delta metadata should be included in the response. + :param include_manifest_capabilities: bool (optional) + Whether to include a manifest containing table capabilities in the response. + + :returns: :class:`TableInfo` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the table. + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include tables in the response for which the principal + can only access selective metadata for. + - name: include_delta_metadata + in: query + required: false + schema: + type: string + description: Whether delta metadata should be included in the response. + - name: include_manifest_capabilities + in: query + required: false + schema: + type: string + description: >- + Whether to include a manifest containing table capabilities in the + response. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TableInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TablesAPI + patch: + operationId: tables_update + summary: >- + Change the owner of the table. The caller must be the owner of the + parent catalog, have the + tags: + - catalog + - tables + description: >- + Change the owner of the table. The caller must be the owner of the + parent catalog, have the + + **USE_CATALOG** privilege on the parent catalog and be the owner of the + parent schema, or be the owner + + of the table and have the **USE_CATALOG** privilege on the parent + catalog and the **USE_SCHEMA** + + privilege on the parent schema. + + + :param full_name: str + Full name of the table. + :param owner: str (optional) + Username of current owner of table. + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the table. + requestBody: + content: + application/json: + schema: + type: object + properties: + owner: + type: string + description: Username of current owner of table. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TablesAPI + /api/2.1/unity-catalog/tables/{full_name}/exists: + get: + operationId: tables_exists + summary: >- + Gets if a table exists in the metastore for a specific catalog and + schema. The caller must satisfy one + tags: + - catalog + - tables + description: >- + Gets if a table exists in the metastore for a specific catalog and + schema. The caller must satisfy one + + of the following requirements: * Be a metastore admin * Be the owner of + the parent catalog * Be the + + owner of the parent schema and have the **USE_CATALOG** privilege on the + parent catalog * Have the + + **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** + privilege on the parent schema, + + and either be the table owner or have the **SELECT** privilege on the + table. * Have **BROWSE** + + privilege on the parent catalog * Have **BROWSE** privilege on the + parent schema + + + :param full_name: str + Full name of the table. + + :returns: :class:`TableExistsResponse` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Full name of the table. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TableExistsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TablesAPI + /api/2.1/unity-catalog/table-summaries: + get: + operationId: tables_list_summaries + summary: >- + Gets an array of summaries for tables for a schema and catalog within + the metastore. The table + tags: + - catalog + - tables + description: >- + Gets an array of summaries for tables for a schema and catalog within + the metastore. The table + + summaries returned are either: + + + * summaries for tables (within the current metastore and parent catalog + and schema), when the user is + + a metastore admin, or: * summaries for tables and schemas (within the + current metastore and parent + + catalog) for which the user has ownership or the **SELECT** privilege on + the table and ownership or + + **USE_SCHEMA** privilege on the schema, provided that the user also has + ownership or the + + **USE_CATALOG** privilege on the parent catalog. + + + There is no guarantee of a specific ordering of the elements in the + array. + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param catalog_name: str + Name of parent catalog for tables of interest. + :param include_manifest_capabilities: bool (optional) + Whether to include a manifest containing table capabilities in the response. + :param max_results: int (optional) + Maximum number of summaries for tables to return. If not set, the page length is set to a server + configured value (10000, as of 1/5/2024). - when set to a value greater than 0, the page length is + the minimum of this value and a server configured value (10000, as of 1/5/2024); - when set to 0, + the page length is set to a server configured value (10000, as of 1/5/2024) (recommended); - when + set to a value less than 0, an invalid parameter error is returned; + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + :param schema_name_pattern: str (optional) + A sql LIKE pattern (% and _) for schema names. All schemas will be returned if not set or empty. + :param table_name_pattern: str (optional) + A sql LIKE pattern (% and _) for table names. All tables will be returned if not set or empty. + + :returns: Iterator over :class:`TableSummary` + parameters: + - name: catalog_name + in: query + required: true + schema: + type: string + description: Name of parent catalog for tables of interest. + - name: include_manifest_capabilities + in: query + required: false + schema: + type: string + description: >- + Whether to include a manifest containing table capabilities in the + response. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of summaries for tables to return. If not set, the + page length is set to a server configured value (10000, as of + 1/5/2024). - when set to a value greater than 0, the page length is + the minimum of this value and a server configured value (10000, as + of 1/5/2024); - when set to 0, the page length is set to a server + configured value (10000, as of 1/5/2024) (recommended); - when set + to a value less than 0, an invalid parameter error is returned; + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + - name: schema_name_pattern + in: query + required: false + schema: + type: string + description: >- + A sql LIKE pattern (% and _) for schema names. All schemas will be + returned if not set or empty. + - name: table_name_pattern + in: query + required: false + schema: + type: string + description: >- + A sql LIKE pattern (% and _) for table names. All tables will be + returned if not set or empty. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListTableSummariesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TablesAPI + /api/2.0/unity-catalog/temporary-path-credentials: + post: + operationId: temporary_path_credentials_generate_temporary_path_credentials + summary: >- + Get a short-lived credential for directly accessing cloud storage + locations registered in Databricks. + tags: + - catalog + - temporary_path_credentials + description: >- + Get a short-lived credential for directly accessing cloud storage + locations registered in Databricks. + + The Generate Temporary Path Credentials API is only supported for + external storage paths, specifically + + external locations and external tables. Managed tables are not supported + by this API. The metastore + + must have **external_access_enabled** flag set to true (default false). + The caller must have the + + **EXTERNAL_USE_LOCATION** privilege on the external location; this + privilege can only be granted by + + external location owners. For requests on existing external tables, the + caller must also have the + + **EXTERNAL_USE_SCHEMA** privilege on the parent schema; this privilege + can only be granted by catalog + + owners. + + + :param url: str + URL for path-based access. + :param operation: :class:`PathOperation` + The operation being performed on the path. + :param dry_run: bool (optional) + Optional. When set to true, the service will not validate that the generated credentials can perform + write operations, therefore no new paths will be created and the response will not contain valid + credentials. Defaults to false. + + :returns: :class:`GenerateTemporaryPathCredentialResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + url: + type: string + description: URL for path-based access. + operation: + type: string + description: The operation being performed on the path. + dry_run: + type: string + description: >- + Optional. When set to true, the service will not validate + that the generated credentials can perform write operations, + therefore no new paths will be created and the response will + not contain valid credentials. Defaults to false. + required: + - url + - operation + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenerateTemporaryPathCredentialResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TemporaryPathCredentialsAPI + /api/2.0/unity-catalog/temporary-table-credentials: + post: + operationId: temporary_table_credentials_generate_temporary_table_credentials + summary: >- + Get a short-lived credential for directly accessing the table data on + cloud storage. The metastore + tags: + - catalog + - temporary_table_credentials + description: >- + Get a short-lived credential for directly accessing the table data on + cloud storage. The metastore + + must have **external_access_enabled** flag set to true (default false). + The caller must have the + + **EXTERNAL_USE_SCHEMA** privilege on the parent schema and this + privilege can only be granted by + + catalog owners. + + + :param operation: :class:`TableOperation` (optional) + The operation performed against the table data, either READ or READ_WRITE. If READ_WRITE is + specified, the credentials returned will have write permissions, otherwise, it will be read only. + :param table_id: str (optional) + UUID of the table to read or write. + + :returns: :class:`GenerateTemporaryTableCredentialResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + operation: + type: string + description: >- + The operation performed against the table data, either READ + or READ_WRITE. If READ_WRITE is specified, the credentials + returned will have write permissions, otherwise, it will be + read only. + table_id: + type: string + description: UUID of the table to read or write. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenerateTemporaryTableCredentialResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TemporaryTableCredentialsAPI + /api/2.1/unity-catalog/volumes: + post: + operationId: volumes_create + summary: Creates a new volume. + tags: + - catalog + - volumes + description: >- + Creates a new volume. + + + The user could create either an external volume or a managed volume. An + external volume will be + + created in the specified external location, while a managed volume will + be located in the default + + location which is specified by the parent schema, or the parent catalog, + or the Metastore. + + + For the volume creation to succeed, the user must satisfy following + conditions: - The caller must be a + + metastore admin, or be the owner of the parent catalog and schema, or + have the **USE_CATALOG** + + privilege on the parent catalog and the **USE_SCHEMA** privilege on the + parent schema. - The caller + + must have **CREATE VOLUME** privilege on the parent schema. + + + For an external volume, following conditions also need to satisfy - The + caller must have **CREATE + + EXTERNAL VOLUME** privilege on the external location. - There are no + other tables, nor volumes + + existing in the specified storage location. - The specified storage + location is not under the location + + of other tables, nor volumes, or catalogs or schemas. + + + :param catalog_name: str + The name of the catalog where the schema and the volume are + :param schema_name: str + The name of the schema where the volume is + :param name: str + The name of the volume + :param volume_type: :class:`VolumeType` + The type of the volume. An external volume is located in the specified external location. A managed + volume is located in the default location which is specified by the parent schema, or the parent + catalog, or the Metastore. [Learn more] + + [Learn more]: https://docs.databricks.com/aws/en/volumes/managed-vs-external + :param comment: str (optional) + The comment attached to the volume + :param storage_location: str (optional) + The storage location on the cloud + + :returns: :class:`VolumeInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + catalog_name: + type: string + description: The name of the catalog where the schema and the volume are + schema_name: + type: string + description: The name of the schema where the volume is + name: + type: string + description: The name of the volume + volume_type: + type: string + description: >- + The type of the volume. An external volume is located in the + specified external location. A managed volume is located in + the default location which is specified by the parent + schema, or the parent catalog, or the Metastore. [Learn + more] [Learn more]: + https://docs.databricks.com/aws/en/volumes/managed-vs-external + comment: + type: string + description: The comment attached to the volume + storage_location: + type: string + description: The storage location on the cloud + required: + - catalog_name + - schema_name + - name + - volume_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VolumeInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VolumesAPI + get: + operationId: volumes_list + summary: >- + Gets an array of volumes for the current metastore under the parent + catalog and schema. + tags: + - catalog + - volumes + description: >- + Gets an array of volumes for the current metastore under the parent + catalog and schema. + + + The returned volumes are filtered based on the privileges of the calling + user. For example, the + + metastore admin is able to list all the volumes. A regular user needs to + be the owner or have the + + **READ VOLUME** privilege on the volume to receive the volumes in the + response. For the latter case, + + the caller must also be the owner or have the **USE_CATALOG** privilege + on the parent catalog and the + + **USE_SCHEMA** privilege on the parent schema. + + + There is no guarantee of a specific ordering of the elements in the + array. + + + PAGINATION BEHAVIOR: The API is by default paginated, a page may contain + zero results while still + + providing a next_page_token. Clients must continue reading pages until + next_page_token is absent, + + which is the only indication that the end of results has been reached. + + + :param catalog_name: str + The identifier of the catalog + :param schema_name: str + The identifier of the schema + :param include_browse: bool (optional) + Whether to include volumes in the response for which the principal can only access selective + metadata for + :param max_results: int (optional) + Maximum number of volumes to return (page length). + + If not set, the page length is set to a server configured value (10000, as of 1/29/2024). - when set + to a value greater than 0, the page length is the minimum of this value and a server configured + value (10000, as of 1/29/2024); - when set to 0, the page length is set to a server configured value + (10000, as of 1/29/2024) (recommended); - when set to a value less than 0, an invalid parameter + error is returned; + + Note: this parameter controls only the maximum number of volumes to return. The actual number of + volumes returned in a page may be smaller than this value, including 0, even if there are more + pages. + :param page_token: str (optional) + Opaque token returned by a previous request. It must be included in the request to retrieve the next + page of results (pagination). + + :returns: Iterator over :class:`VolumeInfo` + parameters: + - name: catalog_name + in: query + required: true + schema: + type: string + description: The identifier of the catalog + - name: schema_name + in: query + required: true + schema: + type: string + description: The identifier of the schema + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include volumes in the response for which the principal + can only access selective metadata for + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of volumes to return (page length). If not set, the + page length is set to a server configured value (10000, as of + 1/29/2024). - when set to a value greater than 0, the page length is + the minimum of this value and a server configured value (10000, as + of 1/29/2024); - when set to 0, the page length is set to a server + configured value (10000, as of 1/29/2024) (recommended); - when set + to a value less than 0, an invalid parameter error is returned; + Note: this parameter controls only the maximum number of volumes to + return. The actual number of volumes returned in a page may be + smaller than this value, including 0, even if there are more pages. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Opaque token returned by a previous request. It must be included in + the request to retrieve the next page of results (pagination). + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListVolumesResponseContent' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VolumesAPI + /api/2.1/unity-catalog/volumes/{name}: + delete: + operationId: volumes_delete + summary: Deletes a volume from the specified parent catalog and schema. + tags: + - catalog + - volumes + description: >- + Deletes a volume from the specified parent catalog and schema. + + + The caller must be a metastore admin or an owner of the volume. For the + latter case, the caller must + + also be the owner or have the **USE_CATALOG** privilege on the parent + catalog and the **USE_SCHEMA** + + privilege on the parent schema. + + + :param name: str + The three-level (fully qualified) name of the volume + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the volume + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VolumesAPI + get: + operationId: volumes_read + summary: Gets a volume from the metastore for a specific catalog and schema. + tags: + - catalog + - volumes + description: >- + Gets a volume from the metastore for a specific catalog and schema. + + + The caller must be a metastore admin or an owner of (or have the **READ + VOLUME** privilege on) the + + volume. For the latter case, the caller must also be the owner or have + the **USE_CATALOG** privilege + + on the parent catalog and the **USE_SCHEMA** privilege on the parent + schema. + + + :param name: str + The three-level (fully qualified) name of the volume + :param include_browse: bool (optional) + Whether to include volumes in the response for which the principal can only access selective + metadata for + + :returns: :class:`VolumeInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the volume + - name: include_browse + in: query + required: false + schema: + type: string + description: >- + Whether to include volumes in the response for which the principal + can only access selective metadata for + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VolumeInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VolumesAPI + patch: + operationId: volumes_update + summary: >- + Updates the specified volume under the specified parent catalog and + schema. + tags: + - catalog + - volumes + description: >- + Updates the specified volume under the specified parent catalog and + schema. + + + The caller must be a metastore admin or an owner of the volume. For the + latter case, the caller must + + also be the owner or have the **USE_CATALOG** privilege on the parent + catalog and the **USE_SCHEMA** + + privilege on the parent schema. + + + Currently only the name, the owner or the comment of the volume could be + updated. + + + :param name: str + The three-level (fully qualified) name of the volume + :param comment: str (optional) + The comment attached to the volume + :param new_name: str (optional) + New name for the volume. + :param owner: str (optional) + The identifier of the user who owns the volume + + :returns: :class:`VolumeInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The three-level (fully qualified) name of the volume + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: The comment attached to the volume + new_name: + type: string + description: New name for the volume. + owner: + type: string + description: The identifier of the user who owns the volume + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VolumeInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VolumesAPI + /api/2.1/unity-catalog/workspace-bindings/catalogs/{name}: + get: + operationId: workspace_bindings_get + summary: >- + Gets workspace bindings of the catalog. The caller must be a metastore + admin or an owner of the + tags: + - catalog + - workspace_bindings + description: >- + Gets workspace bindings of the catalog. The caller must be a metastore + admin or an owner of the + + catalog. + + + :param name: str + The name of the catalog. + + :returns: :class:`GetCatalogWorkspaceBindingsResponse` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the catalog. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetCatalogWorkspaceBindingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceBindingsAPI + patch: + operationId: workspace_bindings_update + summary: >- + Updates workspace bindings of the catalog. The caller must be a + metastore admin or an owner of the + tags: + - catalog + - workspace_bindings + description: >- + Updates workspace bindings of the catalog. The caller must be a + metastore admin or an owner of the + + catalog. + + + :param name: str + The name of the catalog. + :param assign_workspaces: List[int] (optional) + A list of workspace IDs. + :param unassign_workspaces: List[int] (optional) + A list of workspace IDs. + + :returns: :class:`UpdateCatalogWorkspaceBindingsResponse` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the catalog. + requestBody: + content: + application/json: + schema: + type: object + properties: + assign_workspaces: + type: string + description: A list of workspace IDs. + unassign_workspaces: + type: string + description: A list of workspace IDs. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateCatalogWorkspaceBindingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceBindingsAPI + /api/2.1/unity-catalog/bindings/{securable_type}/{securable_name}: + get: + operationId: workspace_bindings_get_bindings + summary: >- + Gets workspace bindings of the securable. The caller must be a metastore + admin or an owner of the + tags: + - catalog + - workspace_bindings + description: >- + Gets workspace bindings of the securable. The caller must be a metastore + admin or an owner of the + + securable. + + + NOTE: we recommend using max_results=0 to use the paginated version of + this API. Unpaginated calls + + will be deprecated soon. + + + PAGINATION BEHAVIOR: When using pagination (max_results >= 0), a page + may contain zero results while + + still providing a next_page_token. Clients must continue reading pages + until next_page_token is + + absent, which is the only indication that the end of results has been + reached. + + + :param securable_type: str + The type of the securable to bind to a workspace (catalog, storage_credential, credential, or + external_location). + :param securable_name: str + The name of the securable. + :param max_results: int (optional) + Maximum number of workspace bindings to return. - When set to 0, the page length is set to a server + configured value (recommended); - When set to a value greater than 0, the page length is the minimum + of this value and a server configured value; - When set to a value less than 0, an invalid parameter + error is returned; - If not set, all the workspace bindings are returned (not recommended). + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`WorkspaceBinding` + parameters: + - name: securable_type + in: path + required: true + schema: + type: string + description: >- + The type of the securable to bind to a workspace (catalog, + storage_credential, credential, or external_location). + - name: securable_name + in: path + required: true + schema: + type: string + description: The name of the securable. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of workspace bindings to return. - When set to 0, the + page length is set to a server configured value (recommended); - + When set to a value greater than 0, the page length is the minimum + of this value and a server configured value; - When set to a value + less than 0, an invalid parameter error is returned; - If not set, + all the workspace bindings are returned (not recommended). + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWorkspaceBindingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceBindingsAPI + patch: + operationId: workspace_bindings_update_bindings + summary: >- + Updates workspace bindings of the securable. The caller must be a + metastore admin or an owner of the + tags: + - catalog + - workspace_bindings + description: >- + Updates workspace bindings of the securable. The caller must be a + metastore admin or an owner of the + + securable. + + + :param securable_type: str + The type of the securable to bind to a workspace (catalog, storage_credential, credential, or + external_location). + :param securable_name: str + The name of the securable. + :param add: List[:class:`WorkspaceBinding`] (optional) + List of workspace bindings to add. If a binding for the workspace already exists with a different + binding_type, adding it again with a new binding_type will update the existing binding (e.g., from + READ_WRITE to READ_ONLY). + :param remove: List[:class:`WorkspaceBinding`] (optional) + List of workspace bindings to remove. + + :returns: :class:`UpdateWorkspaceBindingsResponse` + parameters: + - name: securable_type + in: path + required: true + schema: + type: string + description: >- + The type of the securable to bind to a workspace (catalog, + storage_credential, credential, or external_location). + - name: securable_name + in: path + required: true + schema: + type: string + description: The name of the securable. + requestBody: + content: + application/json: + schema: + type: object + properties: + add: + type: string + description: >- + List of workspace bindings to add. If a binding for the + workspace already exists with a different binding_type, + adding it again with a new binding_type will update the + existing binding (e.g., from READ_WRITE to READ_ONLY). + remove: + type: string + description: List of workspace bindings to remove. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateWorkspaceBindingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceBindingsAPI +components: + schemas: + AccessRequestDestinations: + type: object + properties: + securable: + $ref: '#/components/schemas/Securable' + are_any_destinations_hidden: + type: boolean + description: >- + Indicates whether any destinations are hidden from the caller due to + a lack of permissions. This value is true if the caller does not + have permission to see all destinations. + destination_source_securable: + $ref: '#/components/schemas/Securable' + description: >- + The source securable from which the destinations are inherited. + Either the same value as securable (if destination is set directly + on the securable) or the nearest parent securable with destinations + set. + destinations: + type: array + items: + $ref: '#/components/schemas/NotificationDestination' + description: The access request destinations for the securable. + full_name: + type: string + description: >- + The full name of the securable. Redundant with the name in the + securable object, but necessary for Terraform integration + securable_type: + type: string + description: >- + The type of the securable. Redundant with the type in the securable + object, but necessary for Terraform integration + required: + - securable + AccountsCreateMetastoreAssignmentResponse: + type: object + properties: {} + description: The metastore assignment was successfully created. + AccountsCreateMetastoreResponse: + type: object + properties: + metastore_info: + $ref: '#/components/schemas/MetastoreInfo' + AccountsCreateStorageCredentialInfo: + type: object + properties: + credential_info: + $ref: '#/components/schemas/StorageCredentialInfo' + AccountsDeleteMetastoreAssignmentResponse: + type: object + properties: {} + description: The metastore assignment was successfully deleted. + AccountsDeleteMetastoreResponse: + type: object + properties: {} + description: The metastore was successfully deleted. + AccountsDeleteStorageCredentialResponse: + type: object + properties: {} + description: The storage credential was successfully deleted. + AccountsGetMetastoreResponse: + type: object + properties: + metastore_info: + $ref: '#/components/schemas/MetastoreInfo' + description: The metastore was successfully returned. + AccountsListMetastoresResponse: + type: object + properties: + metastores: + type: array + items: + $ref: '#/components/schemas/MetastoreInfo' + description: An array of metastore information objects. + description: Metastores were returned successfully. + AccountsMetastoreAssignment: + type: object + properties: + metastore_assignment: + $ref: '#/components/schemas/MetastoreAssignment' + description: The workspace metastore assignment was successfully returned. + AccountsStorageCredentialInfo: + type: object + properties: + credential_info: + $ref: '#/components/schemas/StorageCredentialInfo' + description: The storage credential was successfully retrieved. + AccountsUpdateMetastoreAssignmentResponse: + type: object + properties: {} + description: The metastore assignment was successfully updated. + AccountsUpdateMetastoreResponse: + type: object + properties: + metastore_info: + $ref: '#/components/schemas/MetastoreInfo' + description: The metastore update request succeeded. + AccountsUpdateStorageCredentialResponse: + type: object + properties: + credential_info: + $ref: '#/components/schemas/StorageCredentialInfo' + description: The storage credential was successfully updated. + ArtifactAllowlistInfo: + type: object + properties: + artifact_matchers: + type: array + items: + $ref: '#/components/schemas/ArtifactMatcher' + created_at: + type: integer + description: >- + Time at which this artifact allowlist was set, in epoch + milliseconds. + created_by: + type: string + description: Username of the user who set the artifact allowlist. + metastore_id: + type: string + description: Unique identifier of parent metastore. + ArtifactMatcher: + type: object + properties: + artifact: + type: string + match_type: + $ref: '#/components/schemas/MatchType' + description: The pattern matching type of the artifact + required: + - artifact + - match_type + AssignResponse: + type: object + properties: {} + AwsCredentials: + type: object + properties: + access_key_id: + type: string + description: The access key ID that identifies the temporary credentials. + access_point: + type: string + description: >- + The Amazon Resource Name (ARN) of the S3 access point for temporary + credentials related the external location. + secret_access_key: + type: string + description: The secret access key that can be used to sign AWS API requests. + session_token: + type: string + description: >- + The token that users must pass to AWS API to use the temporary + credentials. + description: |- + AWS temporary credentials for API authentication. Read more at + https://docs.aws.amazon.com/STS/latest/APIReference/API_Credentials.html. + AwsIamRole: + type: object + properties: + external_id: + type: string + description: >- + The external ID used in role assumption to prevent the confused + deputy problem. + role_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM role used to vend + temporary credentials. + unity_catalog_iam_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM user managed by + Databricks. This is the identity that is going to assume the AWS IAM + role. + description: The AWS IAM role configuration + AwsIamRoleRequest: + type: object + properties: + role_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM role used to vend + temporary credentials. + required: + - role_arn + description: The AWS IAM role configuration + AwsIamRoleResponse: + type: object + properties: + role_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM role used to vend + temporary credentials. + external_id: + type: string + description: >- + The external ID used in role assumption to prevent the confused + deputy problem. + unity_catalog_iam_arn: + type: string + description: >- + The Amazon Resource Name (ARN) of the AWS IAM user managed by + Databricks. This is the identity that is going to assume the AWS IAM + role. + required: + - role_arn + description: The AWS IAM role configuration + AwsSqsQueue: + type: object + properties: + managed_resource_id: + type: string + queue_url: + type: string + description: >- + The AQS queue url in the format + https://sqs.{region}.amazonaws.com/{account id}/{queue name}. Only + required for provided_sqs. + AzureActiveDirectoryToken: + type: object + properties: + aad_token: + type: string + description: >- + Opaque token that contains claims that you can use in Azure Active + Directory to access cloud services. + description: >- + Azure Active Directory token, essentially the Oauth token for Azure + Service Principal or Managed + Identity. Read more at + https://learn.microsoft.com/en-us/azure/databricks/dev-tools/api/latest/aad/service-prin-aad-token + AzureManagedIdentity: + type: object + properties: + access_connector_id: + type: string + description: >- + The Azure resource ID of the Azure Databricks Access Connector. Use + the format + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.Databricks/accessConnectors/{connector-name}`. + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + managed_identity_id: + type: string + description: >- + The Azure resource ID of the managed identity. Use the format, + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identity-name}` + This is only available for user-assgined identities. For + system-assigned identities, the access_connector_id is used to + identify the identity. If this field is not provided, then we assume + the AzureManagedIdentity is using the system-assigned identity. + required: + - access_connector_id + description: The Azure managed identity configuration. + AzureManagedIdentityRequest: + type: object + properties: + access_connector_id: + type: string + description: >- + The Azure resource ID of the Azure Databricks Access Connector. Use + the format + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.Databricks/accessConnectors/{connector-name}`. + managed_identity_id: + type: string + description: >- + The Azure resource ID of the managed identity. Use the format, + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identity-name}` + This is only available for user-assgined identities. For + system-assigned identities, the access_connector_id is used to + identify the identity. If this field is not provided, then we assume + the AzureManagedIdentity is using the system-assigned identity. + required: + - access_connector_id + description: The Azure managed identity configuration. + AzureManagedIdentityResponse: + type: object + properties: + access_connector_id: + type: string + description: >- + The Azure resource ID of the Azure Databricks Access Connector. Use + the format + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.Databricks/accessConnectors/{connector-name}`. + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + managed_identity_id: + type: string + description: >- + The Azure resource ID of the managed identity. Use the format, + `/subscriptions/{guid}/resourceGroups/{rg-name}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identity-name}` + This is only available for user-assgined identities. For + system-assigned identities, the access_connector_id is used to + identify the identity. If this field is not provided, then we assume + the AzureManagedIdentity is using the system-assigned identity. + required: + - access_connector_id + description: The Azure managed identity configuration. + AzureQueueStorage: + type: object + properties: + managed_resource_id: + type: string + queue_url: + type: string + description: >- + The AQS queue url in the format https://{storage + account}.queue.core.windows.net/{queue name} Only required for + provided_aqs. + resource_group: + type: string + description: >- + Optional resource group for the queue, event grid subscription, and + external location storage account. Only required for locations with + a service principal storage credential + subscription_id: + type: string + description: >- + Optional subscription id for the queue, event grid subscription, and + external location storage account. Required for locations with a + service principal storage credential + AzureServicePrincipal: + type: object + properties: + directory_id: + type: string + description: >- + The directory ID corresponding to the Azure Active Directory (AAD) + tenant of the application. + application_id: + type: string + description: >- + The application ID of the application registration within the + referenced AAD tenant. + client_secret: + type: string + description: The client secret generated for the above app ID in AAD. + required: + - directory_id + - application_id + - client_secret + description: >- + The Azure service principal configuration. Only applicable when purpose + is **STORAGE**. + AzureUserDelegationSas: + type: object + properties: + sas_token: + type: string + description: >- + The signed URI (SAS Token) used to access blob services for a given + path + description: |- + Azure temporary credentials for API authentication. Read more at + https://docs.microsoft.com/en-us/rest/api/storageservices/create-user-delegation-sas + BatchCreateAccessRequestsResponse: + type: object + properties: + responses: + type: array + items: + $ref: '#/components/schemas/CreateAccessRequestResponse' + CancelRefreshResponse: + type: object + properties: {} + CatalogInfo: + type: object + properties: + browse_only: + type: boolean + catalog_type: + $ref: '#/components/schemas/CatalogType' + comment: + type: string + description: User-provided free-form text description. + connection_name: + type: string + description: The name of the connection to an external data source. + created_at: + type: integer + description: Time at which this catalog was created, in epoch milliseconds. + created_by: + type: string + description: Username of catalog creator. + effective_predictive_optimization_flag: + $ref: '#/components/schemas/EffectivePredictiveOptimizationFlag' + enable_predictive_optimization: + $ref: '#/components/schemas/EnablePredictiveOptimization' + description: >- + Whether predictive optimization should be enabled for this object + and objects under it. + full_name: + type: string + description: The full name of the catalog. Corresponds with the name field. + isolation_mode: + $ref: '#/components/schemas/CatalogIsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of catalog. + options: + type: object + description: A map of key-value properties attached to the securable. + owner: + type: string + description: Username of current owner of catalog. + properties: + type: object + description: A map of key-value properties attached to the securable. + provider_name: + type: string + description: >- + The name of delta sharing provider. A Delta Sharing catalog is a + catalog that is based on a Delta share on a remote sharing server. + provisioning_info: + $ref: '#/components/schemas/ProvisioningInfo' + securable_type: + $ref: '#/components/schemas/SecurableType' + share_name: + type: string + description: The name of the share under the share provider. + storage_location: + type: string + description: Storage Location URL (full path) for managed tables within catalog. + storage_root: + type: string + description: Storage root URL for managed tables within catalog. + updated_at: + type: integer + description: Time at which this catalog was last modified, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified catalog. + CloudflareApiToken: + type: object + properties: + access_key_id: + type: string + description: The access key ID associated with the API token. + secret_access_key: + type: string + description: The secret access token generated for the above access key ID. + account_id: + type: string + description: The ID of the account associated with the API token. + required: + - access_key_id + - secret_access_key + description: |- + The Cloudflare API token configuration. Read more at + https://developers.cloudflare.com/r2/api/s3/tokens/ + ColumnInfo: + type: object + properties: + comment: + type: string + mask: + $ref: '#/components/schemas/ColumnMask' + name: + type: string + description: Name of Column. + nullable: + type: boolean + description: 'Whether field may be Null (default: true).' + partition_index: + type: integer + description: Partition index for column. + position: + type: integer + description: Ordinal position of column (starting at position 0). + type_interval_type: + type: string + description: Format of IntervalType. + type_json: + type: string + description: Full data type specification, JSON-serialized. + type_name: + $ref: '#/components/schemas/ColumnTypeName' + type_precision: + type: integer + description: Digits of precision; required for DecimalTypes. + type_scale: + type: integer + description: Digits to right of decimal; Required for DecimalTypes. + type_text: + type: string + description: Full data type specification as SQL/catalogString text. + ColumnMask: + type: object + properties: + function_name: + type: string + using_column_names: + type: array + items: + type: string + description: >- + The list of additional table columns to be passed as input to the + column mask function. The first arg of the mask function should be + of the type of the column being masked and the types of the rest of + the args should match the types of columns in 'using_column_names'. + ColumnMaskOptions: + type: object + properties: + function_name: + type: string + on_column: + type: string + description: >- + The alias of the column to be masked. The alias must refer to one of + matched columns. The values of the column is passed to the column + mask function as the first argument. Required on create and update. + using: + type: array + items: + $ref: '#/components/schemas/FunctionArgument' + description: >- + Optional list of column aliases or constant literals to be passed as + additional arguments to the column mask function. The type of each + column should match the positional argument of the column mask + function. + required: + - function_name + - on_column + ColumnRelationship: + type: object + properties: + source: + type: string + target: + type: string + ConnectionDependency: + type: object + properties: + connection_name: + type: string + description: >- + Full name of the dependent connection, in the form of + __connection_name__. + description: A connection that is dependent on a SQL object. + ConnectionInfo: + type: object + properties: + comment: + type: string + description: User-provided free-form text description. + connection_id: + type: string + description: Unique identifier of the Connection. + connection_type: + $ref: '#/components/schemas/ConnectionType' + description: The type of connection. + created_at: + type: integer + description: Time at which this connection was created, in epoch milliseconds. + created_by: + type: string + description: Username of connection creator. + credential_type: + $ref: '#/components/schemas/CredentialType' + description: The type of credential. + full_name: + type: string + description: Full name of connection. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of the connection. + options: + type: object + description: A map of key-value properties attached to the securable. + owner: + type: string + description: Username of current owner of the connection. + properties: + type: object + description: A map of key-value properties attached to the securable. + provisioning_info: + $ref: '#/components/schemas/ProvisioningInfo' + read_only: + type: boolean + description: If the connection is read only. + securable_type: + $ref: '#/components/schemas/SecurableType' + updated_at: + type: integer + description: Time at which this connection was updated, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified connection. + url: + type: string + description: URL of the remote data source, extracted from options. + description: 'Next ID: 24' + ContinuousUpdateStatus: + type: object + properties: + initial_pipeline_sync_progress: + $ref: '#/components/schemas/PipelineProgress' + description: Progress of the initial data synchronization. + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was synced to the online + table. Note that this Delta version may not be completely synced to + the online table yet. + timestamp: + type: string + description: >- + The timestamp of the last time any data was synchronized from the + source table to the online table. + description: >- + Detailed status of an online table. Shown if the online table is in the + ONLINE_CONTINUOUS_UPDATE + or the ONLINE_UPDATING_PIPELINE_RESOURCES state. + CreateAccessRequest: + type: object + properties: + behalf_of: + $ref: '#/components/schemas/Principal' + comment: + type: string + description: >- + Optional. Comment associated with the request. At most 200 + characters, can only contain lowercase/uppercase letters (a-z, A-Z), + numbers (0-9), punctuation, and spaces. + securable_permissions: + type: array + items: + $ref: '#/components/schemas/SecurablePermissions' + description: >- + List of securables and their corresponding requested UC privileges. + At most 30 securables can be requested for a principal per batched + call. Each securable can only be requested once per principal. + CreateAccessRequestResponse: + type: object + properties: + behalf_of: + $ref: '#/components/schemas/Principal' + request_destinations: + type: array + items: + $ref: '#/components/schemas/AccessRequestDestinations' + description: >- + The access request destinations for all the securables the principal + requested. + CreateAccountsMetastore: + type: object + properties: + name: + type: string + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + region: + type: string + description: >- + Cloud region which the metastore serves (e.g., `us-west-2`, + `westus`). + storage_root: + type: string + description: The storage root URL for metastore + required: + - name + CreateAccountsStorageCredential: + type: object + properties: + name: + type: string + aws_iam_role: + $ref: '#/components/schemas/AwsIamRoleRequest' + description: The AWS IAM role configuration. + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentityRequest' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + cloudflare_api_token: + $ref: '#/components/schemas/CloudflareApiToken' + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccountRequest' + description: The Databricks managed GCP service account configuration. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + required: + - name + CreateFunction: + type: object + properties: + name: + type: string + catalog_name: + type: string + description: Name of parent Catalog. + schema_name: + type: string + description: Name of parent Schema relative to its parent Catalog. + input_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Function input parameters. + data_type: + $ref: '#/components/schemas/ColumnTypeName' + description: Scalar function return data type. + full_data_type: + type: string + description: Pretty printed function data type. + routine_body: + $ref: '#/components/schemas/CreateFunctionRoutineBody' + description: >- + Function language. When **EXTERNAL** is used, the language of the + routine function should be specified in the **external_language** + field, and the **return_params** of the function cannot be used (as + **TABLE** return type is not supported), and the **sql_data_access** + field must be **NO_SQL**. + routine_definition: + type: string + description: Function body. + parameter_style: + $ref: '#/components/schemas/CreateFunctionParameterStyle' + description: Function parameter style. **S** is the value for SQL. + is_deterministic: + type: boolean + description: Whether the function is deterministic. + sql_data_access: + $ref: '#/components/schemas/CreateFunctionSqlDataAccess' + description: Function SQL data access. + is_null_call: + type: boolean + description: Function null call. + security_type: + $ref: '#/components/schemas/CreateFunctionSecurityType' + description: Function security type. + specific_name: + type: string + description: Specific name of the function; Reserved for future use. + comment: + type: string + description: User-provided free-form text description. + external_language: + type: string + description: External function language. + external_name: + type: string + description: External function name. + properties: + type: string + description: JSON-serialized key-value pair map, encoded (escaped) as a string. + return_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Table function return parameters. + routine_dependencies: + $ref: '#/components/schemas/DependencyList' + description: function dependencies. + sql_path: + type: string + description: >- + List of schemes whose objects can be referenced without + qualification. + required: + - name + - catalog_name + - schema_name + - input_params + - data_type + - full_data_type + - routine_body + - routine_definition + - parameter_style + - is_deterministic + - sql_data_access + - is_null_call + - security_type + - specific_name + CreateMetastoreAssignment: + type: object + properties: + workspace_id: + type: integer + metastore_id: + type: string + description: The unique ID of the metastore. + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field is + deprecated. Please use "Default Namespace API" to configure the + default catalog for a Databricks workspace. + required: + - workspace_id + - metastore_id + - default_catalog_name + CreateRequestExternalLineage: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + CredentialDependency: + type: object + properties: + credential_name: + type: string + description: >- + Full name of the dependent credential, in the form of + __credential_name__. + description: A credential that is dependent on a SQL object. + CredentialInfo: + type: object + properties: + aws_iam_role: + $ref: '#/components/schemas/AwsIamRole' + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentity' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + comment: + type: string + description: Comment associated with the credential. + created_at: + type: integer + description: Time at which this credential was created, in epoch milliseconds. + created_by: + type: string + description: Username of credential creator. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccount' + description: The Databricks managed GCP service account configuration. + full_name: + type: string + description: The full name of the credential. + id: + type: string + description: The unique identifier of the credential. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + metastore_id: + type: string + description: Unique identifier of the parent metastore. + name: + type: string + description: >- + The credential name. The name must be unique among storage and + service credentials within the metastore. + owner: + type: string + description: Username of current owner of credential. + purpose: + $ref: '#/components/schemas/CredentialPurpose' + description: Indicates the purpose of the credential. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + updated_at: + type: integer + description: >- + Time at which this credential was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the credential. + used_for_managed_storage: + type: boolean + description: >- + Whether this credential is the current metastore's root storage + credential. Only applicable when purpose is **STORAGE**. + CredentialValidationResult: + type: object + properties: + message: + type: string + result: + $ref: '#/components/schemas/ValidateCredentialResult' + description: The results of the tested operation. + DatabricksGcpServiceAccount: + type: object + properties: + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + email: + type: string + description: The email of the service account. + private_key_id: + type: string + description: The ID that represents the private key for this Service Account + description: >- + GCP long-lived credential. Databricks-created Google Cloud Storage + service account. + DatabricksGcpServiceAccountRequest: + type: object + properties: {} + description: >- + GCP long-lived credential. Databricks-created Google Cloud Storage + service account. + DatabricksGcpServiceAccountResponse: + type: object + properties: + credential_id: + type: string + description: The Databricks internal ID that represents this managed identity. + email: + type: string + description: The email of the service account. + description: >- + GCP long-lived credential. Databricks-created Google Cloud Storage + service account. + DeleteCredentialResponse: + type: object + properties: {} + DeleteMonitorResponse: + type: object + properties: {} + DeletePolicyResponse: + type: object + properties: {} + DeleteRequestExternalLineage: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + id: + type: string + description: Unique identifier of the external lineage relationship. + required: + - source + - target + DeleteResponse: + type: object + properties: {} + DeleteTableConstraintResponse: + type: object + properties: {} + DeltaRuntimePropertiesKvPairs: + type: object + properties: + delta_runtime_properties: + type: object + description: A map of key-value properties attached to the securable. + required: + - delta_runtime_properties + description: >- + Properties pertaining to the current state of the delta table as given + by the commit server. + This does not contain **delta.*** (input) properties in __TableInfo.properties__. + Dependency: + type: object + properties: + connection: + $ref: '#/components/schemas/ConnectionDependency' + credential: + $ref: '#/components/schemas/CredentialDependency' + function: + $ref: '#/components/schemas/FunctionDependency' + table: + $ref: '#/components/schemas/TableDependency' + description: >- + A dependency of a SQL object. One of the following fields must be + defined: __table__, + __function__, __connection__, or __credential__. + DependencyList: + type: object + properties: + dependencies: + type: array + items: + $ref: '#/components/schemas/Dependency' + description: Array of dependencies. + description: A list of dependencies. + DisableResponse: + type: object + properties: {} + EffectivePermissionsList: + type: object + properties: + next_page_token: + type: string + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/EffectivePrivilegeAssignment' + description: >- + The privileges conveyed to each principal (either directly or via + inheritance) + EffectivePredictiveOptimizationFlag: + type: object + properties: + value: + $ref: '#/components/schemas/EnablePredictiveOptimization' + inherited_from_name: + type: string + description: >- + The name of the object from which the flag was inherited. If there + was no inheritance, this field is left blank. + inherited_from_type: + $ref: >- + #/components/schemas/EffectivePredictiveOptimizationFlagInheritedFromType + description: >- + The type of the object from which the flag was inherited. If there + was no inheritance, this field is left blank. + required: + - value + EffectivePrivilege: + type: object + properties: + inherited_from_name: + type: string + inherited_from_type: + $ref: '#/components/schemas/SecurableType' + description: >- + The type of the object that conveys this privilege via inheritance. + This field is omitted when privilege is not inherited (it's assigned + to the securable itself). + privilege: + $ref: '#/components/schemas/Privilege' + description: The privilege assigned to the principal. + EffectivePrivilegeAssignment: + type: object + properties: + principal: + type: string + privileges: + type: array + items: + $ref: '#/components/schemas/EffectivePrivilege' + description: >- + The privileges conveyed to the principal (either directly or via + inheritance). + EnableResponse: + type: object + properties: {} + EncryptionDetails: + type: object + properties: + sse_encryption_details: + $ref: '#/components/schemas/SseEncryptionDetails' + description: >- + Server-Side Encryption properties for clients communicating with AWS + s3. + description: Encryption options that apply to clients connecting to cloud storage. + EntityTagAssignment: + type: object + properties: + entity_name: + type: string + description: The fully qualified name of the entity to which the tag is assigned + tag_key: + type: string + description: The key of the tag + entity_type: + type: string + description: >- + The type of the entity to which the tag is assigned. Allowed values + are: catalogs, schemas, tables, columns, volumes. + tag_value: + type: string + description: The value of the tag + required: + - entity_name + - tag_key + - entity_type + description: Represents a tag assignment to an entity + ExternalLineageExternalMetadata: + type: object + properties: + name: + type: string + ExternalLineageExternalMetadataInfo: + type: object + properties: + entity_type: + type: string + description: Type of entity represented by the external metadata object. + event_time: + type: string + description: Timestamp of the lineage event. + name: + type: string + description: Name of the external metadata object. + system_type: + $ref: '#/components/schemas/SystemType' + description: Type of external system. + description: Represents the external metadata object in the lineage event. + ExternalLineageFileInfo: + type: object + properties: + event_time: + type: string + description: Timestamp of the lineage event. + path: + type: string + description: URL of the path. + securable_name: + type: string + description: The full name of the securable on the path. + securable_type: + type: string + description: The securable type of the securable on the path. + storage_location: + type: string + description: The storage location associated with securable on the path. + description: Represents the path information in the lineage event. + ExternalLineageInfo: + type: object + properties: + external_lineage_info: + $ref: '#/components/schemas/ExternalLineageRelationshipInfo' + description: >- + Information about the edge metadata of the external lineage + relationship. + external_metadata_info: + $ref: '#/components/schemas/ExternalLineageExternalMetadataInfo' + description: >- + Information about external metadata involved in the lineage + relationship. + file_info: + $ref: '#/components/schemas/ExternalLineageFileInfo' + description: Information about the file involved in the lineage relationship. + model_info: + $ref: '#/components/schemas/ExternalLineageModelVersionInfo' + description: >- + Information about the model version involved in the lineage + relationship. + table_info: + $ref: '#/components/schemas/ExternalLineageTableInfo' + description: Information about the table involved in the lineage relationship. + description: Lineage response containing lineage information of a data asset. + ExternalLineageModelVersion: + type: object + properties: + name: + type: string + version: + type: string + ExternalLineageModelVersionInfo: + type: object + properties: + event_time: + type: string + description: Timestamp of the lineage event. + model_name: + type: string + description: Name of the model. + version: + type: integer + description: Version number of the model. + description: Represents the model version information in the lineage event. + ExternalLineageObject: + type: object + properties: + external_metadata: + $ref: '#/components/schemas/ExternalLineageExternalMetadata' + model_version: + $ref: '#/components/schemas/ExternalLineageModelVersion' + path: + $ref: '#/components/schemas/ExternalLineagePath' + table: + $ref: '#/components/schemas/ExternalLineageTable' + ExternalLineagePath: + type: object + properties: + url: + type: string + ExternalLineageRelationship: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + ExternalLineageRelationshipInfo: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + ExternalLineageTable: + type: object + properties: + name: + type: string + ExternalLineageTableInfo: + type: object + properties: + catalog_name: + type: string + description: Name of Catalog. + event_time: + type: string + description: Timestamp of the lineage event. + name: + type: string + description: Name of Table. + schema_name: + type: string + description: Name of Schema. + description: Represents the table information in the lineage event. + ExternalLocationInfo: + type: object + properties: + browse_only: + type: boolean + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: >- + Time at which this external location was created, in epoch + milliseconds. + created_by: + type: string + description: Username of external location creator. + credential_id: + type: string + description: Unique ID of the location's storage credential. + credential_name: + type: string + description: Name of the storage credential used with this location. + enable_file_events: + type: boolean + description: >- + Whether to enable file events on this external location. Default to + `true`. Set to `false` to disable file events. + encryption_details: + $ref: '#/components/schemas/EncryptionDetails' + fallback: + type: boolean + description: >- + Indicates whether fallback mode is enabled for this external + location. When fallback mode is enabled, the access to the location + falls back to cluster credentials if UC credentials are not + sufficient. + file_event_queue: + $ref: '#/components/schemas/FileEventQueue' + description: >- + File event queue settings. If `enable_file_events` is not `false`, + must be defined and have exactly one of the documented properties. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + metastore_id: + type: string + description: Unique identifier of metastore hosting the external location. + name: + type: string + description: Name of the external location. + owner: + type: string + description: The owner of the external location. + read_only: + type: boolean + description: Indicates whether the external location is read-only. + updated_at: + type: integer + description: >- + Time at which external location this was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the external location. + url: + type: string + description: Path URL of the external location. + ExternalMetadata: + type: object + properties: + name: + type: string + system_type: + $ref: '#/components/schemas/SystemType' + description: Type of external system. + entity_type: + type: string + description: Type of entity within the external system. + columns: + type: array + items: + type: string + description: List of columns associated with the external metadata object. + create_time: + type: string + description: Time at which this external metadata object was created. + created_by: + type: string + description: Username of external metadata object creator. + description: + type: string + description: User-provided free-form text description. + id: + type: string + description: Unique identifier of the external metadata object. + metastore_id: + type: string + description: Unique identifier of parent metastore. + owner: + type: string + description: Owner of the external metadata object. + properties: + type: object + description: >- + A map of key-value properties attached to the external metadata + object. + update_time: + type: string + description: Time at which this external metadata object was last modified. + updated_by: + type: string + description: Username of user who last modified external metadata object. + url: + type: string + description: URL associated with the external metadata object. + required: + - name + - system_type + - entity_type + FailedStatus: + type: object + properties: + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was synced to the online + table. Note that this Delta version may only be partially synced to + the online table. Only populated if the table is still online and + available for serving. + timestamp: + type: string + description: >- + The timestamp of the last time any data was synchronized from the + source table to the online table. Only populated if the table is + still online and available for serving. + description: >- + Detailed status of an online table. Shown if the online table is in the + OFFLINE_FAILED or the + ONLINE_PIPELINE_FAILED state. + FileEventQueue: + type: object + properties: + managed_aqs: + $ref: '#/components/schemas/AzureQueueStorage' + managed_pubsub: + $ref: '#/components/schemas/GcpPubsub' + managed_sqs: + $ref: '#/components/schemas/AwsSqsQueue' + provided_aqs: + $ref: '#/components/schemas/AzureQueueStorage' + provided_pubsub: + $ref: '#/components/schemas/GcpPubsub' + provided_sqs: + $ref: '#/components/schemas/AwsSqsQueue' + ForeignKeyConstraint: + type: object + properties: + name: + type: string + child_columns: + type: array + items: + type: string + description: Column names for this constraint. + parent_table: + type: string + description: The full name of the parent constraint. + parent_columns: + type: array + items: + type: string + description: Column names for this constraint. + rely: + type: boolean + description: True if the constraint is RELY, false or unset if NORELY. + required: + - name + - child_columns + - parent_table + - parent_columns + FunctionArgument: + type: object + properties: + alias: + type: string + constant: + type: string + description: A constant literal. + FunctionDependency: + type: object + properties: + function_full_name: + type: string + description: >- + Full name of the dependent function, in the form of + __catalog_name__.__schema_name__.__function_name__. + required: + - function_full_name + description: A function that is dependent on a SQL object. + FunctionInfo: + type: object + properties: + browse_only: + type: boolean + catalog_name: + type: string + description: Name of parent Catalog. + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: Time at which this function was created, in epoch milliseconds. + created_by: + type: string + description: Username of function creator. + data_type: + $ref: '#/components/schemas/ColumnTypeName' + description: Scalar function return data type. + external_language: + type: string + description: External function language. + external_name: + type: string + description: External function name. + full_data_type: + type: string + description: Pretty printed function data type. + full_name: + type: string + description: >- + Full name of Function, in form of + **catalog_name**.**schema_name**.**function_name** + function_id: + type: string + description: Id of Function, relative to parent schema. + input_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Function input parameters. + is_deterministic: + type: boolean + description: Whether the function is deterministic. + is_null_call: + type: boolean + description: Function null call. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of function, relative to parent schema. + owner: + type: string + description: Username of current owner of the function. + parameter_style: + $ref: '#/components/schemas/FunctionInfoParameterStyle' + description: Function parameter style. **S** is the value for SQL. + properties: + type: string + description: JSON-serialized key-value pair map, encoded (escaped) as a string. + return_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: Table function return parameters. + routine_body: + $ref: '#/components/schemas/FunctionInfoRoutineBody' + description: >- + Function language. When **EXTERNAL** is used, the language of the + routine function should be specified in the **external_language** + field, and the **return_params** of the function cannot be used (as + **TABLE** return type is not supported), and the **sql_data_access** + field must be **NO_SQL**. + routine_definition: + type: string + description: Function body. + routine_dependencies: + $ref: '#/components/schemas/DependencyList' + description: function dependencies. + schema_name: + type: string + description: Name of parent Schema relative to its parent Catalog. + security_type: + $ref: '#/components/schemas/FunctionInfoSecurityType' + description: Function security type. + specific_name: + type: string + description: Specific name of the function; Reserved for future use. + sql_data_access: + $ref: '#/components/schemas/FunctionInfoSqlDataAccess' + description: Function SQL data access. + sql_path: + type: string + description: >- + List of schemes whose objects can be referenced without + qualification. + updated_at: + type: integer + description: >- + Time at which this function was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the function. + FunctionParameterInfo: + type: object + properties: + name: + type: string + type_text: + type: string + description: Full data type spec, SQL/catalogString text. + type_name: + $ref: '#/components/schemas/ColumnTypeName' + description: Name of type (INT, STRUCT, MAP, etc.) + position: + type: integer + description: Ordinal position of column (starting at position 0). + comment: + type: string + description: User-provided free-form text description. + parameter_default: + type: string + description: Default value of the parameter. + parameter_mode: + $ref: '#/components/schemas/FunctionParameterMode' + description: Function parameter mode. + parameter_type: + $ref: '#/components/schemas/FunctionParameterType' + description: Function parameter type. + type_interval_type: + type: string + description: Format of IntervalType. + type_json: + type: string + description: Full data type spec, JSON-serialized. + type_precision: + type: integer + description: Digits of precision; required on Create for DecimalTypes. + type_scale: + type: integer + description: Digits to right of decimal; Required on Create for DecimalTypes. + required: + - name + - type_text + - type_name + - position + FunctionParameterInfos: + type: object + properties: + parameters: + type: array + items: + $ref: '#/components/schemas/FunctionParameterInfo' + GcpOauthToken: + type: object + properties: + oauth_token: + type: string + description: |- + GCP temporary credentials for API authentication. Read more at + https://developers.google.com/identity/protocols/oauth2/service-account + GcpPubsub: + type: object + properties: + managed_resource_id: + type: string + subscription_name: + type: string + description: >- + The Pub/Sub subscription name in the format + projects/{project}/subscriptions/{subscription name}. Only required + for provided_pubsub. + GenerateTemporaryPathCredentialResponse: + type: object + properties: + aws_temp_credentials: + $ref: '#/components/schemas/AwsCredentials' + azure_aad: + $ref: '#/components/schemas/AzureActiveDirectoryToken' + azure_user_delegation_sas: + $ref: '#/components/schemas/AzureUserDelegationSas' + expiration_time: + type: integer + description: >- + Server time when the credential will expire, in epoch milliseconds. + The API client is advised to cache the credential given this + expiration time. + gcp_oauth_token: + $ref: '#/components/schemas/GcpOauthToken' + r2_temp_credentials: + $ref: '#/components/schemas/R2Credentials' + url: + type: string + description: The URL of the storage path accessible by the temporary credential. + GenerateTemporaryServiceCredentialAzureOptions: + type: object + properties: + resources: + type: array + items: + type: string + description: >- + The resources to which the temporary Azure credential should apply. + These resources are the scopes that are passed to the token provider + (see + https://learn.microsoft.com/python/api/azure-core/azure.core.credentials.tokencredential?view=azure-python) + description: The Azure cloud options to customize the requested temporary credential + GenerateTemporaryServiceCredentialGcpOptions: + type: object + properties: + scopes: + type: array + items: + type: string + description: >- + The scopes to which the temporary GCP credential should apply. These + resources are the scopes that are passed to the token provider (see + https://google-auth.readthedocs.io/en/latest/reference/google.auth.html#google.auth.credentials.Credentials) + description: The GCP cloud options to customize the requested temporary credential + GenerateTemporaryTableCredentialResponse: + type: object + properties: + aws_temp_credentials: + $ref: '#/components/schemas/AwsCredentials' + azure_aad: + $ref: '#/components/schemas/AzureActiveDirectoryToken' + azure_user_delegation_sas: + $ref: '#/components/schemas/AzureUserDelegationSas' + expiration_time: + type: integer + description: >- + Server time when the credential will expire, in epoch milliseconds. + The API client is advised to cache the credential given this + expiration time. + gcp_oauth_token: + $ref: '#/components/schemas/GcpOauthToken' + r2_temp_credentials: + $ref: '#/components/schemas/R2Credentials' + url: + type: string + description: The URL of the storage path accessible by the temporary credential. + GetCatalogWorkspaceBindingsResponse: + type: object + properties: + workspaces: + type: array + items: + type: integer + GetMetastoreSummaryResponse: + type: object + properties: + cloud: + type: string + created_at: + type: integer + description: Time at which this metastore was created, in epoch milliseconds. + created_by: + type: string + description: Username of metastore creator. + default_data_access_config_id: + type: string + description: >- + Unique identifier of the metastore's (Default) Data Access + Configuration. + delta_sharing_organization_name: + type: string + description: >- + The organization name of a Delta Sharing entity, to be used in + Databricks-to-Databricks Delta Sharing as the official name. + delta_sharing_recipient_token_lifetime_in_seconds: + type: integer + description: The lifetime of delta sharing recipient token in seconds. + delta_sharing_scope: + $ref: '#/components/schemas/DeltaSharingScopeEnum' + description: The scope of Delta Sharing enabled for the metastore. + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + global_metastore_id: + type: string + description: >- + Globally unique metastore ID across clouds and regions, of the form + `cloud:region:metastore_id`. + metastore_id: + type: string + description: Unique identifier of metastore. + name: + type: string + description: The user-specified name of the metastore. + owner: + type: string + description: The owner of the metastore. + privilege_model_version: + type: string + description: >- + Privilege model version of the metastore, of the form `major.minor` + (e.g., `1.0`). + region: + type: string + description: >- + Cloud region which the metastore serves (e.g., `us-west-2`, + `westus`). + storage_root: + type: string + description: The storage root URL for metastore + storage_root_credential_id: + type: string + description: UUID of storage credential to access the metastore storage_root. + storage_root_credential_name: + type: string + description: Name of the storage credential to access the metastore storage_root. + updated_at: + type: integer + description: >- + Time at which the metastore was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the metastore. + GetPermissionsResponse: + type: object + properties: + next_page_token: + type: string + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/PrivilegeAssignment' + description: The privileges assigned to each principal + GetQuotaResponse: + type: object + properties: + quota_info: + $ref: '#/components/schemas/QuotaInfo' + GetWorkspaceBindingsResponse: + type: object + properties: + bindings: + type: array + items: + $ref: '#/components/schemas/WorkspaceBinding' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListAccountMetastoreAssignmentsResponse: + type: object + properties: + workspace_ids: + type: array + items: + type: integer + description: The metastore assignments were successfully returned. + ListAccountStorageCredentialsResponse: + type: object + properties: + storage_credentials: + type: array + items: + $ref: '#/components/schemas/StorageCredentialInfo' + description: An array of metastore storage credentials. + description: The metastore storage credentials were successfully returned. + ListCatalogsResponse: + type: object + properties: + catalogs: + type: array + items: + $ref: '#/components/schemas/CatalogInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListConnectionsResponse: + type: object + properties: + connections: + type: array + items: + $ref: '#/components/schemas/ConnectionInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListCredentialsResponse: + type: object + properties: + credentials: + type: array + items: + $ref: '#/components/schemas/CredentialInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListEntityTagAssignmentsResponse: + type: object + properties: + next_page_token: + type: string + tag_assignments: + type: array + items: + $ref: '#/components/schemas/EntityTagAssignment' + description: The list of tag assignments + ListExternalLineageRelationshipsResponse: + type: object + properties: + external_lineage_relationships: + type: array + items: + $ref: '#/components/schemas/ExternalLineageInfo' + next_page_token: + type: string + ListExternalLocationsResponse: + type: object + properties: + external_locations: + type: array + items: + $ref: '#/components/schemas/ExternalLocationInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListExternalMetadataResponse: + type: object + properties: + external_metadata: + type: array + items: + $ref: '#/components/schemas/ExternalMetadata' + next_page_token: + type: string + ListFunctionsResponse: + type: object + properties: + functions: + type: array + items: + $ref: '#/components/schemas/FunctionInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListMetastoresResponse: + type: object + properties: + metastores: + type: array + items: + $ref: '#/components/schemas/MetastoreInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListModelVersionsResponse: + type: object + properties: + model_versions: + type: array + items: + $ref: '#/components/schemas/ModelVersionInfo' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. __page_token__ should be set to this value for + the next request (for the next page of results). + ListPoliciesResponse: + type: object + properties: + next_page_token: + type: string + policies: + type: array + items: + $ref: '#/components/schemas/PolicyInfo' + description: The list of retrieved policies. + ListQuotasResponse: + type: object + properties: + next_page_token: + type: string + quotas: + type: array + items: + $ref: '#/components/schemas/QuotaInfo' + description: An array of returned QuotaInfos. + ListRegisteredModelsResponse: + type: object + properties: + next_page_token: + type: string + registered_models: + type: array + items: + $ref: '#/components/schemas/RegisteredModelInfo' + ListSchemasResponse: + type: object + properties: + next_page_token: + type: string + schemas: + type: array + items: + $ref: '#/components/schemas/SchemaInfo' + description: An array of schema information objects. + ListStorageCredentialsResponse: + type: object + properties: + next_page_token: + type: string + storage_credentials: + type: array + items: + $ref: '#/components/schemas/StorageCredentialInfo' + ListSystemSchemasResponse: + type: object + properties: + next_page_token: + type: string + schemas: + type: array + items: + $ref: '#/components/schemas/SystemSchemaInfo' + description: An array of system schema information objects. + ListTableSummariesResponse: + type: object + properties: + next_page_token: + type: string + tables: + type: array + items: + $ref: '#/components/schemas/TableSummary' + description: List of table summaries. + ListTablesResponse: + type: object + properties: + next_page_token: + type: string + tables: + type: array + items: + $ref: '#/components/schemas/TableInfo' + description: An array of table information objects. + ListVolumesResponseContent: + type: object + properties: + next_page_token: + type: string + volumes: + type: array + items: + $ref: '#/components/schemas/VolumeInfo' + MatchColumn: + type: object + properties: + alias: + type: string + condition: + type: string + description: The condition expression used to match a table column. + MetastoreAssignment: + type: object + properties: + workspace_id: + type: integer + metastore_id: + type: string + description: The unique ID of the metastore. + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field is + deprecated. Please use "Default Namespace API" to configure the + default catalog for a Databricks workspace. + required: + - workspace_id + - metastore_id + MetastoreInfo: + type: object + properties: + cloud: + type: string + created_at: + type: integer + description: Time at which this metastore was created, in epoch milliseconds. + created_by: + type: string + description: Username of metastore creator. + default_data_access_config_id: + type: string + description: >- + Unique identifier of the metastore's (Default) Data Access + Configuration. + delta_sharing_organization_name: + type: string + description: >- + The organization name of a Delta Sharing entity, to be used in + Databricks-to-Databricks Delta Sharing as the official name. + delta_sharing_recipient_token_lifetime_in_seconds: + type: integer + description: The lifetime of delta sharing recipient token in seconds. + delta_sharing_scope: + $ref: '#/components/schemas/DeltaSharingScopeEnum' + description: The scope of Delta Sharing enabled for the metastore. + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + global_metastore_id: + type: string + description: >- + Globally unique metastore ID across clouds and regions, of the form + `cloud:region:metastore_id`. + metastore_id: + type: string + description: Unique identifier of metastore. + name: + type: string + description: The user-specified name of the metastore. + owner: + type: string + description: The owner of the metastore. + privilege_model_version: + type: string + description: >- + Privilege model version of the metastore, of the form `major.minor` + (e.g., `1.0`). + region: + type: string + description: >- + Cloud region which the metastore serves (e.g., `us-west-2`, + `westus`). + storage_root: + type: string + description: The storage root URL for metastore + storage_root_credential_id: + type: string + description: UUID of storage credential to access the metastore storage_root. + storage_root_credential_name: + type: string + description: Name of the storage credential to access the metastore storage_root. + updated_at: + type: integer + description: >- + Time at which the metastore was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the metastore. + ModelVersionInfo: + type: object + properties: + aliases: + type: array + items: + $ref: '#/components/schemas/RegisteredModelAlias' + catalog_name: + type: string + description: The name of the catalog containing the model version + comment: + type: string + description: The comment attached to the model version + created_at: + type: integer + created_by: + type: string + description: The identifier of the user who created the model version + id: + type: string + description: The unique identifier of the model version + metastore_id: + type: string + description: The unique identifier of the metastore containing the model version + model_name: + type: string + description: >- + The name of the parent registered model of the model version, + relative to parent schema + model_version_dependencies: + $ref: '#/components/schemas/DependencyList' + description: Model version dependencies, for feature-store packaged models + run_id: + type: string + description: >- + MLflow run ID used when creating the model version, if ``source`` + was generated by an experiment run stored in an MLflow tracking + server + run_workspace_id: + type: integer + description: >- + ID of the Databricks workspace containing the MLflow run that + generated this model version, if applicable + schema_name: + type: string + description: >- + The name of the schema containing the model version, relative to + parent catalog + source: + type: string + description: >- + URI indicating the location of the source artifacts (files) for the + model version + status: + $ref: '#/components/schemas/ModelVersionInfoStatus' + description: >- + Current status of the model version. Newly created model versions + start in PENDING_REGISTRATION status, then move to READY status once + the model version files are uploaded and the model version is + finalized. Only model versions in READY status can be loaded for + inference or served. + storage_location: + type: string + description: >- + The storage location on the cloud under which model version data + files are stored + updated_at: + type: integer + updated_by: + type: string + description: The identifier of the user who updated the model version last time + version: + type: integer + description: >- + Integer model version number, used to reference the model version in + API requests. + MonitorCronSchedule: + type: object + properties: + quartz_cron_expression: + type: string + timezone_id: + type: string + description: >- + The timezone id (e.g., ``PST``) in which to evaluate the quartz + expression. + pause_status: + $ref: '#/components/schemas/MonitorCronSchedulePauseStatus' + description: Read only field that indicates whether a schedule is paused or not. + required: + - quartz_cron_expression + - timezone_id + MonitorDataClassificationConfig: + type: object + properties: + enabled: + type: boolean + description: Whether to enable data classification. + description: Data classification related configuration. + MonitorDestination: + type: object + properties: + email_addresses: + type: array + items: + type: string + MonitorInferenceLog: + type: object + properties: + problem_type: + $ref: '#/components/schemas/MonitorInferenceLogProblemType' + timestamp_col: + type: string + description: Column for the timestamp. + granularities: + type: array + items: + type: string + description: >- + List of granularities to use when aggregating data into time windows + based on their timestamp. + prediction_col: + type: string + description: Column for the prediction. + model_id_col: + type: string + description: Column for the model identifier. + label_col: + type: string + description: Column for the label. + prediction_proba_col: + type: string + description: Column for prediction probabilities + required: + - problem_type + - timestamp_col + - granularities + - prediction_col + - model_id_col + MonitorInfo: + type: object + properties: + output_schema_name: + type: string + table_name: + type: string + description: >- + [Create:ERR Update:IGN] UC table to monitor. Format: + `catalog.schema.table_name` + status: + $ref: '#/components/schemas/MonitorInfoStatus' + description: '[Create:ERR Update:IGN] The monitor status.' + profile_metrics_table_name: + type: string + description: >- + [Create:ERR Update:IGN] Table that stores profile metrics data. + Format: `catalog.schema.table_name`. + drift_metrics_table_name: + type: string + description: >- + [Create:ERR Update:IGN] Table that stores drift metrics data. + Format: `catalog.schema.table_name`. + monitor_version: + type: integer + description: >- + [Create:ERR Update:IGN] Represents the current monitor configuration + version in use. The version will be represented in a numeric fashion + (1,2,3...). The field has flexibility to take on negative values, + which can indicate corrupted monitor_version numbers. + assets_dir: + type: string + description: >- + [Create:REQ Update:IGN] Field for specifying the absolute path to a + custom directory to store data-monitoring assets. Normally + prepopulated to a default user location via UI and Python APIs. + baseline_table_name: + type: string + description: >- + [Create:OPT Update:OPT] Baseline table name. Baseline data is used + to compute drift from the data in the monitored `table_name`. The + baseline table and the monitored table shall have the same schema. + custom_metrics: + type: array + items: + $ref: '#/components/schemas/MonitorMetric' + description: '[Create:OPT Update:OPT] Custom metrics.' + dashboard_id: + type: string + description: >- + [Create:ERR Update:OPT] Id of dashboard that visualizes the computed + metrics. This can be empty if the monitor is in PENDING state. + data_classification_config: + $ref: '#/components/schemas/MonitorDataClassificationConfig' + description: '[Create:OPT Update:OPT] Data classification related config.' + inference_log: + $ref: '#/components/schemas/MonitorInferenceLog' + latest_monitor_failure_msg: + type: string + description: >- + [Create:ERR Update:IGN] The latest error message for a monitor + failure. + notifications: + $ref: '#/components/schemas/MonitorNotifications' + description: '[Create:OPT Update:OPT] Field for specifying notification settings.' + schedule: + $ref: '#/components/schemas/MonitorCronSchedule' + description: '[Create:OPT Update:OPT] The monitor schedule.' + slicing_exprs: + type: array + items: + type: string + description: >- + [Create:OPT Update:OPT] List of column expressions to slice data + with for targeted analysis. The data is grouped by each expression + independently, resulting in a separate slice for each predicate and + its complements. For example `slicing_exprs=[“col_1”, “col_2 > 10”]` + will generate the following slices: two slices for `col_2 > 10` + (True and False), and one slice per unique value in `col1`. For + high-cardinality columns, only the top 100 unique values by + frequency will generate slices. + snapshot: + $ref: '#/components/schemas/MonitorSnapshot' + description: Configuration for monitoring snapshot tables. + time_series: + $ref: '#/components/schemas/MonitorTimeSeries' + description: Configuration for monitoring time series tables. + required: + - output_schema_name + - table_name + - status + - profile_metrics_table_name + - drift_metrics_table_name + - monitor_version + MonitorMetric: + type: object + properties: + name: + type: string + description: Name of the metric in the output tables. + definition: + type: string + description: >- + Jinja template for a SQL expression that specifies how to compute + the metric. See [create metric definition]. [create metric + definition]: + https://docs.databricks.com/en/lakehouse-monitoring/custom-metrics.html#create-definition + input_columns: + type: array + items: + type: string + description: >- + A list of column names in the input table the metric should be + computed for. Can use ``":table"`` to indicate that the metric needs + information from multiple columns. + output_data_type: + type: string + description: The output type of the custom metric. + type: + $ref: '#/components/schemas/MonitorMetricType' + description: >- + Can only be one of ``"CUSTOM_METRIC_TYPE_AGGREGATE"``, + ``"CUSTOM_METRIC_TYPE_DERIVED"``, or ``"CUSTOM_METRIC_TYPE_DRIFT"``. + The ``"CUSTOM_METRIC_TYPE_AGGREGATE"`` and + ``"CUSTOM_METRIC_TYPE_DERIVED"`` metrics are computed on a single + table, whereas the ``"CUSTOM_METRIC_TYPE_DRIFT"`` compare metrics + across baseline and input table, or across the two consecutive time + windows. - CUSTOM_METRIC_TYPE_AGGREGATE: only depend on the existing + columns in your table - CUSTOM_METRIC_TYPE_DERIVED: depend on + previously computed aggregate metrics - CUSTOM_METRIC_TYPE_DRIFT: + depend on previously computed aggregate or derived metrics + required: + - name + - definition + - input_columns + - output_data_type + - type + description: Custom metric definition. + MonitorNotifications: + type: object + properties: + on_failure: + $ref: '#/components/schemas/MonitorDestination' + on_new_classification_tag_detected: + $ref: '#/components/schemas/MonitorDestination' + description: >- + Destinations to send notifications on new classification tag + detected. + MonitorRefreshInfo: + type: object + properties: + refresh_id: + type: integer + state: + $ref: '#/components/schemas/MonitorRefreshInfoState' + description: The current state of the refresh. + start_time_ms: + type: integer + description: >- + Time at which refresh operation was initiated (milliseconds since + 1/1/1970 UTC). + end_time_ms: + type: integer + description: >- + Time at which refresh operation completed (milliseconds since + 1/1/1970 UTC). + message: + type: string + description: >- + An optional message to give insight into the current state of the + job (e.g. FAILURE messages). + trigger: + $ref: '#/components/schemas/MonitorRefreshInfoTrigger' + description: The method by which the refresh was triggered. + required: + - refresh_id + - state + - start_time_ms + MonitorRefreshListResponse: + type: object + properties: + refreshes: + type: array + items: + $ref: '#/components/schemas/MonitorRefreshInfo' + MonitorSnapshot: + type: object + properties: {} + description: Snapshot analysis configuration + MonitorTimeSeries: + type: object + properties: + timestamp_col: + type: string + description: Column for the timestamp. + granularities: + type: array + items: + type: string + description: >- + Granularities for aggregating data into time windows based on their + timestamp. Currently the following static granularities are + supported: {``\"5 minutes\"``, ``\"30 minutes\"``, ``\"1 hour\"``, + ``\"1 day\"``, ``\"\u003cn\u003e week(s)\"``, ``\"1 month\"``, ``\"1 + year\"``}. + required: + - timestamp_col + - granularities + description: Time series analysis configuration. + NamedTableConstraint: + type: object + properties: + name: + type: string + required: + - name + NotificationDestination: + type: object + properties: + destination_id: + type: string + destination_type: + $ref: '#/components/schemas/DestinationType' + description: The type of the destination. + special_destination: + $ref: '#/components/schemas/SpecialDestination' + description: >- + This field is used to denote whether the destination is the email of + the owner of the securable object. The special destination cannot be + assigned to a securable and only represents the default destination + of the securable. The securable types that support default special + destinations are: "catalog", "external_location", "connection", + "credential", and "metastore". The **destination_type** of a + **special_destination** is always EMAIL. + OnlineTable: + type: object + properties: + name: + type: string + description: Full three-part (catalog, schema, table) name of the table. + spec: + $ref: '#/components/schemas/OnlineTableSpec' + description: Specification of the online table. + status: + $ref: '#/components/schemas/OnlineTableStatus' + description: Online Table data synchronization status + table_serving_url: + type: string + description: Data serving REST API URL for this table + unity_catalog_provisioning_state: + $ref: '#/components/schemas/ProvisioningInfoState' + description: >- + The provisioning state of the online table entity in Unity Catalog. + This is distinct from the state of the data synchronization pipeline + (i.e. the table may be in "ACTIVE" but the pipeline may be in + "PROVISIONING" as it runs asynchronously). + description: Online Table information. + OnlineTableSpec: + type: object + properties: + perform_full_copy: + type: boolean + description: >- + Whether to create a full-copy pipeline -- a pipeline that stops + after creates a full copy of the source table upon initialization + and does not process any change data feeds (CDFs) afterwards. The + pipeline can still be manually triggered afterwards, but it always + perform a full copy of the source table and there are no incremental + updates. This mode is useful for syncing views or tables without + CDFs to online tables. Note that the full-copy pipeline only + supports "triggered" scheduling policy. + pipeline_id: + type: string + description: >- + ID of the associated pipeline. Generated by the server - cannot be + set by the caller. + primary_key_columns: + type: array + items: + type: string + description: >- + Primary Key columns to be used for data insert/update in the + destination. + run_continuously: + $ref: '#/components/schemas/OnlineTableSpecContinuousSchedulingPolicy' + description: Pipeline runs continuously after generating the initial data. + run_triggered: + $ref: '#/components/schemas/OnlineTableSpecTriggeredSchedulingPolicy' + description: >- + Pipeline stops after generating the initial data and can be + triggered later (manually, through a cron job or through data + triggers) + source_table_full_name: + type: string + description: Three-part (catalog, schema, table) name of the source Delta table. + timeseries_key: + type: string + description: >- + Time series key to deduplicate (tie-break) rows with the same + primary key. + description: Specification of an online table. + OnlineTableSpecContinuousSchedulingPolicy: + type: object + properties: {} + OnlineTableSpecTriggeredSchedulingPolicy: + type: object + properties: {} + OnlineTableStatus: + type: object + properties: + continuous_update_status: + $ref: '#/components/schemas/ContinuousUpdateStatus' + detailed_state: + $ref: '#/components/schemas/OnlineTableState' + description: The state of the online table. + failed_status: + $ref: '#/components/schemas/FailedStatus' + message: + type: string + description: A text description of the current state of the online table. + provisioning_status: + $ref: '#/components/schemas/ProvisioningStatus' + triggered_update_status: + $ref: '#/components/schemas/TriggeredUpdateStatus' + description: Status of an online table. + OptionSpec: + type: object + properties: + allowed_values: + type: array + items: + type: string + description: >- + For drop down / radio button selections, UI will want to know the + possible input values, it can also be used by other option types to + limit input selections. + default_value: + type: string + description: >- + The default value of the option, for example, value '443' for 'port' + option. + description: + type: string + description: >- + A concise user facing description of what the input value of this + option should look like. + hint: + type: string + description: >- + The hint is used on the UI to suggest what the input value can + possibly be like, for example: example.com for 'host' option. Unlike + default value, it will not be applied automatically without user + input. + is_copiable: + type: boolean + description: >- + Indicates whether an option should be displayed with copy button on + the UI. + is_creatable: + type: boolean + description: >- + Indicates whether an option can be provided by users in the + create/update path of an entity. + is_hidden: + type: boolean + description: >- + Is the option value not user settable and is thus not shown on the + UI. + is_loggable: + type: boolean + description: >- + Specifies whether this option is safe to log, i.e. no sensitive + information. + is_required: + type: boolean + description: Is the option required. + is_secret: + type: boolean + description: Is the option value considered secret and thus redacted on the UI. + is_updatable: + type: boolean + description: Is the option updatable by users. + name: + type: string + description: The unique name of the option. + oauth_stage: + $ref: '#/components/schemas/OptionSpecOauthStage' + description: >- + Specifies when the option value is displayed on the UI within the + OAuth flow. + type: + $ref: '#/components/schemas/OptionSpecOptionType' + description: The type of the option. + description: >- + Spec of an allowed option on a securable kind and its attributes. This + is mostly used by UI to + provide user friendly hints and descriptions in order to facilitate the securable creation + process. + PermissionsChange: + type: object + properties: + add: + type: array + items: + $ref: '#/components/schemas/Privilege' + principal: + type: string + description: >- + The principal whose privileges we are changing. Only one of + principal or principal_id should be specified, never both at the + same time. + remove: + type: array + items: + $ref: '#/components/schemas/Privilege' + description: The set of privileges to remove. + PipelineProgress: + type: object + properties: + estimated_completion_time_seconds: + type: number + description: The estimated time remaining to complete this update in seconds. + latest_version_currently_processing: + type: integer + description: >- + The source table Delta version that was last processed by the + pipeline. The pipeline may not have completely processed this + version yet. + sync_progress_completion: + type: number + description: >- + The completion ratio of this update. This is a number between 0 and + 1. + synced_row_count: + type: integer + description: The number of rows that have been synced in this update. + total_row_count: + type: integer + description: >- + The total number of rows that need to be synced in this update. This + number may be an estimate. + description: Progress information of the Online Table data synchronization pipeline. + PolicyInfo: + type: object + properties: + to_principals: + type: array + items: + type: string + for_securable_type: + $ref: '#/components/schemas/SecurableType' + description: >- + Type of securables that the policy should take effect on. Only + `TABLE` is supported at this moment. Required on create and optional + on update. + policy_type: + $ref: '#/components/schemas/PolicyType' + description: Type of the policy. Required on create. + column_mask: + $ref: '#/components/schemas/ColumnMaskOptions' + description: >- + Options for column mask policies. Valid only if `policy_type` is + `POLICY_TYPE_COLUMN_MASK`. Required on create and optional on + update. When specified on update, the new options will replace the + existing options as a whole. + comment: + type: string + description: Optional description of the policy. + created_at: + type: integer + description: >- + Time at which the policy was created, in epoch milliseconds. Output + only. + created_by: + type: string + description: Username of the user who created the policy. Output only. + except_principals: + type: array + items: + type: string + description: >- + Optional list of user or group names that should be excluded from + the policy. + id: + type: string + description: >- + Unique identifier of the policy. This field is output only and is + generated by the system. + match_columns: + type: array + items: + $ref: '#/components/schemas/MatchColumn' + description: >- + Optional list of condition expressions used to match table columns. + Only valid when `for_securable_type` is `TABLE`. When specified, the + policy only applies to tables whose columns satisfy all match + conditions. + name: + type: string + description: >- + Name of the policy. Required on create and optional on update. To + rename the policy, set `name` to a different value on update. + on_securable_fullname: + type: string + description: >- + Full name of the securable on which the policy is defined. Required + on create. + on_securable_type: + $ref: '#/components/schemas/SecurableType' + description: >- + Type of the securable on which the policy is defined. Only + `CATALOG`, `SCHEMA` and `TABLE` are supported at this moment. + Required on create. + row_filter: + $ref: '#/components/schemas/RowFilterOptions' + description: >- + Options for row filter policies. Valid only if `policy_type` is + `POLICY_TYPE_ROW_FILTER`. Required on create and optional on update. + When specified on update, the new options will replace the existing + options as a whole. + updated_at: + type: integer + description: >- + Time at which the policy was last modified, in epoch milliseconds. + Output only. + updated_by: + type: string + description: Username of the user who last modified the policy. Output only. + when_condition: + type: string + description: Optional condition when the policy should take effect. + required: + - to_principals + - for_securable_type + - policy_type + PrimaryKeyConstraint: + type: object + properties: + name: + type: string + child_columns: + type: array + items: + type: string + description: Column names for this constraint. + rely: + type: boolean + description: True if the constraint is RELY, false or unset if NORELY. + timeseries_columns: + type: array + items: + type: string + description: Column names that represent a timeseries. + required: + - name + - child_columns + Principal: + type: object + properties: + id: + type: string + principal_type: + $ref: '#/components/schemas/PrincipalType' + PrivilegeAssignment: + type: object + properties: + principal: + type: string + privileges: + type: array + items: + $ref: '#/components/schemas/Privilege' + description: The privileges assigned to the principal. + ProvisioningInfo: + type: object + properties: + state: + $ref: '#/components/schemas/ProvisioningInfoState' + description: The provisioning state of the resource. + description: Status of an asynchronously provisioned resource. + ProvisioningStatus: + type: object + properties: + initial_pipeline_sync_progress: + $ref: '#/components/schemas/PipelineProgress' + description: >- + Details about initial data synchronization. Only populated when in + the PROVISIONING_INITIAL_SNAPSHOT state. + description: |- + Detailed status of an online table. Shown if the online table is in the + PROVISIONING_PIPELINE_RESOURCES or the PROVISIONING_INITIAL_SNAPSHOT state. + QuotaInfo: + type: object + properties: + last_refreshed_at: + type: integer + parent_full_name: + type: string + description: >- + Name of the parent resource. Returns metastore ID if the parent is a + metastore. + parent_securable_type: + $ref: '#/components/schemas/SecurableType' + description: The quota parent securable type. + quota_count: + type: integer + description: The current usage of the resource quota. + quota_limit: + type: integer + description: The current limit of the resource quota. + quota_name: + type: string + description: The name of the quota. + R2Credentials: + type: object + properties: + access_key_id: + type: string + description: The access key ID that identifies the temporary credentials. + secret_access_key: + type: string + description: The secret access key associated with the access key. + session_token: + type: string + description: >- + The generated JWT that users must pass to use the temporary + credentials. + description: |- + R2 temporary credentials for API authentication. Read more at + https://developers.cloudflare.com/r2/api/s3/tokens/. + RegenerateDashboardResponse: + type: object + properties: + dashboard_id: + type: string + parent_folder: + type: string + description: Parent folder is equivalent to {assets_dir}/{tableName} + RegisteredModelAlias: + type: object + properties: + alias_name: + type: string + catalog_name: + type: string + description: The name of the catalog containing the model version + id: + type: string + description: The unique identifier of the alias + model_name: + type: string + description: >- + The name of the parent registered model of the model version, + relative to parent schema + schema_name: + type: string + description: >- + The name of the schema containing the model version, relative to + parent catalog + version_num: + type: integer + description: >- + Integer version number of the model version to which this alias + points. + RegisteredModelInfo: + type: object + properties: + aliases: + type: array + items: + $ref: '#/components/schemas/RegisteredModelAlias' + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: >- + The name of the catalog where the schema and the registered model + reside + comment: + type: string + description: The comment attached to the registered model + created_at: + type: integer + description: >- + Creation timestamp of the registered model in milliseconds since the + Unix epoch + created_by: + type: string + description: The identifier of the user who created the registered model + full_name: + type: string + description: The three-level (fully qualified) name of the registered model + metastore_id: + type: string + description: The unique identifier of the metastore + name: + type: string + description: The name of the registered model + owner: + type: string + description: The identifier of the user who owns the registered model + schema_name: + type: string + description: The name of the schema where the registered model resides + storage_location: + type: string + description: >- + The storage location on the cloud under which model version data + files are stored + updated_at: + type: integer + description: >- + Last-update timestamp of the registered model in milliseconds since + the Unix epoch + updated_by: + type: string + description: >- + The identifier of the user who updated the registered model last + time + RowFilterOptions: + type: object + properties: + function_name: + type: string + using: + type: array + items: + $ref: '#/components/schemas/FunctionArgument' + description: >- + Optional list of column aliases or constant literals to be passed as + arguments to the row filter function. The type of each column should + match the positional argument of the row filter function. + required: + - function_name + SchemaInfo: + type: object + properties: + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: Name of parent catalog. + catalog_type: + $ref: '#/components/schemas/CatalogType' + description: The type of the parent catalog. + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: Time at which this schema was created, in epoch milliseconds. + created_by: + type: string + description: Username of schema creator. + effective_predictive_optimization_flag: + $ref: '#/components/schemas/EffectivePredictiveOptimizationFlag' + enable_predictive_optimization: + $ref: '#/components/schemas/EnablePredictiveOptimization' + description: >- + Whether predictive optimization should be enabled for this object + and objects under it. + full_name: + type: string + description: Full name of schema, in form of __catalog_name__.__schema_name__. + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of schema, relative to parent catalog. + owner: + type: string + description: Username of current owner of schema. + properties: + type: object + description: A map of key-value properties attached to the securable. + schema_id: + type: string + description: The unique identifier of the schema. + storage_location: + type: string + description: Storage location for managed tables within schema. + storage_root: + type: string + description: Storage root URL for managed tables within schema. + updated_at: + type: integer + description: Time at which this schema was created, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified schema. + description: 'Next ID: 45' + Securable: + type: object + properties: + full_name: + type: string + description: >- + Required. The full name of the catalog/schema/table. Optional if + resource_name is present. + provider_share: + type: string + description: >- + Optional. The name of the Share object that contains the securable + when the securable is getting shared in D2D Delta Sharing. + type: + $ref: '#/components/schemas/SecurableType' + description: >- + Required. The type of securable (catalog/schema/table). Optional if + resource_name is present. + description: >- + Generic definition of a securable, which is uniquely defined in a + metastore by its type and full + name. + SecurableKindManifest: + type: object + properties: + assignable_privileges: + type: array + items: + type: string + description: Privileges that can be assigned to the securable. + capabilities: + type: array + items: + type: string + description: A list of capabilities in the securable kind. + options: + type: array + items: + $ref: '#/components/schemas/OptionSpec' + description: Detailed specs of allowed options. + securable_kind: + $ref: '#/components/schemas/SecurableKind' + description: Securable kind to get manifest of. + securable_type: + $ref: '#/components/schemas/SecurableType' + description: Securable Type of the kind. + description: Manifest of a specific securable kind. + SecurablePermissions: + type: object + properties: + permissions: + type: array + items: + type: string + securable: + $ref: '#/components/schemas/Securable' + description: >- + The securable for which the access request destinations are being + requested. + SseEncryptionDetails: + type: object + properties: + algorithm: + $ref: '#/components/schemas/SseEncryptionDetailsAlgorithm' + description: >- + Sets the value of the 'x-amz-server-side-encryption' header in S3 + request. + aws_kms_key_arn: + type: string + description: >- + Optional. The ARN of the SSE-KMS key used with the S3 location, when + algorithm = "SSE-KMS". Sets the value of the + 'x-amz-server-side-encryption-aws-kms-key-id' header. + description: Server-Side Encryption properties for clients communicating with AWS s3. + StorageCredentialInfo: + type: object + properties: + aws_iam_role: + $ref: '#/components/schemas/AwsIamRoleResponse' + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentityResponse' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + cloudflare_api_token: + $ref: '#/components/schemas/CloudflareApiToken' + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + created_at: + type: integer + description: Time at which this credential was created, in epoch milliseconds. + created_by: + type: string + description: Username of credential creator. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccountResponse' + description: The Databricks managed GCP service account configuration. + full_name: + type: string + description: The full name of the credential. + id: + type: string + description: The unique identifier of the credential. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + metastore_id: + type: string + description: Unique identifier of the parent metastore. + name: + type: string + description: >- + The credential name. The name must be unique among storage and + service credentials within the metastore. + owner: + type: string + description: Username of current owner of credential. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + updated_at: + type: integer + description: >- + Time at which this credential was last modified, in epoch + milliseconds. + updated_by: + type: string + description: Username of user who last modified the credential. + used_for_managed_storage: + type: boolean + description: >- + Whether this credential is the current metastore's root storage + credential. Only applicable when purpose is **STORAGE**. + SystemSchemaInfo: + type: object + properties: + schema: + type: string + state: + type: string + description: >- + The current state of enablement for the system schema. An empty + string means the system schema is available and ready for opt-in. + Possible values: AVAILABLE | ENABLE_INITIALIZED | ENABLE_COMPLETED | + DISABLE_INITIALIZED | UNAVAILABLE | MANAGED + required: + - schema + - state + TableConstraint: + type: object + properties: + foreign_key_constraint: + $ref: '#/components/schemas/ForeignKeyConstraint' + named_table_constraint: + $ref: '#/components/schemas/NamedTableConstraint' + primary_key_constraint: + $ref: '#/components/schemas/PrimaryKeyConstraint' + description: >- + A table constraint, as defined by *one* of the following fields being + set: + __primary_key_constraint__, __foreign_key_constraint__, __named_table_constraint__. + TableDependency: + type: object + properties: + table_full_name: + type: string + description: >- + Full name of the dependent table, in the form of + __catalog_name__.__schema_name__.__table_name__. + required: + - table_full_name + description: A table that is dependent on a SQL object. + TableExistsResponse: + type: object + properties: + table_exists: + type: boolean + TableInfo: + type: object + properties: + access_point: + type: string + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: Name of parent catalog. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnInfo' + description: The array of __ColumnInfo__ definitions of the table's columns. + comment: + type: string + description: User-provided free-form text description. + created_at: + type: integer + description: Time at which this table was created, in epoch milliseconds. + created_by: + type: string + description: Username of table creator. + data_access_configuration_id: + type: string + description: >- + Unique ID of the Data Access Configuration to use with the table + data. + data_source_format: + $ref: '#/components/schemas/DataSourceFormat' + deleted_at: + type: integer + description: >- + Time at which this table was deleted, in epoch milliseconds. Field + is omitted if table is not deleted. + delta_runtime_properties_kvpairs: + $ref: '#/components/schemas/DeltaRuntimePropertiesKvPairs' + description: Information pertaining to current state of the delta table. + effective_predictive_optimization_flag: + $ref: '#/components/schemas/EffectivePredictiveOptimizationFlag' + enable_predictive_optimization: + $ref: '#/components/schemas/EnablePredictiveOptimization' + encryption_details: + $ref: '#/components/schemas/EncryptionDetails' + full_name: + type: string + description: >- + Full name of table, in form of + __catalog_name__.__schema_name__.__table_name__ + metastore_id: + type: string + description: Unique identifier of parent metastore. + name: + type: string + description: Name of table, relative to parent schema. + owner: + type: string + description: Username of current owner of table. + pipeline_id: + type: string + description: >- + The pipeline ID of the table. Applicable for tables created by + pipelines (Materialized View, Streaming Table, etc.). + properties: + type: object + description: A map of key-value properties attached to the securable. + row_filter: + $ref: '#/components/schemas/TableRowFilter' + schema_name: + type: string + description: Name of parent schema relative to its parent catalog. + securable_kind_manifest: + $ref: '#/components/schemas/SecurableKindManifest' + description: >- + SecurableKindManifest of table, including capabilities the table + has. + sql_path: + type: string + description: >- + List of schemes whose objects can be referenced without + qualification. + storage_credential_name: + type: string + description: >- + Name of the storage credential, when a storage credential is + configured for use with this table. + storage_location: + type: string + description: Storage root URL for table (for **MANAGED**, **EXTERNAL** tables). + table_constraints: + type: array + items: + $ref: '#/components/schemas/TableConstraint' + description: >- + List of table constraints. Note: this field is not set in the output + of the __listTables__ API. + table_id: + type: string + description: The unique identifier of the table. + table_type: + $ref: '#/components/schemas/TableType' + updated_at: + type: integer + description: Time at which this table was last modified, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified the table. + view_definition: + type: string + description: >- + View definition SQL (when __table_type__ is **VIEW**, + **MATERIALIZED_VIEW**, or **STREAMING_TABLE**) + view_dependencies: + $ref: '#/components/schemas/DependencyList' + description: >- + View dependencies (when table_type == **VIEW** or + **MATERIALIZED_VIEW**, **STREAMING_TABLE**) - when DependencyList is + None, the dependency is not provided; - when DependencyList is an + empty list, the dependency is provided but is empty; - when + DependencyList is not an empty list, dependencies are provided and + recorded. Note: this field is not set in the output of the + __listTables__ API. + TableRowFilter: + type: object + properties: + function_name: + type: string + input_column_names: + type: array + items: + type: string + description: >- + The list of table columns to be passed as input to the row filter + function. The column types should match the types of the filter + function arguments. + required: + - function_name + - input_column_names + TableSummary: + type: object + properties: + full_name: + type: string + securable_kind_manifest: + $ref: '#/components/schemas/SecurableKindManifest' + description: >- + SecurableKindManifest of table, including capabilities the table + has. + table_type: + $ref: '#/components/schemas/TableType' + TagKeyValue: + type: object + properties: + key: + type: string + value: + type: string + description: value of the tag associated with the key, could be optional + TemporaryCredentials: + type: object + properties: + aws_temp_credentials: + $ref: '#/components/schemas/AwsCredentials' + azure_aad: + $ref: '#/components/schemas/AzureActiveDirectoryToken' + expiration_time: + type: integer + description: >- + Server time when the credential will expire, in epoch milliseconds. + The API client is advised to cache the credential given this + expiration time. + gcp_oauth_token: + $ref: '#/components/schemas/GcpOauthToken' + TriggeredUpdateStatus: + type: object + properties: + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was synced to the online + table. Note that this Delta version may not be completely synced to + the online table yet. + timestamp: + type: string + description: >- + The timestamp of the last time any data was synchronized from the + source table to the online table. + triggered_update_progress: + $ref: '#/components/schemas/PipelineProgress' + description: Progress of the active data synchronization pipeline. + description: >- + Detailed status of an online table. Shown if the online table is in the + ONLINE_TRIGGERED_UPDATE + or the ONLINE_NO_PENDING_UPDATE state. + UnassignResponse: + type: object + properties: {} + UpdateAccountsMetastore: + type: object + properties: + delta_sharing_organization_name: + type: string + delta_sharing_recipient_token_lifetime_in_seconds: + type: integer + description: The lifetime of delta sharing recipient token in seconds. + delta_sharing_scope: + $ref: '#/components/schemas/DeltaSharingScopeEnum' + description: The scope of Delta Sharing enabled for the metastore. + external_access_enabled: + type: boolean + description: >- + Whether to allow non-DBR clients to directly access entities under + the metastore. + owner: + type: string + description: The owner of the metastore. + privilege_model_version: + type: string + description: >- + Privilege model version of the metastore, of the form `major.minor` + (e.g., `1.0`). + storage_root_credential_id: + type: string + description: UUID of storage credential to access the metastore storage_root. + UpdateAccountsStorageCredential: + type: object + properties: + aws_iam_role: + $ref: '#/components/schemas/AwsIamRoleRequest' + azure_managed_identity: + $ref: '#/components/schemas/AzureManagedIdentityResponse' + description: The Azure managed identity configuration. + azure_service_principal: + $ref: '#/components/schemas/AzureServicePrincipal' + description: The Azure service principal configuration. + cloudflare_api_token: + $ref: '#/components/schemas/CloudflareApiToken' + description: The Cloudflare API token configuration. + comment: + type: string + description: Comment associated with the credential. + databricks_gcp_service_account: + $ref: '#/components/schemas/DatabricksGcpServiceAccountRequest' + description: The Databricks managed GCP service account configuration. + isolation_mode: + $ref: '#/components/schemas/IsolationMode' + description: >- + Whether the current securable is accessible from all workspaces or a + specific set of workspaces. + owner: + type: string + description: Username of current owner of credential. + read_only: + type: boolean + description: >- + Whether the credential is usable only for read operations. Only + applicable when purpose is **STORAGE**. + UpdateAssignmentResponse: + type: object + properties: {} + UpdateCatalogWorkspaceBindingsResponse: + type: object + properties: + workspaces: + type: array + items: + type: integer + UpdateMetastoreAssignment: + type: object + properties: + workspace_id: + type: integer + default_catalog_name: + type: string + description: >- + The name of the default catalog in the metastore. This field is + deprecated. Please use "Default Namespace API" to configure the + default catalog for a Databricks workspace. + metastore_id: + type: string + description: The unique ID of the metastore. + required: + - workspace_id + UpdatePermissionsResponse: + type: object + properties: + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/PrivilegeAssignment' + UpdateRequestExternalLineage: + type: object + properties: + source: + $ref: '#/components/schemas/ExternalLineageObject' + target: + $ref: '#/components/schemas/ExternalLineageObject' + description: Target object of the external lineage relationship. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnRelationship' + description: List of column relationships between source and target objects. + id: + type: string + description: Unique identifier of the external lineage relationship. + properties: + type: object + description: >- + Key-value properties associated with the external lineage + relationship. + required: + - source + - target + UpdateResponse: + type: object + properties: {} + UpdateWorkspaceBindingsResponse: + type: object + properties: + bindings: + type: array + items: + $ref: '#/components/schemas/WorkspaceBinding' + description: List of workspace bindings. + description: A list of workspace IDs that are bound to the securable + ValidateCredentialResponse: + type: object + properties: + isDir: + type: boolean + results: + type: array + items: + $ref: '#/components/schemas/CredentialValidationResult' + description: The results of the validation check. + ValidateStorageCredentialResponse: + type: object + properties: + isDir: + type: boolean + results: + type: array + items: + $ref: '#/components/schemas/ValidationResult' + description: The results of the validation check. + ValidationResult: + type: object + properties: + message: + type: string + operation: + $ref: '#/components/schemas/ValidationResultOperation' + description: The operation tested. + result: + $ref: '#/components/schemas/ValidationResultResult' + description: The results of the tested operation. + VolumeInfo: + type: object + properties: + access_point: + type: string + browse_only: + type: boolean + description: >- + Indicates whether the principal is limited to retrieving metadata + for the associated object through the BROWSE privilege when + include_browse is enabled in the request. + catalog_name: + type: string + description: The name of the catalog where the schema and the volume are + comment: + type: string + description: The comment attached to the volume + created_at: + type: integer + created_by: + type: string + description: The identifier of the user who created the volume + encryption_details: + $ref: '#/components/schemas/EncryptionDetails' + full_name: + type: string + description: The three-level (fully qualified) name of the volume + metastore_id: + type: string + description: The unique identifier of the metastore + name: + type: string + description: The name of the volume + owner: + type: string + description: The identifier of the user who owns the volume + schema_name: + type: string + description: The name of the schema where the volume is + storage_location: + type: string + description: The storage location on the cloud + updated_at: + type: integer + updated_by: + type: string + description: The identifier of the user who updated the volume last time + volume_id: + type: string + description: The unique identifier of the volume + volume_type: + $ref: '#/components/schemas/VolumeType' + description: >- + The type of the volume. An external volume is located in the + specified external location. A managed volume is located in the + default location which is specified by the parent schema, or the + parent catalog, or the Metastore. [Learn more] [Learn more]: + https://docs.databricks.com/aws/en/volumes/managed-vs-external + WorkspaceBinding: + type: object + properties: + workspace_id: + type: integer + binding_type: + $ref: '#/components/schemas/WorkspaceBindingBindingType' + description: One of READ_WRITE/READ_ONLY. Default is READ_WRITE. + required: + - workspace_id + ArtifactType: + type: string + enum: + - INIT_SCRIPT + - LIBRARY_JAR + - LIBRARY_MAVEN + description: The artifact type + CatalogIsolationMode: + type: string + enum: + - ISOLATED + - OPEN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CatalogType: + type: string + enum: + - DELTASHARING_CATALOG + - FOREIGN_CATALOG + - INTERNAL_CATALOG + - MANAGED_CATALOG + - MANAGED_ONLINE_CATALOG + - SYSTEM_CATALOG + description: The type of the catalog. + ColumnTypeName: + type: string + enum: + - ARRAY + - BINARY + - BOOLEAN + - BYTE + - CHAR + - DATE + - DECIMAL + - DOUBLE + - FLOAT + - GEOGRAPHY + - GEOMETRY + - INT + - INTERVAL + - LONG + - MAP + - 'NULL' + - SHORT + - STRING + - STRUCT + - TABLE_TYPE + - TIMESTAMP + - TIMESTAMP_NTZ + - USER_DEFINED_TYPE + - VARIANT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ConnectionType: + type: string + enum: + - BIGQUERY + - DATABRICKS + - GA4_RAW_DATA + - GLUE + - HIVE_METASTORE + - HTTP + - MYSQL + - ORACLE + - POSTGRESQL + - POWER_BI + - REDSHIFT + - SALESFORCE + - SALESFORCE_DATA_CLOUD + - SERVICENOW + - SNOWFLAKE + - SQLDW + - SQLSERVER + - TERADATA + - UNKNOWN_CONNECTION_TYPE + - WORKDAY_RAAS + description: 'Next Id: 72' + CreateFunctionParameterStyle: + type: string + enum: + - S + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CreateFunctionRoutineBody: + type: string + enum: + - EXTERNAL + - SQL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CreateFunctionSecurityType: + type: string + enum: + - DEFINER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CreateFunctionSqlDataAccess: + type: string + enum: + - CONTAINS_SQL + - NO_SQL + - READS_SQL_DATA + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CredentialPurpose: + type: string + enum: + - SERVICE + - STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CredentialType: + type: string + enum: + - ANY_STATIC_CREDENTIAL + - BEARER_TOKEN + - OAUTH_ACCESS_TOKEN + - OAUTH_M2M + - OAUTH_MTLS + - OAUTH_REFRESH_TOKEN + - OAUTH_RESOURCE_OWNER_PASSWORD + - OAUTH_U2M + - OAUTH_U2M_MAPPING + - OIDC_TOKEN + - PEM_PRIVATE_KEY + - SERVICE_CREDENTIAL + - SSWS_TOKEN + - UNKNOWN_CREDENTIAL_TYPE + - USERNAME_PASSWORD + description: 'Next Id: 15' + DataSourceFormat: + type: string + enum: + - AVRO + - BIGQUERY_FORMAT + - CSV + - DATABRICKS_FORMAT + - DATABRICKS_ROW_STORE_FORMAT + - DELTA + - DELTASHARING + - DELTA_UNIFORM_HUDI + - DELTA_UNIFORM_ICEBERG + - HIVE + - ICEBERG + - JSON + - MONGODB_FORMAT + - MYSQL_FORMAT + - NETSUITE_FORMAT + - ORACLE_FORMAT + - ORC + - PARQUET + - POSTGRESQL_FORMAT + - REDSHIFT_FORMAT + - SALESFORCE_DATA_CLOUD_FORMAT + - SALESFORCE_FORMAT + - SNOWFLAKE_FORMAT + - SQLDW_FORMAT + - SQLSERVER_FORMAT + - TERADATA_FORMAT + - TEXT + - UNITY_CATALOG + - VECTOR_INDEX_FORMAT + - WORKDAY_RAAS_FORMAT + description: Data source format + DeltaSharingScopeEnum: + type: string + enum: + - INTERNAL + - INTERNAL_AND_EXTERNAL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DestinationType: + type: string + enum: + - EMAIL + - GENERIC_WEBHOOK + - MICROSOFT_TEAMS + - SLACK + - URL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EffectivePredictiveOptimizationFlagInheritedFromType: + type: string + enum: + - CATALOG + - SCHEMA + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EnablePredictiveOptimization: + type: string + enum: + - DISABLE + - ENABLE + - INHERIT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoParameterStyle: + type: string + enum: + - S + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoRoutineBody: + type: string + enum: + - EXTERNAL + - SQL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoSecurityType: + type: string + enum: + - DEFINER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionInfoSqlDataAccess: + type: string + enum: + - CONTAINS_SQL + - NO_SQL + - READS_SQL_DATA + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionParameterMode: + type: string + enum: + - IN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionParameterType: + type: string + enum: + - COLUMN + - PARAM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + IsolationMode: + type: string + enum: + - ISOLATION_MODE_ISOLATED + - ISOLATION_MODE_OPEN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + LineageDirection: + type: string + enum: + - DOWNSTREAM + - UPSTREAM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MatchType: + type: string + enum: + - PREFIX_MATCH + description: The artifact pattern matching type + ModelVersionInfoStatus: + type: string + enum: + - FAILED_REGISTRATION + - MODEL_VERSION_STATUS_UNKNOWN + - PENDING_REGISTRATION + - READY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MonitorCronSchedulePauseStatus: + type: string + enum: + - PAUSED + - UNPAUSED + - UNSPECIFIED + description: >- + Source link: + + https://src.dev.databricks.com/databricks/universe/-/blob/elastic-spark-common/api/messages/schedule.proto + + Monitoring workflow schedule pause status. + MonitorInferenceLogProblemType: + type: string + enum: + - PROBLEM_TYPE_CLASSIFICATION + - PROBLEM_TYPE_REGRESSION + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MonitorInfoStatus: + type: string + enum: + - MONITOR_STATUS_ACTIVE + - MONITOR_STATUS_DELETE_PENDING + - MONITOR_STATUS_ERROR + - MONITOR_STATUS_FAILED + - MONITOR_STATUS_PENDING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MonitorMetricType: + type: string + enum: + - CUSTOM_METRIC_TYPE_AGGREGATE + - CUSTOM_METRIC_TYPE_DERIVED + - CUSTOM_METRIC_TYPE_DRIFT + description: >- + Can only be one of ``"CUSTOM_METRIC_TYPE_AGGREGATE"``, + ``"CUSTOM_METRIC_TYPE_DERIVED"``, or + + ``"CUSTOM_METRIC_TYPE_DRIFT"``. The ``"CUSTOM_METRIC_TYPE_AGGREGATE"`` + and + + ``"CUSTOM_METRIC_TYPE_DERIVED"`` metrics are computed on a single table, + whereas the + + ``"CUSTOM_METRIC_TYPE_DRIFT"`` compare metrics across baseline and input + table, or across the + + two consecutive time windows. - CUSTOM_METRIC_TYPE_AGGREGATE: only + depend on the existing + + columns in your table - CUSTOM_METRIC_TYPE_DERIVED: depend on previously + computed aggregate + + metrics - CUSTOM_METRIC_TYPE_DRIFT: depend on previously computed + aggregate or derived metrics + MonitorRefreshInfoState: + type: string + enum: + - CANCELED + - FAILED + - PENDING + - RUNNING + - SUCCESS + - UNKNOWN + description: The current state of the refresh. + MonitorRefreshInfoTrigger: + type: string + enum: + - MANUAL + - SCHEDULE + - UNKNOWN_TRIGGER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + OnlineTableState: + type: string + enum: + - OFFLINE + - OFFLINE_FAILED + - ONLINE + - ONLINE_CONTINUOUS_UPDATE + - ONLINE_NO_PENDING_UPDATE + - ONLINE_PIPELINE_FAILED + - ONLINE_TRIGGERED_UPDATE + - ONLINE_UPDATING_PIPELINE_RESOURCES + - PROVISIONING + - PROVISIONING_INITIAL_SNAPSHOT + - PROVISIONING_PIPELINE_RESOURCES + description: The state of an online table. + OptionSpecOauthStage: + type: string + enum: + - BEFORE_ACCESS_TOKEN + - BEFORE_AUTHORIZATION_CODE + description: >- + During the OAuth flow, specifies which stage the option should be + displayed in the UI. + + OAUTH_STAGE_UNSPECIFIED is the default value for options unrelated to + the OAuth flow. + + BEFORE_AUTHORIZATION_CODE corresponds to options necessary to initiate + the OAuth process. + + BEFORE_ACCESS_TOKEN corresponds to options that are necessary to create + a foreign connection, + + but that should be displayed after the authorization code has already + been received. + OptionSpecOptionType: + type: string + enum: + - OPTION_BIGINT + - OPTION_BOOLEAN + - OPTION_ENUM + - OPTION_MULTILINE_STRING + - OPTION_NUMBER + - OPTION_SERVICE_CREDENTIAL + - OPTION_STRING + description: >- + Type of the option, we purposely follow JavaScript types so that the UI + can map the options to + + JS types. https://www.w3schools.com/js/js_datatypes.asp Enum is a + special case that it's just + + string with selections. + PathOperation: + type: string + enum: + - PATH_CREATE_TABLE + - PATH_READ + - PATH_READ_WRITE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PolicyType: + type: string + enum: + - POLICY_TYPE_COLUMN_MASK + - POLICY_TYPE_ROW_FILTER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PrincipalType: + type: string + enum: + - GROUP_PRINCIPAL + - SERVICE_PRINCIPAL + - USER_PRINCIPAL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + Privilege: + type: string + enum: + - ACCESS + - ALL_PRIVILEGES + - APPLY_TAG + - BROWSE + - CREATE + - CREATE_CATALOG + - CREATE_CLEAN_ROOM + - CREATE_CONNECTION + - CREATE_EXTERNAL_LOCATION + - CREATE_EXTERNAL_TABLE + - CREATE_EXTERNAL_VOLUME + - CREATE_FOREIGN_CATALOG + - CREATE_FOREIGN_SECURABLE + - CREATE_FUNCTION + - CREATE_MANAGED_STORAGE + - CREATE_MATERIALIZED_VIEW + - CREATE_MODEL + - CREATE_PROVIDER + - CREATE_RECIPIENT + - CREATE_SCHEMA + - CREATE_SERVICE_CREDENTIAL + - CREATE_SHARE + - CREATE_STORAGE_CREDENTIAL + - CREATE_TABLE + - CREATE_VIEW + - CREATE_VOLUME + - EXECUTE + - EXECUTE_CLEAN_ROOM_TASK + - EXTERNAL_USE_SCHEMA + - MANAGE + - MANAGE_ALLOWLIST + - MODIFY + - MODIFY_CLEAN_ROOM + - READ_FILES + - READ_PRIVATE_FILES + - READ_VOLUME + - REFRESH + - SELECT + - SET_SHARE_PERMISSION + - USAGE + - USE_CATALOG + - USE_CONNECTION + - USE_MARKETPLACE_ASSETS + - USE_PROVIDER + - USE_RECIPIENT + - USE_SCHEMA + - USE_SHARE + - WRITE_FILES + - WRITE_PRIVATE_FILES + - WRITE_VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ProvisioningInfoState: + type: string + enum: + - ACTIVE + - DEGRADED + - DELETING + - FAILED + - PROVISIONING + - UPDATING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SecurableKind: + type: string + enum: + - TABLE_DB_STORAGE + - TABLE_DELTA + - TABLE_DELTASHARING + - TABLE_DELTASHARING_MUTABLE + - TABLE_DELTASHARING_OPEN_DIR_BASED + - TABLE_DELTA_EXTERNAL + - TABLE_DELTA_ICEBERG_DELTASHARING + - TABLE_DELTA_ICEBERG_MANAGED + - TABLE_DELTA_UNIFORM_HUDI_EXTERNAL + - TABLE_DELTA_UNIFORM_ICEBERG_EXTERNAL + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_DELTASHARING + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_HIVE_METASTORE_EXTERNAL + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_HIVE_METASTORE_MANAGED + - TABLE_DELTA_UNIFORM_ICEBERG_FOREIGN_SNOWFLAKE + - TABLE_EXTERNAL + - TABLE_FEATURE_STORE + - TABLE_FEATURE_STORE_EXTERNAL + - TABLE_FOREIGN_BIGQUERY + - TABLE_FOREIGN_DATABRICKS + - TABLE_FOREIGN_DELTASHARING + - TABLE_FOREIGN_HIVE_METASTORE + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_SHALLOW_CLONE_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_SHALLOW_CLONE_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_DBFS_VIEW + - TABLE_FOREIGN_HIVE_METASTORE_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_SHALLOW_CLONE_EXTERNAL + - TABLE_FOREIGN_HIVE_METASTORE_SHALLOW_CLONE_MANAGED + - TABLE_FOREIGN_HIVE_METASTORE_VIEW + - TABLE_FOREIGN_MONGODB + - TABLE_FOREIGN_MYSQL + - TABLE_FOREIGN_NETSUITE + - TABLE_FOREIGN_ORACLE + - TABLE_FOREIGN_POSTGRESQL + - TABLE_FOREIGN_REDSHIFT + - TABLE_FOREIGN_SALESFORCE + - TABLE_FOREIGN_SALESFORCE_DATA_CLOUD + - TABLE_FOREIGN_SALESFORCE_DATA_CLOUD_FILE_SHARING + - TABLE_FOREIGN_SALESFORCE_DATA_CLOUD_FILE_SHARING_VIEW + - TABLE_FOREIGN_SNOWFLAKE + - TABLE_FOREIGN_SQLDW + - TABLE_FOREIGN_SQLSERVER + - TABLE_FOREIGN_TERADATA + - TABLE_FOREIGN_WORKDAY_RAAS + - TABLE_ICEBERG_UNIFORM_MANAGED + - TABLE_INTERNAL + - TABLE_MANAGED_POSTGRESQL + - TABLE_MATERIALIZED_VIEW + - TABLE_MATERIALIZED_VIEW_DELTASHARING + - TABLE_METRIC_VIEW + - TABLE_METRIC_VIEW_DELTASHARING + - TABLE_ONLINE_VECTOR_INDEX_DIRECT + - TABLE_ONLINE_VECTOR_INDEX_REPLICA + - TABLE_ONLINE_VIEW + - TABLE_STANDARD + - TABLE_STREAMING_LIVE_TABLE + - TABLE_STREAMING_LIVE_TABLE_DELTASHARING + - TABLE_SYSTEM + - TABLE_SYSTEM_DELTASHARING + - TABLE_VIEW + - TABLE_VIEW_DELTASHARING + description: 'Latest kind: CONNECTION_OKTA_SYSTEM_LOGS_SSWS_TOKEN = 295; Next id: 296' + SecurableType: + type: string + enum: + - CATALOG + - CLEAN_ROOM + - CONNECTION + - CREDENTIAL + - EXTERNAL_LOCATION + - EXTERNAL_METADATA + - FUNCTION + - METASTORE + - PIPELINE + - PROVIDER + - RECIPIENT + - SCHEMA + - SHARE + - STAGING_TABLE + - STORAGE_CREDENTIAL + - TABLE + - VOLUME + description: The type of Unity Catalog securable. + SpecialDestination: + type: string + enum: + - SPECIAL_DESTINATION_CATALOG_OWNER + - SPECIAL_DESTINATION_CONNECTION_OWNER + - SPECIAL_DESTINATION_CREDENTIAL_OWNER + - SPECIAL_DESTINATION_EXTERNAL_LOCATION_OWNER + - SPECIAL_DESTINATION_METASTORE_OWNER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SseEncryptionDetailsAlgorithm: + type: string + enum: + - AWS_SSE_KMS + - AWS_SSE_S3 + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SystemType: + type: string + enum: + - AMAZON_REDSHIFT + - AZURE_SYNAPSE + - CONFLUENT + - DATABRICKS + - GOOGLE_BIGQUERY + - KAFKA + - LOOKER + - MICROSOFT_FABRIC + - MICROSOFT_SQL_SERVER + - MONGODB + - MYSQL + - ORACLE + - OTHER + - POSTGRESQL + - POWER_BI + - SALESFORCE + - SAP + - SERVICENOW + - SNOWFLAKE + - STREAM_NATIVE + - TABLEAU + - TERADATA + - WORKDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TableOperation: + type: string + enum: + - READ + - READ_WRITE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TableType: + type: string + enum: + - EXTERNAL + - EXTERNAL_SHALLOW_CLONE + - FOREIGN + - MANAGED + - MANAGED_SHALLOW_CLONE + - MATERIALIZED_VIEW + - METRIC_VIEW + - STREAMING_TABLE + - VIEW + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ValidateCredentialResult: + type: string + enum: + - FAIL + - PASS + - SKIP + description: A enum represents the result of the file operation + ValidationResultOperation: + type: string + enum: + - DELETE + - LIST + - PATH_EXISTS + - READ + - WRITE + description: >- + A enum represents the file operation performed on the external location + with the storage + + credential + ValidationResultResult: + type: string + enum: + - FAIL + - PASS + - SKIP + description: A enum represents the result of the file operation + VolumeType: + type: string + enum: + - EXTERNAL + - MANAGED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + WorkspaceBindingBindingType: + type: string + enum: + - BINDING_TYPE_READ_ONLY + - BINDING_TYPE_READ_WRITE + description: >- + Using `BINDING_TYPE_` prefix here to avoid conflict with + `TableOperation` enum in + + `credentials_common.proto`. + x-stackQL-resources: + artifact_allowlists: + id: databricks_workspace.catalog.artifact_allowlists + name: artifact_allowlists + title: Artifact Allowlists + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1artifact-allowlists~1{artifact_type.value}/get + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1artifact-allowlists~1{artifact_type.value}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/artifact_allowlists/methods/get' + insert: [] + update: [] + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/artifact_allowlists/methods/replace + catalogs: + id: databricks_workspace.catalog.catalogs + name: catalogs + title: Catalogs + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.catalogs + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/catalogs/methods/get' + - $ref: '#/components/x-stackQL-resources/catalogs/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/catalogs/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/catalogs/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/catalogs/methods/delete' + replace: [] + connections: + id: databricks_workspace.catalog.connections + name: connections + title: Connections + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1connections/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1connections/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.connections + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1connections~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1connections~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1connections~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/connections/methods/get' + - $ref: '#/components/x-stackQL-resources/connections/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/connections/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/connections/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/connections/methods/delete' + replace: [] + entity_tag_assignments: + id: databricks_workspace.catalog.entity_tag_assignments + name: entity_tag_assignments + title: Entity Tag Assignments + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1entity-tag-assignments/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1entity-tag-assignments~1{entity_type}~1{entity_name}~1tags~1{tag_key}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1entity-tag-assignments~1{entity_type}~1{entity_name}~1tags~1{tag_key}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1entity-tag-assignments~1{entity_type}~1{entity_name}~1tags~1{tag_key}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1entity-tag-assignments~1{entity_type}~1{entity_name}~1tags/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.tag_assignments + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/entity_tag_assignments/methods/get + - $ref: >- + #/components/x-stackQL-resources/entity_tag_assignments/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/entity_tag_assignments/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/entity_tag_assignments/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/entity_tag_assignments/methods/delete + replace: [] + external_lineage: + id: databricks_workspace.catalog.external_lineage + name: external_lineage + title: External Lineage + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1lineage-tracking~1external-lineage/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1lineage-tracking~1external-lineage/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1lineage-tracking~1external-lineage/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.external_lineage_relationships + update: + operation: + $ref: '#/paths/~1api~12.0~1lineage-tracking~1external-lineage/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/external_lineage/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/external_lineage/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/external_lineage/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/external_lineage/methods/delete' + replace: [] + external_locations: + id: databricks_workspace.catalog.external_locations + name: external_locations + title: External Locations + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.external_locations + delete: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1external-locations~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1external-locations~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/external_locations/methods/get' + - $ref: '#/components/x-stackQL-resources/external_locations/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/external_locations/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/external_locations/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/external_locations/methods/delete' + replace: [] + external_metadata: + id: databricks_workspace.catalog.external_metadata + name: external_metadata + title: External Metadata + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1lineage-tracking~1external-metadata/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1lineage-tracking~1external-metadata/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.external_metadata + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1lineage-tracking~1external-metadata~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1lineage-tracking~1external-metadata~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1lineage-tracking~1external-metadata~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/external_metadata/methods/get' + - $ref: '#/components/x-stackQL-resources/external_metadata/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/external_metadata/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/external_metadata/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/external_metadata/methods/delete' + replace: [] + functions: + id: databricks_workspace.catalog.functions + name: functions + title: Functions + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1functions/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1functions/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.functions + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1functions~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1functions~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1functions~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/functions/methods/get' + - $ref: '#/components/x-stackQL-resources/functions/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/functions/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/functions/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/functions/methods/delete' + replace: [] + grants: + id: databricks_workspace.catalog.grants + name: grants + title: Grants + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1permissions~1{securable_type}~1{full_name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1permissions~1{securable_type}~1{full_name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/grants/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/grants/methods/update' + delete: [] + replace: [] + effective_grants: + id: databricks_workspace.catalog.effective_grants + name: effective_grants + title: Effective Grants + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1effective-permissions~1{securable_type}~1{full_name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/effective_grants/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + metastores: + id: databricks_workspace.catalog.metastores + name: metastores + title: Metastores + methods: + assign: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1workspaces~1{workspace_id}~1metastore/put + response: + mediaType: application/json + openAPIDocKey: '200' + unassign: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1workspaces~1{workspace_id}~1metastore/delete + response: + mediaType: application/json + openAPIDocKey: '200' + update_assignment: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1workspaces~1{workspace_id}~1metastore/patch + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.metastores + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/metastores/methods/get' + - $ref: '#/components/x-stackQL-resources/metastores/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/metastores/methods/create' + update: + - $ref: >- + #/components/x-stackQL-resources/metastores/methods/update_assignment + - $ref: '#/components/x-stackQL-resources/metastores/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/metastores/methods/unassign' + - $ref: '#/components/x-stackQL-resources/metastores/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/metastores/methods/assign' + current_metastore_assignment: + id: databricks_workspace.catalog.current_metastore_assignment + name: current_metastore_assignment + title: Current Metastore Assignment + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1current-metastore-assignment/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/current_metastore_assignment/methods/get + insert: [] + update: [] + delete: [] + replace: [] + metastore_summary: + id: databricks_workspace.catalog.metastore_summary + name: metastore_summary + title: Metastore Summary + methods: + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1metastore_summary/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/metastore_summary/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + model_versions: + id: databricks_workspace.catalog.model_versions + name: model_versions + title: Model Versions + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions~1{version}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions~1{version}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions~1{version}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + get_by_alias: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1aliases~1{alias}/get + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.model_versions + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/model_versions/methods/get' + - $ref: >- + #/components/x-stackQL-resources/model_versions/methods/get_by_alias + - $ref: '#/components/x-stackQL-resources/model_versions/methods/list' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/model_versions/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/model_versions/methods/delete' + replace: [] + registered_models: + id: databricks_workspace.catalog.registered_models + name: registered_models + title: Registered Models + methods: + delete_alias: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1aliases~1{alias}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + set_alias: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1aliases~1{alias}/put + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1models/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1models/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.registered_models + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/registered_models/methods/get' + - $ref: '#/components/x-stackQL-resources/registered_models/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/registered_models/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/registered_models/methods/update' + delete: + - $ref: >- + #/components/x-stackQL-resources/registered_models/methods/delete_alias + - $ref: '#/components/x-stackQL-resources/registered_models/methods/delete' + replace: + - $ref: >- + #/components/x-stackQL-resources/registered_models/methods/set_alias + online_tables: + id: databricks_workspace.catalog.online_tables + name: online_tables + title: Online Tables + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1online-tables/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1online-tables~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1online-tables~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/online_tables/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/online_tables/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/online_tables/methods/delete' + replace: [] + policies: + id: databricks_workspace.catalog.policies + name: policies + title: Policies + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1policies/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1policies~1{on_securable_type}~1{on_securable_fullname}~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1policies~1{on_securable_type}~1{on_securable_fullname}~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1policies~1{on_securable_type}~1{on_securable_fullname}~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1policies~1{on_securable_type}~1{on_securable_fullname}/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.policies + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/policies/methods/get' + - $ref: '#/components/x-stackQL-resources/policies/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/policies/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/policies/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/policies/methods/delete' + replace: [] + quality_monitors: + id: databricks_workspace.catalog.quality_monitors + name: quality_monitors + title: Quality Monitors + methods: + cancel_refresh: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor~1refreshes~1{refresh_id}~1cancel/post + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/post + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/put + response: + mediaType: application/json + openAPIDocKey: '200' + regenerate_dashboard: + operation: + $ref: >- + #/paths/~1api~12.1~1quality-monitoring~1tables~1{table_name}~1monitor~1dashboard/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/update' + quality_monitor_refreshes: + id: databricks_workspace.catalog.quality_monitor_refreshes + name: quality_monitor_refreshes + title: Quality Monitor Refreshes + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor~1refreshes~1{refresh_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor~1refreshes/get + response: + mediaType: application/json + openAPIDocKey: '200' + run: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor~1refreshes/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/quality_monitor_refreshes/methods/get + - $ref: >- + #/components/x-stackQL-resources/quality_monitor_refreshes/methods/list + insert: [] + update: [] + delete: [] + replace: [] + resource_quotas: + id: databricks_workspace.catalog.resource_quotas + name: resource_quotas + title: Resource Quotas + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1resource-quotas~1{parent_securable_type}~1{parent_full_name}~1{quota_name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1resource-quotas~1all-resource-quotas/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.quotas + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/resource_quotas/methods/get' + - $ref: '#/components/x-stackQL-resources/resource_quotas/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + rfa: + id: databricks_workspace.catalog.rfa + name: rfa + title: Rfa + methods: + batch_create: + operation: + $ref: '#/paths/~1api~13.0~1rfa~1requests/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get_destinations: + operation: + $ref: >- + #/paths/~1api~13.0~1rfa~1destinations~1{securable_type}~1{full_name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update_destinations: + operation: + $ref: '#/paths/~1api~13.0~1rfa~1destinations/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/rfa/methods/get_destinations' + insert: + - $ref: '#/components/x-stackQL-resources/rfa/methods/batch_create' + update: + - $ref: '#/components/x-stackQL-resources/rfa/methods/update_destinations' + delete: [] + replace: [] + schemas: + id: databricks_workspace.catalog.schemas + name: schemas + title: Schemas + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.schemas + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas~1{full_name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas~1{full_name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas~1{full_name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/schemas/methods/get' + - $ref: '#/components/x-stackQL-resources/schemas/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/schemas/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/schemas/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/schemas/methods/delete' + replace: [] + storage_credentials: + id: databricks_workspace.catalog.storage_credentials + name: storage_credentials + title: Storage Credentials + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.storage_credentials + delete: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1storage-credentials~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1storage-credentials~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + validate: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1validate-storage-credentials/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/get' + - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/list' + insert: + - $ref: >- + #/components/x-stackQL-resources/storage_credentials/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/storage_credentials/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/storage_credentials/methods/delete + replace: [] + system_schemas: + id: databricks_workspace.catalog.system_schemas + name: system_schemas + title: System Schemas + methods: + disable: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1metastores~1{metastore_id}~1systemschemas~1{schema_name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + enable: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1metastores~1{metastore_id}~1systemschemas~1{schema_name}/put + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1metastores~1{metastore_id}~1systemschemas/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.schemas + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/system_schemas/methods/list' + insert: [] + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/system_schemas/methods/disable' + replace: + - $ref: '#/components/x-stackQL-resources/system_schemas/methods/enable' + table_constraints: + id: databricks_workspace.catalog.table_constraints + name: table_constraints + title: Table Constraints + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1constraints/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1constraints~1{full_name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: '#/components/x-stackQL-resources/table_constraints/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/table_constraints/methods/delete' + replace: [] + tables: + id: databricks_workspace.catalog.tables + name: tables + title: Tables + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1tables/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1tables/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.tables + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{full_name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{full_name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{full_name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + exists: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{full_name}~1exists/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/tables/methods/get' + - $ref: '#/components/x-stackQL-resources/tables/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/tables/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/tables/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/tables/methods/delete' + replace: [] + table_summaries: + id: databricks_workspace.catalog.table_summaries + name: table_summaries + title: Table Summaries + methods: + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1table-summaries/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.tables + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/table_summaries/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + temporary_path_credentials: + id: databricks_workspace.catalog.temporary_path_credentials + name: temporary_path_credentials + title: Temporary Path Credentials + methods: + generate: + operation: + $ref: '#/paths/~1api~12.0~1unity-catalog~1temporary-path-credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: >- + #/components/x-stackQL-resources/temporary_path_credentials/methods/generate + update: [] + delete: [] + replace: [] + temporary_table_credentials: + id: databricks_workspace.catalog.temporary_table_credentials + name: temporary_table_credentials + title: Temporary Table Credentials + methods: + generate: + operation: + $ref: >- + #/paths/~1api~12.0~1unity-catalog~1temporary-table-credentials/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: >- + #/components/x-stackQL-resources/temporary_table_credentials/methods/generate + update: [] + delete: [] + replace: [] + volumes: + id: databricks_workspace.catalog.volumes + name: volumes + title: Volumes + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.volumes + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + read: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/volumes/methods/read' + - $ref: '#/components/x-stackQL-resources/volumes/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/volumes/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/volumes/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/volumes/methods/delete' + replace: [] + workspace_bindings: + id: databricks_workspace.catalog.workspace_bindings + name: workspace_bindings + title: Workspace Bindings + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1workspace-bindings~1catalogs~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1workspace-bindings~1catalogs~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + get_bindings: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1bindings~1{securable_type}~1{securable_name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.bindings + update_bindings: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1bindings~1{securable_type}~1{securable_name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/workspace_bindings/methods/get_bindings + - $ref: '#/components/x-stackQL-resources/workspace_bindings/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/workspace_bindings/methods/update_bindings + - $ref: '#/components/x-stackQL-resources/workspace_bindings/methods/update' + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/cleanrooms.yaml b/providers/src/databricks_workspace/v00.00.00000/services/cleanrooms.yaml index 5bcb6f2e..9d9058e9 100644 --- a/providers/src/databricks_workspace/v00.00.00000/services/cleanrooms.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/services/cleanrooms.yaml @@ -1,1849 +1,2234 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Cleanrooms API - description: Cleanrooms -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/clean-rooms/{clean_room_name}/assets: - post: - operationId: assets-create - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanroomassets/create - x-stackQL-resource: assets - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: clean_room_name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - asset_type: - type: string - added_at: - type: integer - status: - type: string - owner_collaborator_alias: - type: string - table_local_details: - type: object - properties: - local_name: - type: string - partitions: - type: array - items: - type: object - properties: - values: - type: array - items: - type: object - properties: - name: - type: string - value: - type: string - recipient_property_key: - type: string - op: - type: string - volume_local_details: - type: object - properties: - local_name: - type: string - view_local_details: - type: object - properties: - local_name: - type: string - foreign_table_local_details: - type: object - properties: - local_name: - type: string - table: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - notebook: - type: object - properties: - notebook_content: - type: string - etag: - type: string - view: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - foreign_table: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - asset_type: - type: string - table_local_details: - type: object - volume_local_details: - type: object - view_local_details: - type: object - foreign_table_local_details: - type: object - table: - type: object - notebook: - type: object - view: - type: object - foreign_table: - type: object - example: - name: creator.sales.products - asset_type: TABLE - table_local_details: - local_name: demo.sales.products - partitions: - - values: - - name: string - value: string - recipient_property_key: string - op: EQUAL - volume_local_details: - local_name: demo.sales.products - view_local_details: - local_name: demo.sales.products - foreign_table_local_details: - local_name: demo.sales.products - table: - columns: - - name: string - type_text: string - type_name: BOOLEAN - position: 0 - type_precision: 0 - type_scale: 0 - type_interval_type: string - type_json: string - comment: string - nullable: true - partition_index: 0 - mask: - function_name: string - using_column_names: - - string - notebook: - notebook_content: string - etag: string - view: - columns: - - name: string - type_text: string - type_name: BOOLEAN - position: 0 - type_precision: 0 - type_scale: 0 - type_interval_type: string - type_json: string - comment: string - nullable: true - partition_index: 0 - mask: - function_name: string - using_column_names: - - string - foreign_table: - columns: - - name: string - type_text: string - type_name: BOOLEAN - position: 0 - type_precision: 0 - type_scale: 0 - type_interval_type: string - type_json: string - comment: string - nullable: true - partition_index: 0 - mask: - function_name: string - using_column_names: - - string - description: Metadata of the clean room asset - description: "Create a clean room asset \u2014share an asset like a notebook\ - \ or table into the clean room. For each UC asset that is added through this\ - \ method, the clean room owner must also have enough privilege on the asset\ - \ to consume it. The privilege must be maintained indefinitely for the clean\ - \ room to be able to access the asset. Typically, you should use a group as\ - \ the clean room owner." - get: - operationId: assets-list - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanroomassets/list - x-stackQL-resource: assets - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: clean_room_name - required: true - description: string - in: path - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - assets: - type: array - items: - type: object - properties: - name: - type: string - asset_type: - type: string - added_at: - type: integer - status: - type: string - owner_collaborator_alias: - type: string - table_local_details: - type: object - properties: - local_name: - type: string - partitions: - type: array - items: - type: object - properties: - values: - type: array - items: - type: object - properties: - name: - type: string - value: - type: string - recipient_property_key: - type: string - op: - type: string - volume_local_details: - type: object - properties: - local_name: - type: string - view_local_details: - type: object - properties: - local_name: - type: string - foreign_table_local_details: - type: object - properties: - local_name: - type: string - table: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - notebook: - type: object - properties: - notebook_content: - type: string - etag: - type: string - view: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - foreign_table: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.assets - /api/2.0/clean-rooms/{clean_room_name}/assets/{asset_type}/{asset_full_name}: - delete: - operationId: assets-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanroomassets/delete - x-stackQL-resource: assets - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: clean_room_name - required: true - description: string - in: path - - name: asset_type - required: true - description: string - in: path - - name: asset_full_name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Delete a clean room asset - unshare/remove the asset from the clean - room - /api/2.0/clean-rooms/{clean_room_name}/assets/{asset_type}/{name}: - patch: - operationId: assets-update - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanroomassets/update - x-stackQL-resource: assets - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 3 - parameters: - - name: clean_room_name - required: true - description: string - in: path - - name: asset_type - required: true - description: string - in: path - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - asset_type: - type: string - added_at: - type: integer - status: - type: string - owner_collaborator_alias: - type: string - table_local_details: - type: object - properties: - local_name: - type: string - partitions: - type: array - items: - type: object - properties: - values: - type: array - items: - type: object - properties: - name: - type: string - value: - type: string - recipient_property_key: - type: string - op: - type: string - volume_local_details: - type: object - properties: - local_name: - type: string - view_local_details: - type: object - properties: - local_name: - type: string - foreign_table_local_details: - type: object - properties: - local_name: - type: string - table: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - notebook: - type: object - properties: - notebook_content: - type: string - etag: - type: string - view: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - foreign_table: - type: object - properties: - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - asset_type: - type: string - table_local_details: - type: object - volume_local_details: - type: object - view_local_details: - type: object - foreign_table_local_details: - type: object - table: - type: object - notebook: - type: object - view: - type: object - foreign_table: - type: object - example: - name: creator.sales.products - asset_type: TABLE - table_local_details: - local_name: demo.sales.products - partitions: - - values: - - name: string - value: string - recipient_property_key: string - op: EQUAL - volume_local_details: - local_name: demo.sales.products - view_local_details: - local_name: demo.sales.products - foreign_table_local_details: - local_name: demo.sales.products - table: - columns: - - name: string - type_text: string - type_name: BOOLEAN - position: 0 - type_precision: 0 - type_scale: 0 - type_interval_type: string - type_json: string - comment: string - nullable: true - partition_index: 0 - mask: - function_name: string - using_column_names: - - string - notebook: - notebook_content: string - etag: string - view: - columns: - - name: string - type_text: string - type_name: BOOLEAN - position: 0 - type_precision: 0 - type_scale: 0 - type_interval_type: string - type_json: string - comment: string - nullable: true - partition_index: 0 - mask: - function_name: string - using_column_names: - - string - foreign_table: - columns: - - name: string - type_text: string - type_name: BOOLEAN - position: 0 - type_precision: 0 - type_scale: 0 - type_interval_type: string - type_json: string - comment: string - nullable: true - partition_index: 0 - mask: - function_name: string - using_column_names: - - string - description: Metadata of the clean room asset - description: Update a clean room asset. For example, updating the content of - a notebook; changing the shared partitions of a table; etc. - /api/2.0/clean-rooms: - post: - operationId: clean-rooms-create - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanrooms/create - x-stackQL-resource: clean_rooms - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - remote_detailed_info: - type: object - properties: - central_clean_room_id: - type: string - cloud_vendor: - type: string - region: - type: string - collaborators: - type: array - items: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - creator: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - egress_network_policy: - type: object - properties: - internet_access: - type: object - properties: - restriction_mode: - type: string - allowed_internet_destinations: - type: array - items: - type: object - properties: - destination: - type: string - type: - type: string - protocol: - type: string - allowed_storage_destinations: - type: array - items: - type: object - properties: - bucket_name: - type: string - region: - type: string - type: - type: string - azure_storage_account: - type: string - allowed_paths: - type: array - items: - type: string - azure_storage_service: - type: string - azure_dns_zone: - type: string - azure_container: - type: string - log_only_mode: - type: object - properties: - log_only_mode_type: - type: string - workloads: - type: array - items: - type: string - compliance_security_profile: - type: object - properties: - is_enabled: - type: boolean - compliance_standards: - type: array - items: - type: string - owner: - type: string - comment: - type: string - created_at: - type: integer - updated_at: - type: integer - status: - type: string - local_collaborator_alias: - type: string - output_catalog: - type: object - properties: - status: - type: string - catalog_name: - type: string - access_restricted: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - remote_detailed_info: - type: object - example: - name: test-clean-room1 - remote_detailed_info: - central_clean_room_id: b5c8d856-ff41-4c5f-8ccd-2b13b44fec27 - cloud_vendor: aws - region: us-west-2 - collaborators: - - global_metastore_id: aws:us-west-2:aff56c64-a34e-4c1f-a24c-c2dd2889517a - organization_name: acme corporation - invite_recipient_workspace_id: 6822898386300992 - invite_recipient_email: john.doe@databricks.com - collaborator_alias: creator - display_name: acme corporation - creator: - global_metastore_id: aws:us-west-2:aff56c64-a34e-4c1f-a24c-c2dd2889517a - organization_name: acme corporation - invite_recipient_workspace_id: 6822898386300992 - invite_recipient_email: john.doe@databricks.com - collaborator_alias: creator - display_name: acme corporation - egress_network_policy: - internet_access: - restriction_mode: FULL_ACCESS - allowed_internet_destinations: - - destination: string - type: FQDN - protocol: TCP - allowed_storage_destinations: - - bucket_name: string - region: string - type: AWS_S3 - azure_storage_account: string - allowed_paths: - - string - azure_storage_service: string - azure_dns_zone: string - azure_container: string - log_only_mode: - log_only_mode_type: ALL_SERVICES - workloads: - - DBSQL - compliance_security_profile: - is_enabled: true - compliance_standards: - - NONE - owner: alice@example.com - comment: This is a clean room for demo - description: Create a new clean room with the specified collaborators. This - method is asynchronous; the returned name field inside the clean_room field - can be used to poll the clean room status, using the - get: - operationId: clean-rooms-list - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanrooms/list - x-stackQL-resource: clean_rooms - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_size - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - clean_rooms: - type: array - items: - type: object - properties: - name: - type: string - remote_detailed_info: - type: object - properties: - central_clean_room_id: - type: string - cloud_vendor: - type: string - region: - type: string - collaborators: - type: array - items: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - creator: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - egress_network_policy: - type: object - properties: - internet_access: - type: object - properties: - restriction_mode: - type: string - allowed_internet_destinations: - type: array - items: - type: object - properties: - destination: - type: string - type: - type: string - protocol: - type: string - allowed_storage_destinations: - type: array - items: - type: object - properties: - bucket_name: - type: string - region: - type: string - type: - type: string - azure_storage_account: - type: string - allowed_paths: - type: array - items: - type: string - azure_storage_service: - type: string - azure_dns_zone: - type: string - azure_container: - type: string - log_only_mode: - type: object - properties: - log_only_mode_type: - type: string - workloads: - type: array - items: - type: string - compliance_security_profile: - type: object - properties: - is_enabled: - type: boolean - compliance_standards: - type: array - items: - type: string - owner: - type: string - comment: - type: string - created_at: - type: integer - updated_at: - type: integer - status: - type: string - local_collaborator_alias: - type: string - output_catalog: - type: object - properties: - status: - type: string - catalog_name: - type: string - access_restricted: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.clean_rooms - description: Get a list of all clean rooms of the metastore. Only clean rooms - the caller has access to are returned. - /api/2.0/clean-rooms/{name}: - delete: - operationId: clean-rooms-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanrooms/delete - x-stackQL-resource: clean_rooms - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Delete a clean room. After deletion, the clean room will be removed - from the metastore. If the other collaborators have not deleted the clean - room, they will still have the clean room in their metastore, but it will - be in a DELETED state and no operations other than deletion can be performed - on it. - get: - operationId: clean-rooms-get - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanrooms/get - x-stackQL-resource: clean_rooms - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - remote_detailed_info: - type: object - properties: - central_clean_room_id: - type: string - cloud_vendor: - type: string - region: - type: string - collaborators: - type: array - items: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - creator: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - egress_network_policy: - type: object - properties: - internet_access: - type: object - properties: - restriction_mode: - type: string - allowed_internet_destinations: - type: array - items: - type: object - properties: - destination: - type: string - type: - type: string - protocol: - type: string - allowed_storage_destinations: - type: array - items: - type: object - properties: - bucket_name: - type: string - region: - type: string - type: - type: string - azure_storage_account: - type: string - allowed_paths: - type: array - items: - type: string - azure_storage_service: - type: string - azure_dns_zone: - type: string - azure_container: - type: string - log_only_mode: - type: object - properties: - log_only_mode_type: - type: string - workloads: - type: array - items: - type: string - compliance_security_profile: - type: object - properties: - is_enabled: - type: boolean - compliance_standards: - type: array - items: - type: string - owner: - type: string - comment: - type: string - created_at: - type: integer - updated_at: - type: integer - status: - type: string - local_collaborator_alias: - type: string - output_catalog: - type: object - properties: - status: - type: string - catalog_name: - type: string - access_restricted: - type: string - description: Get the details of a clean room given its name. - patch: - operationId: clean-rooms-update - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanrooms/update - x-stackQL-resource: clean_rooms - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - remote_detailed_info: - type: object - properties: - central_clean_room_id: - type: string - cloud_vendor: - type: string - region: - type: string - collaborators: - type: array - items: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - creator: - type: object - properties: - global_metastore_id: - type: string - organization_name: - type: string - invite_recipient_workspace_id: - type: integer - invite_recipient_email: - type: string - collaborator_alias: - type: string - display_name: - type: string - egress_network_policy: - type: object - properties: - internet_access: - type: object - properties: - restriction_mode: - type: string - allowed_internet_destinations: - type: array - items: - type: object - properties: - destination: - type: string - type: - type: string - protocol: - type: string - allowed_storage_destinations: - type: array - items: - type: object - properties: - bucket_name: - type: string - region: - type: string - type: - type: string - azure_storage_account: - type: string - allowed_paths: - type: array - items: - type: string - azure_storage_service: - type: string - azure_dns_zone: - type: string - azure_container: - type: string - log_only_mode: - type: object - properties: - log_only_mode_type: - type: string - workloads: - type: array - items: - type: string - compliance_security_profile: - type: object - properties: - is_enabled: - type: boolean - compliance_standards: - type: array - items: - type: string - owner: - type: string - comment: - type: string - created_at: - type: integer - updated_at: - type: integer - status: - type: string - local_collaborator_alias: - type: string - output_catalog: - type: object - properties: - status: - type: string - catalog_name: - type: string - access_restricted: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - clean_room: - type: object - example: - clean_room: - name: test-clean-room1 - remote_detailed_info: - central_clean_room_id: b5c8d856-ff41-4c5f-8ccd-2b13b44fec27 - cloud_vendor: aws - region: us-west-2 - collaborators: - - global_metastore_id: aws:us-west-2:aff56c64-a34e-4c1f-a24c-c2dd2889517a - organization_name: acme corporation - invite_recipient_workspace_id: 6822898386300992 - invite_recipient_email: john.doe@databricks.com - collaborator_alias: creator - display_name: acme corporation - creator: - global_metastore_id: aws:us-west-2:aff56c64-a34e-4c1f-a24c-c2dd2889517a - organization_name: acme corporation - invite_recipient_workspace_id: 6822898386300992 - invite_recipient_email: john.doe@databricks.com - collaborator_alias: creator - display_name: acme corporation - egress_network_policy: - internet_access: - restriction_mode: FULL_ACCESS - allowed_internet_destinations: - - destination: string - type: FQDN - protocol: TCP - allowed_storage_destinations: - - bucket_name: string - region: string - type: AWS_S3 - azure_storage_account: string - allowed_paths: - - string - azure_storage_service: string - azure_dns_zone: string - azure_container: string - log_only_mode: - log_only_mode_type: ALL_SERVICES - workloads: - - DBSQL - compliance_security_profile: - is_enabled: true - compliance_standards: - - NONE - owner: alice@example.com - comment: This is a clean room for demo - created_at: 1716424663321 - updated_at: 1716424663321 - status: ACTIVE - local_collaborator_alias: creator - output_catalog: - status: NOT_ELIGIBLE - catalog_name: string - access_restricted: NO_RESTRICTION - description: Update a clean room. The caller must be the owner of the clean - room, have - /api/2.0/clean-rooms/{clean_room_name}/output-catalogs: - post: - operationId: output-catalog-createoutputcatalog - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanrooms/createoutputcatalog - x-stackQL-resource: output_catalog - x-stackQL-method: createoutputcatalog - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: clean_room_name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - output_catalog: - type: object - properties: - status: - type: string - catalog_name: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - catalog_name: - type: string - example: - catalog_name: string - description: Create the output catalog of the clean room. - /api/2.0/clean-rooms/{clean_room_name}/runs: - get: - operationId: task-runs-list - externalDocs: - url: https://docs.databricks.com/api/workspace/cleanroomtaskruns/list - x-stackQL-resource: task_runs - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: clean_room_name - required: true - description: string - in: path - - name: notebook_name - in: query - - name: page_size - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - runs: - type: array - items: - type: object - properties: - notebook_name: - type: string - start_time: - type: integer - run_duration: - type: integer - notebook_job_run_state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - collaborator_job_run_info: - type: object - properties: - collaborator_job_id: - type: integer - collaborator_job_run_id: - type: integer - collaborator_task_run_id: - type: integer - collaborator_workspace_id: - type: integer - collaborator_alias: - type: string - output_schema_name: - type: string - output_schema_expiration_time: - type: integer - next_page_token: - type: string - x-stackQL-objectKey: $.runs - description: List all the historical notebook task runs in a clean room. -components: - x-stackQL-resources: - assets: - id: databricks_workspace.cleanrooms.assets - name: assets - title: Assets - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.assets - delete: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type}~1{asset_full_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type}~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/assets/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/assets/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/assets/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/assets/methods/delete' - exec: [] - clean_rooms: - id: databricks_workspace.cleanrooms.clean_rooms - name: clean_rooms - title: Clean_rooms - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.clean_rooms - delete: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/get' - - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/delete' - exec: [] - output_catalog: - id: databricks_workspace.cleanrooms.output_catalog - name: output_catalog - title: Output_catalog - methods: - createoutputcatalog: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1output-catalogs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: - - $ref: '#/components/x-stackQL-resources/output_catalog/methods/createoutputcatalog' - update: [] - replace: [] - delete: [] - exec: [] - task_runs: - id: databricks_workspace.cleanrooms.task_runs - name: task_runs - title: Task_runs - methods: - list: - operation: - $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1runs/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.runs - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/task_runs/methods/list' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] +openapi: 3.0.0 +info: + title: Databricks Cleanrooms API (workspace) + description: >- + OpenAPI specification for the Databricks cleanrooms service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.cleanrooms +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/clean-rooms/{clean_room_name}/assets/{asset_type.value}/{name}/revisions/{etag}: + get: + operationId: clean_room_asset_revisions_get + summary: Get a specific revision of an asset + tags: + - cleanrooms + - clean_room_asset_revisions + description: |- + Get a specific revision of an asset + + :param clean_room_name: str + Name of the clean room. + :param asset_type: :class:`CleanRoomAssetAssetType` + Asset type. Only NOTEBOOK_FILE is supported. + :param name: str + Name of the asset. + :param etag: str + Revision etag to fetch. If not provided, the latest revision will be returned. + + :returns: :class:`CleanRoomAsset` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + - name: asset_type.value + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: Name of the asset. + - name: etag + in: path + required: true + schema: + type: string + description: >- + Revision etag to fetch. If not provided, the latest revision will be + returned. + - name: asset_type + in: query + required: true + schema: + type: string + description: Asset type. Only NOTEBOOK_FILE is supported. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoomAsset' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetRevisionsAPI + /api/2.0/clean-rooms/{clean_room_name}/assets/{asset_type.value}/{name}/revisions: + get: + operationId: clean_room_asset_revisions_list + summary: List revisions for an asset + tags: + - cleanrooms + - clean_room_asset_revisions + description: |- + List revisions for an asset + + :param clean_room_name: str + Name of the clean room. + :param asset_type: :class:`CleanRoomAssetAssetType` + Asset type. Only NOTEBOOK_FILE is supported. + :param name: str + Name of the asset. + :param page_size: int (optional) + Maximum number of asset revisions to return. Defaults to 10. + :param page_token: str (optional) + Opaque pagination token to go to next page based on the previous query. + + :returns: Iterator over :class:`CleanRoomAsset` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + - name: asset_type.value + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: Name of the asset. + - name: asset_type + in: query + required: true + schema: + type: string + description: Asset type. Only NOTEBOOK_FILE is supported. + - name: page_size + in: query + required: false + schema: + type: string + description: Maximum number of asset revisions to return. Defaults to 10. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Opaque pagination token to go to next page based on the previous + query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCleanRoomAssetRevisionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetRevisionsAPI + /api/2.0/clean-rooms/{clean_room_name}/assets: + post: + operationId: clean_room_assets_create + summary: >- + Create a clean room asset —share an asset like a notebook or table into + the clean room. For each UC + tags: + - cleanrooms + - clean_room_assets + description: >- + Create a clean room asset —share an asset like a notebook or table into + the clean room. For each UC + + asset that is added through this method, the clean room owner must also + have enough privilege on the + + asset to consume it. The privilege must be maintained indefinitely for + the clean room to be able to + + access the asset. Typically, you should use a group as the clean room + owner. + + + :param clean_room_name: str + The name of the clean room this asset belongs to. This field is required for create operations and + populated by the server for responses. + :param asset: :class:`CleanRoomAsset` + + + :returns: :class:`CleanRoomAsset` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: >- + The name of the clean room this asset belongs to. This field is + required for create operations and populated by the server for + responses. + requestBody: + content: + application/json: + schema: + type: object + properties: + asset: + type: string + description: ':returns: :class:`CleanRoomAsset`' + required: + - asset + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoomAsset' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetsAPI + get: + operationId: clean_room_assets_list + summary: List assets. + tags: + - cleanrooms + - clean_room_assets + description: |- + List assets. + + :param clean_room_name: str + Name of the clean room. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`CleanRoomAsset` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCleanRoomAssetRevisionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetsAPI + /api/2.0/clean-rooms/{clean_room_name}/assets/{asset_type.value}/{name}/reviews: + post: + operationId: clean_room_assets_create_clean_room_asset_review + summary: Submit an asset review + tags: + - cleanrooms + - clean_room_assets + description: |- + Submit an asset review + + :param clean_room_name: str + Name of the clean room + :param asset_type: :class:`CleanRoomAssetAssetType` + Asset type. Can either be NOTEBOOK_FILE or JAR_ANALYSIS. + :param name: str + Name of the asset + :param notebook_review: :class:`NotebookVersionReview` (optional) + + :returns: :class:`CreateCleanRoomAssetReviewResponse` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room + - name: asset_type.value + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: Name of the asset + requestBody: + content: + application/json: + schema: + type: object + properties: + asset_type: + type: string + description: Asset type. Can either be NOTEBOOK_FILE or JAR_ANALYSIS. + notebook_review: + type: string + description: ':returns: :class:`CreateCleanRoomAssetReviewResponse`' + required: + - asset_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateCleanRoomAssetReviewResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetsAPI + /api/2.0/clean-rooms/{clean_room_name}/assets/{asset_type.value}/{name}: + delete: + operationId: clean_room_assets_delete + summary: Delete a clean room asset - unshare/remove the asset from the clean room + tags: + - cleanrooms + - clean_room_assets + description: |- + Delete a clean room asset - unshare/remove the asset from the clean room + + :param clean_room_name: str + Name of the clean room. + :param asset_type: :class:`CleanRoomAssetAssetType` + The type of the asset. + :param name: str + The fully qualified name of the asset, it is same as the name field in CleanRoomAsset. + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + - name: asset_type.value + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: >- + The fully qualified name of the asset, it is same as the name field + in CleanRoomAsset. + - name: asset_type + in: query + required: true + schema: + type: string + description: The type of the asset. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetsAPI + get: + operationId: clean_room_assets_get + summary: Get the details of a clean room asset by its type and full name. + tags: + - cleanrooms + - clean_room_assets + description: |- + Get the details of a clean room asset by its type and full name. + + :param clean_room_name: str + Name of the clean room. + :param asset_type: :class:`CleanRoomAssetAssetType` + The type of the asset. + :param name: str + The fully qualified name of the asset, it is same as the name field in CleanRoomAsset. + + :returns: :class:`CleanRoomAsset` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + - name: asset_type.value + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: >- + The fully qualified name of the asset, it is same as the name field + in CleanRoomAsset. + - name: asset_type + in: query + required: true + schema: + type: string + description: The type of the asset. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoomAsset' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetsAPI + patch: + operationId: clean_room_assets_update + summary: >- + Update a clean room asset. For example, updating the content of a + notebook; changing the shared + tags: + - cleanrooms + - clean_room_assets + description: >- + Update a clean room asset. For example, updating the content of a + notebook; changing the shared + + partitions of a table; etc. + + + :param clean_room_name: str + Name of the clean room. + :param asset_type: :class:`CleanRoomAssetAssetType` + The type of the asset. + :param name: str + A fully qualified name that uniquely identifies the asset within the clean room. This is also the + name displayed in the clean room UI. + + For UC securable assets (tables, volumes, etc.), the format is + *shared_catalog*.*shared_schema*.*asset_name* + + For notebooks, the name is the notebook file name. For jar analyses, the name is the jar analysis + name. + :param asset: :class:`CleanRoomAsset` + The asset to update. The asset's `name` and `asset_type` fields are used to identify the asset to + update. + + :returns: :class:`CleanRoomAsset` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + - name: asset_type.value + in: path + required: true + schema: + type: string + - name: name + in: path + required: true + schema: + type: string + description: >- + A fully qualified name that uniquely identifies the asset within the + clean room. This is also the name displayed in the clean room UI. + For UC securable assets (tables, volumes, etc.), the format is + *shared_catalog*.*shared_schema*.*asset_name* For notebooks, the + name is the notebook file name. For jar analyses, the name is the + jar analysis name. + requestBody: + content: + application/json: + schema: + type: object + properties: + asset_type: + type: string + description: The type of the asset. + asset: + type: string + description: >- + The asset to update. The asset's `name` and `asset_type` + fields are used to identify the asset to update. + required: + - asset_type + - asset + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoomAsset' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAssetsAPI + /api/2.0/clean-rooms/{clean_room_name}/auto-approval-rules: + post: + operationId: clean_room_auto_approval_rules_create + summary: Create an auto-approval rule + tags: + - cleanrooms + - clean_room_auto_approval_rules + description: |- + Create an auto-approval rule + + :param clean_room_name: str + The name of the clean room this auto-approval rule belongs to. + :param auto_approval_rule: :class:`CleanRoomAutoApprovalRule` + + :returns: :class:`CleanRoomAutoApprovalRule` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: The name of the clean room this auto-approval rule belongs to. + requestBody: + content: + application/json: + schema: + type: object + properties: + auto_approval_rule: + type: string + description: ':returns: :class:`CleanRoomAutoApprovalRule`' + required: + - auto_approval_rule + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoomAutoApprovalRule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAutoApprovalRulesAPI + get: + operationId: clean_room_auto_approval_rules_list + summary: List all auto-approval rules for the caller + tags: + - cleanrooms + - clean_room_auto_approval_rules + description: |- + List all auto-approval rules for the caller + + :param clean_room_name: str + :param page_size: int (optional) + Maximum number of auto-approval rules to return. Defaults to 100. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`CleanRoomAutoApprovalRule` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: >- + :param page_size: int (optional) Maximum number of auto-approval + rules to return. Defaults to 100. + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCleanRoomAutoApprovalRulesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAutoApprovalRulesAPI + /api/2.0/clean-rooms/{clean_room_name}/auto-approval-rules/{rule_id}: + delete: + operationId: clean_room_auto_approval_rules_delete + summary: Delete a auto-approval rule by rule ID + tags: + - cleanrooms + - clean_room_auto_approval_rules + description: |- + Delete a auto-approval rule by rule ID + + :param clean_room_name: str + :param rule_id: str + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: ':param rule_id: str' + - name: rule_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAutoApprovalRulesAPI + get: + operationId: clean_room_auto_approval_rules_get + summary: Get a auto-approval rule by rule ID + tags: + - cleanrooms + - clean_room_auto_approval_rules + description: |- + Get a auto-approval rule by rule ID + + :param clean_room_name: str + :param rule_id: str + + :returns: :class:`CleanRoomAutoApprovalRule` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: ':param rule_id: str' + - name: rule_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoomAutoApprovalRule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAutoApprovalRulesAPI + patch: + operationId: clean_room_auto_approval_rules_update + summary: Update a auto-approval rule by rule ID + tags: + - cleanrooms + - clean_room_auto_approval_rules + description: |- + Update a auto-approval rule by rule ID + + :param clean_room_name: str + The name of the clean room this auto-approval rule belongs to. + :param rule_id: str + A generated UUID identifying the rule. + :param auto_approval_rule: :class:`CleanRoomAutoApprovalRule` + The auto-approval rule to update. The rule_id field is used to identify the rule to update. + + :returns: :class:`CleanRoomAutoApprovalRule` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: The name of the clean room this auto-approval rule belongs to. + - name: rule_id + in: path + required: true + schema: + type: string + description: A generated UUID identifying the rule. + requestBody: + content: + application/json: + schema: + type: object + properties: + auto_approval_rule: + type: string + description: >- + The auto-approval rule to update. The rule_id field is used + to identify the rule to update. + required: + - auto_approval_rule + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoomAutoApprovalRule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomAutoApprovalRulesAPI + /api/2.0/clean-rooms/{clean_room_name}/runs: + get: + operationId: clean_room_task_runs_list + summary: List all the historical notebook task runs in a clean room. + tags: + - cleanrooms + - clean_room_task_runs + description: |- + List all the historical notebook task runs in a clean room. + + :param clean_room_name: str + Name of the clean room. + :param notebook_name: str (optional) + Notebook name + :param page_size: int (optional) + The maximum number of task runs to return. Currently ignored - all runs will be returned. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`CleanRoomNotebookTaskRun` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + - name: notebook_name + in: query + required: false + schema: + type: string + description: Notebook name + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of task runs to return. Currently ignored - all + runs will be returned. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCleanRoomNotebookTaskRunsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomTaskRunsAPI + /api/2.0/clean-rooms: + post: + operationId: clean_rooms_create + summary: >- + Create a new clean room with the specified collaborators. This method is + asynchronous; the returned + tags: + - cleanrooms + - clean_rooms + description: >- + Create a new clean room with the specified collaborators. This method is + asynchronous; the returned + + name field inside the clean_room field can be used to poll the clean + room status, using the + + :method:cleanrooms/get method. When this method returns, the clean room + will be in a PROVISIONING + + state, with only name, owner, comment, created_at and status populated. + The clean room will be usable + + once it enters an ACTIVE state. + + + The caller must be a metastore admin or have the **CREATE_CLEAN_ROOM** + privilege on the metastore. + + + :param clean_room: :class:`CleanRoom` + + + :returns: + Long-running operation waiter for :class:`CleanRoom`. + See :method:wait_get_clean_room_active for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + clean_room: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`CleanRoom`. See :method:wait_get_clean_room_active + for more details. + required: + - clean_room + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoom' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomsAPI + get: + operationId: clean_rooms_list + summary: >- + Get a list of all clean rooms of the metastore. Only clean rooms the + caller has access to are + tags: + - cleanrooms + - clean_rooms + description: >- + Get a list of all clean rooms of the metastore. Only clean rooms the + caller has access to are + + returned. + + + :param page_size: int (optional) + Maximum number of clean rooms to return (i.e., the page length). Defaults to 100. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`CleanRoom` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: >- + Maximum number of clean rooms to return (i.e., the page length). + Defaults to 100. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCleanRoomsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomsAPI + /api/2.0/clean-rooms/{clean_room_name}/output-catalogs: + post: + operationId: clean_rooms_create_output_catalog + summary: Create the output catalog of the clean room. + tags: + - cleanrooms + - clean_rooms + description: |- + Create the output catalog of the clean room. + + :param clean_room_name: str + Name of the clean room. + :param output_catalog: :class:`CleanRoomOutputCatalog` + + :returns: :class:`CreateCleanRoomOutputCatalogResponse` + parameters: + - name: clean_room_name + in: path + required: true + schema: + type: string + description: Name of the clean room. + requestBody: + content: + application/json: + schema: + type: object + properties: + output_catalog: + type: string + description: ':returns: :class:`CreateCleanRoomOutputCatalogResponse`' + required: + - output_catalog + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateCleanRoomOutputCatalogResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomsAPI + /api/2.0/clean-rooms/{name}: + delete: + operationId: clean_rooms_delete + summary: >- + Delete a clean room. After deletion, the clean room will be removed from + the metastore. If the other + tags: + - cleanrooms + - clean_rooms + description: >- + Delete a clean room. After deletion, the clean room will be removed from + the metastore. If the other + + collaborators have not deleted the clean room, they will still have the + clean room in their metastore, + + but it will be in a DELETED state and no operations other than deletion + can be performed on it. + + + :param name: str + Name of the clean room. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the clean room. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomsAPI + get: + operationId: clean_rooms_get + summary: Get the details of a clean room given its name. + tags: + - cleanrooms + - clean_rooms + description: |- + Get the details of a clean room given its name. + + :param name: str + + :returns: :class:`CleanRoom` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: ':returns: :class:`CleanRoom`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoom' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomsAPI + patch: + operationId: clean_rooms_update + summary: >- + Update a clean room. The caller must be the owner of the clean room, + have **MODIFY_CLEAN_ROOM** + tags: + - cleanrooms + - clean_rooms + description: >- + Update a clean room. The caller must be the owner of the clean room, + have **MODIFY_CLEAN_ROOM** + + privilege, or be metastore admin. + + + When the caller is a metastore admin, only the __owner__ field can be + updated. + + + :param name: str + Name of the clean room. + :param clean_room: :class:`CleanRoom` (optional) + + + :returns: :class:`CleanRoom` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the clean room. + requestBody: + content: + application/json: + schema: + type: object + properties: + clean_room: + type: string + description: ':returns: :class:`CleanRoom`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CleanRoom' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CleanRoomsAPI +components: + schemas: + CleanRoom: + type: object + properties: + access_restricted: + $ref: '#/components/schemas/CleanRoomAccessRestricted' + comment: + type: string + created_at: + type: integer + description: When the clean room was created, in epoch milliseconds. + local_collaborator_alias: + type: string + description: The alias of the collaborator tied to the local clean room. + name: + type: string + description: >- + The name of the clean room. It should follow [UC securable naming + requirements]. [UC securable naming requirements]: + https://docs.databricks.com/en/data-governance/unity-catalog/index.html#securable-object-naming-requirements + output_catalog: + $ref: '#/components/schemas/CleanRoomOutputCatalog' + description: >- + Output catalog of the clean room. It is an output only field. Output + catalog is manipulated using the separate + CreateCleanRoomOutputCatalog API. + owner: + type: string + description: >- + This is the Databricks username of the owner of the local clean room + securable for permission management. + remote_detailed_info: + $ref: '#/components/schemas/CleanRoomRemoteDetail' + description: >- + Central clean room details. During creation, users need to specify + cloud_vendor, region, and collaborators.global_metastore_id. This + field will not be filled in the ListCleanRooms call. + status: + $ref: '#/components/schemas/CleanRoomStatusEnum' + description: Clean room status. + updated_at: + type: integer + description: When the clean room was last updated, in epoch milliseconds. + CleanRoomAsset: + type: object + properties: + name: + type: string + description: >- + A fully qualified name that uniquely identifies the asset within the + clean room. This is also the name displayed in the clean room UI. + For UC securable assets (tables, volumes, etc.), the format is + *shared_catalog*.*shared_schema*.*asset_name* For notebooks, the + name is the notebook file name. For jar analyses, the name is the + jar analysis name. + asset_type: + $ref: '#/components/schemas/CleanRoomAssetAssetType' + description: The type of the asset. + added_at: + type: integer + description: When the asset is added to the clean room, in epoch milliseconds. + clean_room_name: + type: string + description: >- + The name of the clean room this asset belongs to. This field is + required for create operations and populated by the server for + responses. + foreign_table: + $ref: '#/components/schemas/CleanRoomAssetForeignTable' + description: >- + Foreign table details available to all collaborators of the clean + room. Present if and only if **asset_type** is **FOREIGN_TABLE** + foreign_table_local_details: + $ref: '#/components/schemas/CleanRoomAssetForeignTableLocalDetails' + description: >- + Local details for a foreign that are only available to its owner. + Present if and only if **asset_type** is **FOREIGN_TABLE** + notebook: + $ref: '#/components/schemas/CleanRoomAssetNotebook' + description: >- + Notebook details available to all collaborators of the clean room. + Present if and only if **asset_type** is **NOTEBOOK_FILE** + owner_collaborator_alias: + type: string + description: The alias of the collaborator who owns this asset + status: + $ref: '#/components/schemas/CleanRoomAssetStatusEnum' + description: Status of the asset + table: + $ref: '#/components/schemas/CleanRoomAssetTable' + description: >- + Table details available to all collaborators of the clean room. + Present if and only if **asset_type** is **TABLE** + table_local_details: + $ref: '#/components/schemas/CleanRoomAssetTableLocalDetails' + description: >- + Local details for a table that are only available to its owner. + Present if and only if **asset_type** is **TABLE** + view: + $ref: '#/components/schemas/CleanRoomAssetView' + description: >- + View details available to all collaborators of the clean room. + Present if and only if **asset_type** is **VIEW** + view_local_details: + $ref: '#/components/schemas/CleanRoomAssetViewLocalDetails' + description: >- + Local details for a view that are only available to its owner. + Present if and only if **asset_type** is **VIEW** + volume_local_details: + $ref: '#/components/schemas/CleanRoomAssetVolumeLocalDetails' + description: >- + Local details for a volume that are only available to its owner. + Present if and only if **asset_type** is **VOLUME** + required: + - name + - asset_type + description: Metadata of the clean room asset + CleanRoomAssetForeignTable: + type: object + properties: + columns: + type: string + CleanRoomAssetForeignTableLocalDetails: + type: object + properties: + local_name: + type: string + required: + - local_name + CleanRoomAssetNotebook: + type: object + properties: + notebook_content: + type: string + etag: + type: string + description: Server generated etag that represents the notebook version. + review_state: + $ref: '#/components/schemas/CleanRoomNotebookReviewNotebookReviewState' + description: Top-level status derived from all reviews + reviews: + type: array + items: + $ref: '#/components/schemas/CleanRoomNotebookReview' + description: All existing approvals or rejections + runner_collaborator_aliases: + type: array + items: + type: string + description: Aliases of collaborators that can run the notebook. + required: + - notebook_content + CleanRoomAssetTable: + type: object + properties: + columns: + type: string + CleanRoomAssetTableLocalDetails: + type: object + properties: + local_name: + type: string + partitions: + type: string + description: Partition filtering specification for a shared table. + required: + - local_name + CleanRoomAssetView: + type: object + properties: + columns: + type: string + CleanRoomAssetViewLocalDetails: + type: object + properties: + local_name: + type: string + required: + - local_name + CleanRoomAssetVolumeLocalDetails: + type: object + properties: + local_name: + type: string + required: + - local_name + CleanRoomAutoApprovalRule: + type: object + properties: + author_collaborator_alias: + type: string + author_scope: + $ref: '#/components/schemas/CleanRoomAutoApprovalRuleAuthorScope' + description: >- + Scope of authors covered by the rule. Only one of + `author_collaborator_alias` and `author_scope` can be set. + clean_room_name: + type: string + description: The name of the clean room this auto-approval rule belongs to. + created_at: + type: integer + description: Timestamp of when the rule was created, in epoch milliseconds. + rule_id: + type: string + description: A generated UUID identifying the rule. + rule_owner_collaborator_alias: + type: string + description: The owner of the rule to whom the rule applies. + runner_collaborator_alias: + type: string + description: Collaborator alias of the runner covered by the rule. + CleanRoomCollaborator: + type: object + properties: + collaborator_alias: + type: string + description: >- + Collaborator alias specified by the clean room creator. It is unique + across all collaborators of this clean room, and used to derive + multiple values internally such as catalog alias and clean room name + for single metastore clean rooms. It should follow [UC securable + naming requirements]. [UC securable naming requirements]: + https://docs.databricks.com/en/data-governance/unity-catalog/index.html#securable-object-naming-requirements + display_name: + type: string + description: >- + Generated display name for the collaborator. In the case of a single + metastore clean room, it is the clean room name. For x-metastore + clean rooms, it is the organization name of the metastore. It is not + restricted to these values and could change in the future + global_metastore_id: + type: string + description: >- + The global Unity Catalog metastore ID of the collaborator. The + identifier is of format cloud:region:metastore-uuid. + invite_recipient_email: + type: string + description: >- + Email of the user who is receiving the clean room "invitation". It + should be empty for the creator of the clean room, and non-empty for + the invitees of the clean room. It is only returned in the output + when clean room creator calls GET + invite_recipient_workspace_id: + type: integer + description: >- + Workspace ID of the user who is receiving the clean room + "invitation". Must be specified if invite_recipient_email is + specified. It should be empty when the collaborator is the creator + of the clean room. + organization_name: + type: string + description: >- + [Organization + name](:method:metastores/list#metastores-delta_sharing_organization_name) + configured in the metastore + required: + - collaborator_alias + description: Publicly visible clean room collaborator. + CleanRoomNotebookReview: + type: object + properties: + comment: + type: string + created_at_millis: + type: integer + description: When the review was submitted, in epoch milliseconds + review_state: + $ref: '#/components/schemas/CleanRoomNotebookReviewNotebookReviewState' + description: Review outcome + review_sub_reason: + $ref: '#/components/schemas/CleanRoomNotebookReviewNotebookReviewSubReason' + description: Specified when the review was not explicitly made by a user + reviewer_collaborator_alias: + type: string + description: Collaborator alias of the reviewer + CleanRoomNotebookTaskRun: + type: object + properties: + collaborator_job_run_info: + $ref: '#/components/schemas/CollaboratorJobRunInfo' + description: >- + Job run info of the task in the runner's local workspace. This field + is only included in the LIST API. if the task was run within the + same workspace the API is being called. If the task run was in a + different workspace under the same metastore, only the workspace_id + is included. + notebook_etag: + type: string + description: >- + Etag of the notebook executed in this task run, used to identify the + notebook version. + notebook_job_run_state: + type: string + description: State of the task run. + notebook_name: + type: string + description: Asset name of the notebook executed in this task run. + notebook_updated_at: + type: integer + description: The timestamp of when the notebook was last updated. + output_schema_expiration_time: + type: integer + description: >- + Expiration time of the output schema of the task run (if any), in + epoch milliseconds. + output_schema_name: + type: string + description: >- + Name of the output schema associated with the clean rooms notebook + task run. + run_duration: + type: integer + description: Duration of the task run, in milliseconds. + start_time: + type: integer + description: When the task run started, in epoch milliseconds. + description: Stores information about a single task run. + CleanRoomOutputCatalog: + type: object + properties: + catalog_name: + type: string + status: + $ref: '#/components/schemas/CleanRoomOutputCatalogOutputCatalogStatus' + CleanRoomRemoteDetail: + type: object + properties: + central_clean_room_id: + type: string + description: Central clean room ID. + cloud_vendor: + type: string + description: Cloud vendor (aws,azure,gcp) of the central clean room. + collaborators: + type: array + items: + $ref: '#/components/schemas/CleanRoomCollaborator' + description: >- + Collaborators in the central clean room. There should one and only + one collaborator in the list that satisfies the owner condition: 1. + It has the creator's global_metastore_id (determined by caller of + CreateCleanRoom). 2. Its invite_recipient_email is empty. + compliance_security_profile: + $ref: '#/components/schemas/ComplianceSecurityProfile' + creator: + $ref: '#/components/schemas/CleanRoomCollaborator' + description: Collaborator who creates the clean room. + egress_network_policy: + type: string + description: Egress network policy to apply to the central clean room workspace. + region: + type: string + description: Region of the central clean room. + description: Publicly visible central clean room details. + CollaboratorJobRunInfo: + type: object + properties: + collaborator_alias: + type: string + collaborator_job_id: + type: integer + description: Job ID of the task run in the collaborator's workspace. + collaborator_job_run_id: + type: integer + description: Job run ID of the task run in the collaborator's workspace. + collaborator_task_run_id: + type: integer + description: Task run ID of the task run in the collaborator's workspace. + collaborator_workspace_id: + type: integer + description: ID of the collaborator's workspace that triggered the task run. + ComplianceSecurityProfile: + type: object + properties: + compliance_standards: + type: string + description: >- + The list of compliance standards that the compliance security + profile is configured to enforce. + is_enabled: + type: boolean + description: Whether the compliance security profile is enabled. + description: >- + The compliance security profile used to process regulated data following + compliance standards. + CreateCleanRoomAssetReviewResponse: + type: object + properties: + notebook_review_state: + $ref: '#/components/schemas/CleanRoomNotebookReviewNotebookReviewState' + notebook_reviews: + type: array + items: + $ref: '#/components/schemas/CleanRoomNotebookReview' + description: All existing notebook approvals or rejections + CreateCleanRoomOutputCatalogResponse: + type: object + properties: + output_catalog: + $ref: '#/components/schemas/CleanRoomOutputCatalog' + DeleteCleanRoomAssetResponse: + type: object + properties: {} + description: >- + Response for delete clean room request. Using an empty message since the + generic Empty proto + does not externd UnshadedMessageMarker. + ListCleanRoomAssetRevisionsResponse: + type: object + properties: + next_page_token: + type: string + revisions: + type: array + items: + $ref: '#/components/schemas/CleanRoomAsset' + ListCleanRoomAssetsResponse: + type: object + properties: + assets: + type: array + items: + $ref: '#/components/schemas/CleanRoomAsset' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. page_token should be set to this value for the + next request (for the next page of results). + ListCleanRoomAutoApprovalRulesResponse: + type: object + properties: + next_page_token: + type: string + rules: + type: array + items: + $ref: '#/components/schemas/CleanRoomAutoApprovalRule' + ListCleanRoomNotebookTaskRunsResponse: + type: object + properties: + next_page_token: + type: string + runs: + type: array + items: + $ref: '#/components/schemas/CleanRoomNotebookTaskRun' + description: Name of the clean room. + ListCleanRoomsResponse: + type: object + properties: + clean_rooms: + type: array + items: + $ref: '#/components/schemas/CleanRoom' + next_page_token: + type: string + description: >- + Opaque token to retrieve the next page of results. Absent if there + are no more pages. page_token should be set to this value for the + next request (for the next page of results). + NotebookVersionReview: + type: object + properties: + etag: + type: string + review_state: + $ref: '#/components/schemas/CleanRoomNotebookReviewNotebookReviewState' + description: Review outcome + comment: + type: string + description: Review comment + required: + - etag + - review_state + CleanRoomAccessRestricted: + type: string + enum: + - CSP_MISMATCH + - NO_RESTRICTION + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomAssetAssetType: + type: string + enum: + - FOREIGN_TABLE + - NOTEBOOK_FILE + - TABLE + - VIEW + - VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomAssetStatusEnum: + type: string + enum: + - ACTIVE + - PENDING + - PERMISSION_DENIED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomAutoApprovalRuleAuthorScope: + type: string + enum: + - ANY_AUTHOR + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomNotebookReviewNotebookReviewState: + type: string + enum: + - APPROVED + - PENDING + - REJECTED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomNotebookReviewNotebookReviewSubReason: + type: string + enum: + - AUTO_APPROVED + - BACKFILLED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomOutputCatalogOutputCatalogStatus: + type: string + enum: + - CREATED + - NOT_CREATED + - NOT_ELIGIBLE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomStatusEnum: + type: string + enum: + - ACTIVE + - DELETED + - FAILED + - PROVISIONING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + asset_revisions: + id: databricks_workspace.cleanrooms.asset_revisions + name: asset_revisions + title: Asset Revisions + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type.value}~1{name}~1revisions~1{etag}/get + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type.value}~1{name}~1revisions/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.revisions + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/asset_revisions/methods/get' + - $ref: '#/components/x-stackQL-resources/asset_revisions/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + assets: + id: databricks_workspace.cleanrooms.assets + name: assets + title: Assets + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.assets + review: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type.value}~1{name}~1reviews/post + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type.value}~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type.value}~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1assets~1{asset_type.value}~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/assets/methods/get' + - $ref: '#/components/x-stackQL-resources/assets/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/assets/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/assets/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/assets/methods/delete' + replace: [] + auto_approval_rules: + id: databricks_workspace.cleanrooms.auto_approval_rules + name: auto_approval_rules + title: Auto Approval Rules + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1auto-approval-rules/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1auto-approval-rules/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.rules + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1auto-approval-rules~1{rule_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1auto-approval-rules~1{rule_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1auto-approval-rules~1{rule_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/auto_approval_rules/methods/get' + - $ref: '#/components/x-stackQL-resources/auto_approval_rules/methods/list' + insert: + - $ref: >- + #/components/x-stackQL-resources/auto_approval_rules/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/auto_approval_rules/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/auto_approval_rules/methods/delete + replace: [] + task_runs: + id: databricks_workspace.cleanrooms.task_runs + name: task_runs + title: Task Runs + methods: + list: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1runs/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.runs + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/task_runs/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + clean_rooms: + id: databricks_workspace.cleanrooms.clean_rooms + name: clean_rooms + title: Clean Rooms + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.clean_rooms + create_output_catalog: + operation: + $ref: >- + #/paths/~1api~12.0~1clean-rooms~1{clean_room_name}~1output-catalogs/post + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1clean-rooms~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/get' + - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/clean_rooms/methods/delete' + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/compute.yaml b/providers/src/databricks_workspace/v00.00.00000/services/compute.yaml index b4499f14..5af5c54a 100644 --- a/providers/src/databricks_workspace/v00.00.00000/services/compute.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/services/compute.yaml @@ -1,4609 +1,9581 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Compute API - description: Compute -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/policies/clusters/list-compliance: - get: - operationId: all-clusters-policy-compliance-listcompliance - externalDocs: - url: https://docs.databricks.com/api/workspace/policycomplianceforclusters/listcompliance - x-stackQL-resource: all_clusters_policy_compliance - x-stackQL-method: listcompliance - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: policy_id - required: true - description: string - in: query - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - clusters: - type: array - items: - type: object - properties: - cluster_id: - type: string - is_compliant: - type: boolean - violations: - type: object - properties: - custom_tags.test_tag: - type: string - next_page_token: - type: string - prev_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - x-stackQL-objectKey: $.clusters - description: Returns the policy compliance status of all clusters that use a - given policy. Clusters could be out of compliance if their policy was updated - after the cluster was last edited. - /api/2.1/clusters/change-owner: - post: - operationId: clusters-changeowner - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/changeowner - x-stackQL-resource: clusters - x-stackQL-method: changeowner - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - owner_username: - type: string - example: - cluster_id: string - owner_username: string - description: Change the owner of the cluster. You must be an admin and the cluster - must be terminated to perform this operation. The service principal application - ID can be supplied as an argument to - /api/2.1/clusters/create: - post: - operationId: clusters-create - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/create - x-stackQL-resource: clusters - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - num_workers: - type: int32 - kind: - type: string - cluster_name: - type: string - spark_version: - type: required - use_ml_runtime: - type: string - is_single_node: - type: boolean - node_type_id: - type: boolean - driver_node_type_id: - type: string - ssh_public_keys: - type: string - autotermination_minutes: - type: Array of string - enable_elastic_disk: - type: int32 - instance_pool_id: - type: boolean - policy_id: - type: string - enable_local_disk_encryption: - type: string - driver_instance_pool_id: - type: boolean - runtime_engine: - type: string - data_security_mode: - type: string - single_user_name: - type: string - apply_policy_default_values: - type: string - autoscale: - type: object - spark_conf: - type: object - aws_attributes: - type: object - custom_tags: - type: object - cluster_log_conf: - type: object - init_scripts: - type: Array of object - spark_env_vars: - type: object - workload_type: - type: object - docker_image: - type: object - clone_from: - type: object - example: - cluster_name: single-node-with-kind-cluster - is_single_node: true - kind: CLASSIC_PREVIEW - spark_version: 14.3.x-scala2.12 - node_type_id: i3.xlarge - aws_attributes: - first_on_demand: 1 - availability: SPOT_WITH_FALLBACK - zone_id: auto - spot_bid_price_percent: 100 - ebs_volume_count: 0 - description: Creates a new Spark cluster. This method will acquire new instances - from the cloud provider if necessary. This method is asynchronous; the returned - /api/2.1/clusters/delete: - post: - operationId: clusters-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/delete - x-stackQL-resource: clusters - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - example: - cluster_id: 1234-567890-frays123 - description: Terminates the Spark cluster with the specified ID. The cluster - is removed asynchronously. Once the termination has completed, the cluster - will be in a - /api/2.1/clusters/edit: - post: - operationId: clusters-edit - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/edit - x-stackQL-resource: clusters - x-stackQL-method: edit - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - num_workers: - type: string - kind: - type: int32 - cluster_name: - type: string - spark_version: - type: string - use_ml_runtime: - type: required - is_single_node: - type: string - node_type_id: - type: boolean - driver_node_type_id: - type: boolean - ssh_public_keys: - type: string - autotermination_minutes: - type: string - enable_elastic_disk: - type: Array of string - instance_pool_id: - type: int32 - policy_id: - type: boolean - enable_local_disk_encryption: - type: string - driver_instance_pool_id: - type: string - runtime_engine: - type: boolean - data_security_mode: - type: string - single_user_name: - type: string - apply_policy_default_values: - type: string - autoscale: - type: object - spark_conf: - type: object - aws_attributes: - type: object - custom_tags: - type: object - cluster_log_conf: - type: object - init_scripts: - type: Array of object - spark_env_vars: - type: object - workload_type: - type: object - docker_image: - type: object - example: - cluster_id: 1202-211320-brick1 - num_workers: 10 - spark_version: 3.3.x-scala2.11 - node_type_id: i3.2xlarge - description: Updates the configuration of a cluster to match the provided attributes - and size. A cluster can be updated if it is in a - /api/2.1/clusters/get: - get: - operationId: clusters-get - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/get - x-stackQL-resource: clusters - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: cluster_id - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: integer - cluster_name: - type: string - spark_version: - type: string - aws_attributes: - type: object - properties: - zone_id: - type: string - first_on_demand: - type: integer - availability: - type: string - spot_bid_price_percent: - type: integer - ebs_volume_count: - type: integer - node_type_id: - type: string - driver_node_type_id: - type: string - autotermination_minutes: - type: integer - enable_elastic_disk: - type: boolean - disk_spec: - type: object - properties: - disk_count: - type: integer - cluster_source: - type: string - enable_local_disk_encryption: - type: boolean - instance_source: - type: object - properties: - node_type_id: - type: string - driver_instance_source: - type: object - properties: - node_type_id: - type: string - state: - type: string - state_message: - type: string - start_time: - type: integer - terminated_time: - type: integer - last_state_loss_time: - type: integer - num_workers: - type: integer - default_tags: - type: object - properties: - Vendor: - type: string - Creator: - type: string - ClusterName: - type: string - ClusterId: - type: string - creator_user_name: - type: string - termination_reason: - type: object - properties: - code: - type: string - parameters: - type: object - properties: - inactivity_duration_min: - type: string - type: - type: string - init_scripts_safe_mode: - type: boolean - spec: - type: object - properties: - spark_version: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves the information for a cluster given its identifier. Clusters - can be described while they are running, or up to 60 days after they are terminated. - /api/2.1/clusters/list: - get: - operationId: clusters-list - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/list - x-stackQL-resource: clusters - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - - name: null - in: query - - name: null - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - clusters: - type: array - items: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: integer - cluster_name: - type: string - spark_version: - type: string - aws_attributes: - type: object - properties: - zone_id: - type: string - first_on_demand: - type: integer - availability: - type: string - spot_bid_price_percent: - type: integer - ebs_volume_count: - type: integer - node_type_id: - type: string - driver_node_type_id: - type: string - autotermination_minutes: - type: integer - enable_elastic_disk: - type: boolean - disk_spec: - type: object - properties: - disk_count: - type: integer - cluster_source: - type: string - enable_local_disk_encryption: - type: boolean - instance_source: - type: object - properties: - node_type_id: - type: string - driver_instance_source: - type: object - properties: - node_type_id: - type: string - state: - type: string - state_message: - type: string - start_time: - type: integer - terminated_time: - type: integer - last_state_loss_time: - type: integer - num_workers: - type: integer - default_tags: - type: object - properties: - Vendor: - type: string - Creator: - type: string - ClusterName: - type: string - ClusterId: - type: string - creator_user_name: - type: string - termination_reason: - type: object - properties: - code: - type: string - parameters: - type: object - properties: - inactivity_duration_min: - type: string - type: - type: string - init_scripts_safe_mode: - type: boolean - prev_page_token: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.clusters - description: Return information about all pinned and active clusters, and all - clusters terminated within the last 30 days. Clusters terminated prior to - this period are not included. - /api/2.1/clusters/permanent-delete: - post: - operationId: clusters-permanentdelete - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/permanentdelete - x-stackQL-resource: clusters - x-stackQL-method: permanentdelete - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - example: - cluster_id: 1202-211320-brick1 - description: Permanently deletes a Spark cluster. This cluster is terminated - and resources are asynchronously removed. - /api/2.1/clusters/pin: - post: - operationId: clusters-pin - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/pin - x-stackQL-resource: clusters - x-stackQL-method: pin - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - example: - cluster_id: 1234-567890-reef123 - description: Pinning a cluster ensures that the cluster will always be returned - by the ListClusters API. Pinning a cluster that is already pinned will have - no effect. This API can only be called by workspace admins. - /api/2.1/clusters/resize: - post: - operationId: clusters-resize - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/resize - x-stackQL-resource: clusters - x-stackQL-method: resize - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - num_workers: - type: int32 - cluster_id: - type: required - autoscale: - type: object - example: - cluster_id: 1202-211320-brick1 - num_workers: 30 - description: Resizes a cluster to have a desired number of workers. This will - fail unless the cluster is in a - /api/2.1/clusters/restart: - post: - operationId: clusters-restart - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/restart - x-stackQL-resource: clusters - x-stackQL-method: restart - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - restart_user: - type: string - example: - cluster_id: 1202-211320-brick1 - description: Restarts a Spark cluster with the supplied ID. If the cluster is - not currently in a - /api/2.1/clusters/start: - post: - operationId: clusters-start - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/start - x-stackQL-resource: clusters - x-stackQL-method: start - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - example: - cluster_id: 1202-211320-brick1 - description: Starts a terminated Spark cluster with the supplied ID. This works - similar to - /api/2.1/clusters/unpin: - post: - operationId: clusters-unpin - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/unpin - x-stackQL-resource: clusters - x-stackQL-method: unpin - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - example: - cluster_id: 1234-567890-reef123 - description: Unpinning a cluster will allow the cluster to eventually be removed - from the ListClusters API. Unpinning a cluster that is not pinned will have - no effect. This API can only be called by workspace admins. - /api/2.1/clusters/update: - post: - operationId: clusters-update - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/update - x-stackQL-resource: clusters - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - update_mask: - type: string - cluster: - type: object - example: - cluster_id: 1202-211320-brick1 - cluster: - num_workers: 10 - spark_version: 14.3.x-scala2.12 - spark_conf: - spark.databricks.cluster.profile: singleNode - spark.master: '[*, 4]' - update_mask: num_workers,spark_version,spark_conf - description: Updates the configuration of a cluster to match the partial set - of attributes and size. Denote which fields to update using the - /api/2.1/clusters/events: - post: - operationId: cluster-events-events - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/events - x-stackQL-resource: cluster_events - x-stackQL-method: events - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - events: - type: array - items: - type: object - properties: - cluster_id: - type: string - timestamp: - type: integer - event_type: - type: string - event_details: - type: object - properties: - username: - type: string - next_page: - type: object - properties: - cluster_id: - type: string - end_time: - type: integer - order: - type: string - offset: - type: integer - total_count: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - start_time: - type: string - end_time: - type: integer - format: int64 - order: - type: integer - format: int64 - event_types: - type: string - offset: - type: Array of string - limit: - type: integer - format: int64 - example: - start_time: 1617238800000 - end_time: 1619485200000 - offset: 5 - order: DESC - event_types: - - RUNNING - cluster_id: 1234-567890-reef123 - limit: 5 - description: Retrieves a list of events about the activity of a cluster. This - API is paginated. If there are more events to read, the response includes - all the nparameters necessary to request the next page of events. - /api/2.0/permissions/clusters/{cluster_id}: - get: - operationId: cluster-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/getpermissions - x-stackQL-resource: cluster_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: cluster_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a cluster. Clusters can inherit permissions - from their root object. - put: - operationId: cluster-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/setpermissions - x-stackQL-resource: cluster_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: cluster_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: cluster-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/updatepermissions - x-stackQL-resource: cluster_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: cluster_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a cluster. Clusters can inherit permissions - from their root object. - /api/2.0/permissions/clusters/{cluster_id}/permissionLevels: - get: - operationId: cluster-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/getpermissionlevels - x-stackQL-resource: cluster_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: cluster_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/policies/clusters/create: - post: - operationId: cluster-policies-create - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/create - x-stackQL-resource: cluster_policies - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - policy_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - definition: - type: JSON - description: - type: string - policy_family_id: - type: string - policy_family_definition_overrides: - type: JSON - max_clusters_per_user: - type: integer - format: int64 - libraries: - type: Array of object - example: - definition: '{ "custom_tags.test_tag": { "type": "fixed", "value": - "test_value" } }' - name: Test policy - description: Creates a new policy with prescribed settings. - /api/2.0/policies/clusters/delete: - post: - operationId: cluster-policies-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/delete - x-stackQL-resource: cluster_policies - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - policy_id: - type: required - example: - policy_id: string - description: Delete a policy for a cluster. Clusters governed by this policy - can still run, but cannot be edited. - /api/2.0/policies/clusters/edit: - post: - operationId: cluster-policies-edit - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/edit - x-stackQL-resource: cluster_policies - x-stackQL-method: edit - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - policy_id: - type: required - name: - type: string - definition: - type: string - description: - type: JSON - policy_family_id: - type: string - policy_family_definition_overrides: - type: string - max_clusters_per_user: - type: JSON - libraries: - type: Array of object - example: - definition: '{ "custom_tags.test_tag": { "type": "fixed", "value": - "test_value" } }' - name: Test policy - policy_id: ABCD000000000000 - description: Update an existing policy for cluster. This operation may make - some clusters governed by the previous policy invalid. - /api/2.0/policies/clusters/get: - get: - operationId: cluster-policies-get - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/get - x-stackQL-resource: cluster_policies - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: policy_id - required: true - description: string - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - policy_id: - type: string - name: - type: string - definition: - type: object - properties: - custom_tags.test_tag: - type: object - properties: - type: - type: string - value: - type: string - description: - type: string - policy_family_id: - type: string - policy_family_definition_overrides: - type: object - properties: - custom_tags.test_tag: - type: object - properties: - type: - type: string - value: - type: string - max_clusters_per_user: - type: integer - libraries: - type: array - items: - type: object - properties: - jar: - type: string - egg: - type: string - pypi: - type: object - properties: - package: - type: string - repo: - type: string - maven: - type: object - properties: - coordinates: - type: string - repo: - type: string - exclusions: - type: array - items: - type: string - cran: - type: object - properties: - package: - type: string - repo: - type: string - whl: - type: string - requirements: - type: string - creator_user_name: - type: string - created_at_timestamp: - type: integer - is_default: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: Get a cluster policy entity. Creation and editing is available - to admins only. - /api/2.0/policies/clusters/list: - get: - operationId: cluster-policies-list - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/list - x-stackQL-resource: cluster_policies - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: sort_order - in: query - - name: sort_column - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - policies: - type: array - items: - type: object - properties: - policy_id: - type: string - name: - type: string - definition: - type: object - properties: - custom_tags.test_tag: - type: object - properties: - type: - type: string - value: - type: string - description: - type: string - policy_family_id: - type: string - policy_family_definition_overrides: - type: object - properties: - custom_tags.test_tag: - type: object - properties: - type: - type: string - value: - type: string - max_clusters_per_user: - type: integer - libraries: - type: array - items: - type: object - properties: - jar: - type: string - egg: - type: string - pypi: - type: object - properties: - package: - type: string - repo: - type: string - maven: - type: object - properties: - coordinates: - type: string - repo: - type: string - exclusions: - type: array - items: - type: string - cran: - type: object - properties: - package: - type: string - repo: - type: string - whl: - type: string - requirements: - type: string - creator_user_name: - type: string - created_at_timestamp: - type: integer - is_default: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - '503': - description: Service is currently unavailable. - '504': - description: Deadline expired before the operation could complete. - x-stackQL-objectKey: $.policies - description: Returns a list of policies accessible by the requesting user. - /api/2.0/permissions/cluster-policies/{cluster_policy_id}: - get: - operationId: cluster-policies-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/getpermissions - x-stackQL-resource: cluster_policies_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: cluster_policy_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a cluster policy. Cluster policies can - inherit permissions from their root object. - put: - operationId: cluster-policies-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/setpermissions - x-stackQL-resource: cluster_policies_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: cluster_policy_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_USE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: cluster-policies-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/updatepermissions - x-stackQL-resource: cluster_policies_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: cluster_policy_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_USE - description: Updates the permissions on a cluster policy. Cluster policies can - inherit permissions from their root object. - /api/2.0/policies/clusters/enforce-compliance: - post: - operationId: cluster-policy-compliance-enforcecompliance - externalDocs: - url: https://docs.databricks.com/api/workspace/policycomplianceforclusters/enforcecompliance - x-stackQL-resource: cluster_policy_compliance - x-stackQL-method: enforcecompliance - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - has_changes: - type: boolean - changes: - type: array - items: - type: object - properties: - field: - type: string - previous_value: - type: string - new_value: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - validate_only: - type: string - example: - cluster_id: string - validate_only: false - description: Updates a cluster to be compliant with the current version of its - policy. A cluster can be updated if it is in a - /api/2.0/policies/clusters/get-compliance: - get: - operationId: cluster-policy-compliance-getcompliance - externalDocs: - url: https://docs.databricks.com/api/workspace/policycomplianceforclusters/getcompliance - x-stackQL-resource: cluster_policy_compliance - x-stackQL-method: getcompliance - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: cluster_id - required: true - description: string - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - is_compliant: - type: boolean - violations: - type: object - properties: - custom_tags.test_tag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: Returns the policy compliance status of a cluster. Clusters could - be out of compliance if their policy was updated after the cluster was last - edited. - /api/1.2/commands/cancel: - post: - operationId: commands-cancel - externalDocs: - url: https://docs.databricks.com/api/workspace/commandexecution/cancel - x-stackQL-resource: commands - x-stackQL-method: cancel - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Status was returned successfully. - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - clusterId: - type: string - contextId: - type: string - commandId: - type: string - example: - clusterId: string - contextId: string - commandId: string - description: Cancels a currently running command within an execution context. - /api/1.2/commands/status: - get: - operationId: commands-commandstatus - externalDocs: - url: https://docs.databricks.com/api/workspace/commandexecution/commandstatus - x-stackQL-resource: commands - x-stackQL-method: commandstatus - x-stackQL-verb: select - x-numReqParams: 3 - parameters: - - name: clusterId - required: true - description: string - in: query - - name: contextId - required: true - description: string - in: query - - name: commandId - required: true - description: string - in: query - responses: - '200': - description: Status was returned successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - status: - type: string - results: - type: object - properties: - resultType: - type: string - data: - type: string - description: Gets the status of and, if available, the results from a currently - executing command. - /api/1.2/contexts/create: - post: - operationId: commands-create - externalDocs: - url: https://docs.databricks.com/api/workspace/commandexecution/create - x-stackQL-resource: commands - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Status was returned successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - clusterId: - type: string - language: - type: string - example: - clusterId: string - language: python - description: Creates an execution context for running cluster commands. - /api/1.2/contexts/destroy: - post: - operationId: commands-destroy - externalDocs: - url: https://docs.databricks.com/api/workspace/commandexecution/destroy - x-stackQL-resource: commands - x-stackQL-method: destroy - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Status was returned successfully. - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - clusterId: - type: required - contextId: - type: string - example: - clusterId: string - contextId: string - description: Deletes an execution context. - /api/1.2/commands/execute: - post: - operationId: commands-execute - externalDocs: - url: https://docs.databricks.com/api/workspace/commandexecution/execute - x-stackQL-resource: commands - x-stackQL-method: execute - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Status was returned successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - clusterId: - type: string - contextId: - type: string - language: - type: string - command: - type: string - example: - clusterId: 1234-567890-span123 - contextId: '1234567890123456789' - language: python - command: print('Hello, World!') - description: Runs a cluster command in the given execution context, using the - provided language. - /api/1.2/contexts/status: - get: - operationId: command-status-contextstatus - externalDocs: - url: https://docs.databricks.com/api/workspace/commandexecution/contextstatus - x-stackQL-resource: command_status - x-stackQL-method: contextstatus - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: clusterId - required: true - description: string - in: query - - name: contextId - required: true - description: string - in: query - responses: - '200': - description: Status was returned successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - status: - type: string - description: Gets the status for an execution context. - /api/2.0/global-init-scripts: - post: - operationId: global-init-scripts-create - externalDocs: - url: https://docs.databricks.com/api/workspace/globalinitscripts/create - x-stackQL-resource: global_init_scripts - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Response indicating that the script was created successfully. - content: - application/json: - schema: - type: object - properties: - script_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - script: - type: string - position: - type: required - enabled: - type: byte - example: - name: My example script name - script: ZWNobyBoZWxsbw== - position: 0 - enabled: false - description: Manages global init scripts for this workspace. - get: - operationId: global-init-scripts-list - externalDocs: - url: https://docs.databricks.com/api/workspace/globalinitscripts/list - x-stackQL-resource: global_init_scripts - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Response indicating that the scripts were retrieved successfully. - content: - application/json: - schema: - type: object - properties: - scripts: - type: array - items: - type: object - properties: - script_id: - type: string - name: - type: string - position: - type: integer - enabled: - type: boolean - created_by: - type: string - created_at: - type: integer - updated_by: - type: string - updated_at: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.scripts - description: Manages global init scripts for this workspace. - /api/2.0/global-init-scripts/{script_id}: - delete: - operationId: global-init-scripts-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/globalinitscripts/delete - x-stackQL-resource: global_init_scripts - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: script_id - required: true - description: string - in: path - responses: - '200': - description: The script was deleted successfully. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Manage a specific global init script with ID - get: - operationId: global-init-scripts-get - externalDocs: - url: https://docs.databricks.com/api/workspace/globalinitscripts/get - x-stackQL-resource: global_init_scripts - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: script_id - required: true - description: string - in: path - responses: - '200': - description: Response indicating that the script was retrieved successfully. - content: - application/json: - schema: - type: object - properties: - script_id: - type: string - name: - type: string - position: - type: integer - enabled: - type: boolean - created_by: - type: string - created_at: - type: integer - updated_by: - type: string - updated_at: - type: integer - script: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Manage a specific global init script with ID - patch: - operationId: global-init-scripts-update - externalDocs: - url: https://docs.databricks.com/api/workspace/globalinitscripts/update - x-stackQL-resource: global_init_scripts - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: script_id - required: true - description: string - in: path - responses: - '200': - description: The script was updated successfully. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - script: - type: string - position: - type: required - enabled: - type: byte - example: - name: My example script name - script: ZWNobyBoZWxsbw== - position: 0 - enabled: false - description: Manage a specific global init script with ID - /api/2.0/instance-pools/create: - post: - operationId: instance-pools-create - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/create - x-stackQL-resource: instance_pools - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - instance_pool_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - instance_pool_name: - type: required - min_idle_instances: - type: string - max_capacity: - type: int32 - node_type_id: - type: int32 - idle_instance_autotermination_minutes: - type: required - enable_elastic_disk: - type: string - preloaded_spark_versions: - type: int32 - aws_attributes: - type: object - custom_tags: - type: object - disk_spec: - type: object - preloaded_docker_images: - type: Array of object - example: - instance_pool_name: my-pool - node_type_id: i3.xlarge - min_idle_instances: 10 - aws_attributes: - availability: SPOT - custom_tags: - - key: my-key - value: my-value - description: Creates a new instance pool using idle and ready-to-use cloud instances. - /api/2.0/instance-pools/delete: - post: - operationId: instance-pools-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/delete - x-stackQL-resource: instance_pools - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - instance_pool_id: - type: required - example: - instance_pool_id: 1234-567890-fetch12-pool-A3BcdEFg - description: Deletes the instance pool permanently. The idle instances in the - pool are terminated asynchronously. - /api/2.0/instance-pools/edit: - post: - operationId: instance-pools-edit - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/edit - x-stackQL-resource: instance_pools - x-stackQL-method: edit - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - instance_pool_name: - type: required - min_idle_instances: - type: string - max_capacity: - type: int32 - idle_instance_autotermination_minutes: - type: int32 - instance_pool_id: - type: int32 - node_type_id: - type: required - custom_tags: - type: object - example: - max_capacity: 200 - idle_instance_autotermination_minutes: 30 - min_idle_instances: 5 - instance_pool_id: 1234-567890-fetch12-pool-A3BcdEFg - instance_pool_name: my-edited-pool - description: Modifies the configuration of an existing instance pool. - /api/2.0/instance-pools/get: - get: - operationId: instance-pools-get - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/get - x-stackQL-resource: instance_pools - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: instance_pool_id - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - instance_pool_name: - type: string - min_idle_instances: - type: integer - aws_attributes: - type: object - properties: - availability: - type: string - zone_id: - type: string - spot_bid_price_percent: - type: integer - node_type_id: - type: string - custom_tags: - type: object - properties: - my-key: - type: string - idle_instance_autotermination_minutes: - type: integer - enable_elastic_disk: - type: boolean - disk_spec: - type: object - properties: - disk_type: - type: object - properties: - ebs_volume_type: - type: string - disk_count: - type: integer - disk_size: - type: integer - preloaded_spark_versions: - type: array - items: - type: string - instance_pool_id: - type: string - default_tags: - type: object - properties: - Vendor: - type: string - DatabricksInstancePoolCreatorId: - type: string - DatabricksInstancePoolId: - type: string - state: - type: string - stats: - type: object - properties: - used_count: - type: integer - idle_count: - type: integer - pending_used_count: - type: integer - pending_idle_count: - type: integer - status: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieve the information for an instance pool based on its identifier. - /api/2.0/instance-pools/list: - get: - operationId: instance-pools-list - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/list - x-stackQL-resource: instance_pools - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - instance_pools: - type: array - items: - type: object - properties: - instance_pool_name: - type: string - min_idle_instances: - type: integer - aws_attributes: - type: object - properties: - availability: - type: string - zone_id: - type: string - spot_bid_price_percent: - type: integer - node_type_id: - type: string - idle_instance_autotermination_minutes: - type: integer - enable_elastic_disk: - type: boolean - disk_spec: - type: object - properties: - disk_type: - type: object - properties: - ebs_volume_type: - type: string - disk_count: - type: integer - disk_size: - type: integer - preloaded_spark_versions: - type: array - items: - type: string - instance_pool_id: - type: string - default_tags: - type: object - properties: - Vendor: - type: string - DatabricksInstancePoolCreatorId: - type: string - DatabricksInstancePoolId: - type: string - state: - type: string - stats: - type: object - properties: - used_count: - type: integer - idle_count: - type: integer - pending_used_count: - type: integer - pending_idle_count: - type: integer - status: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.instance_pools - description: Gets a list of instance pools with their statistics. - /api/2.0/permissions/instance-pools/{instance_pool_id}: - get: - operationId: instance-pool-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/getpermissions - x-stackQL-resource: instance_pool_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: instance_pool_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of an instance pool. Instance pools can inherit - permissions from their root object. - put: - operationId: instance-pool-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/setpermissions - x-stackQL-resource: instance_pool_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: instance_pool_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: instance-pool-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/updatepermissions - x-stackQL-resource: instance_pool_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: instance_pool_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on an instance pool. Instance pools can - inherit permissions from their root object. - /api/2.0/permissions/instance-pools/{instance_pool_id}/permissionLevels: - get: - operationId: instance-pool-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/instancepools/getpermissionlevels - x-stackQL-resource: instance_pool_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: instance_pool_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/instance-profiles/add: - post: - operationId: instance-profiles-add - externalDocs: - url: https://docs.databricks.com/api/workspace/instanceprofiles/add - x-stackQL-resource: instance_profiles - x-stackQL-method: add - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - skip_validation: - type: boolean - instance_profile_arn: - type: required - iam_role_arn: - type: string - is_meta_instance_profile: - type: string - example: - instance_profile_arn: arn:aws:iam::123456789012:instance-profile/my-profile - description: In the UI, you can select the instance profile when launching clusters. - This API is only available to admin users. - /api/2.0/instance-profiles/edit: - post: - operationId: instance-profiles-edit - externalDocs: - url: https://docs.databricks.com/api/workspace/instanceprofiles/edit - x-stackQL-resource: instance_profiles - x-stackQL-method: edit - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - instance_profile_arn: - type: required - iam_role_arn: - type: string - is_meta_instance_profile: - type: string - example: - instance_profile_arn: arn:aws:iam::123456789012:instance-profile/my-profile - iam_role_arn: arn:aws:iam::123456789012:role/my-profile2 - description: 'The only supported field to change is the optional IAM role ARN - associated with the instance profile. It is required to specify the IAM role - ARN if both of the following are true:' - /api/2.0/instance-profiles/list: - get: - operationId: instance-profiles-list - externalDocs: - url: https://docs.databricks.com/api/workspace/instanceprofiles/list - x-stackQL-resource: instance_profiles - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - instance_profiles: - type: array - items: - type: object - properties: - instance_profile_arn: - type: string - is_meta_instance_profile: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.instance_profiles - description: List the instance profiles that the calling user can use to launch - a cluster. - /api/2.0/instance-profiles/remove: - post: - operationId: instance-profiles-remove - externalDocs: - url: https://docs.databricks.com/api/workspace/instanceprofiles/remove - x-stackQL-resource: instance_profiles - x-stackQL-method: remove - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - instance_profile_arn: - type: required - example: - instance_profile_arn: arn:aws:iam::123456789012:instance-profile/my-profile - description: Remove the instance profile with the provided ARN. Existing clusters - with this instance profile will continue to function. - /api/2.0/libraries/all-cluster-statuses: - get: - operationId: libraries-allclusterlibrarystatuses - externalDocs: - url: https://docs.databricks.com/api/workspace/libraries/allclusterlibrarystatuses - x-stackQL-resource: libraries - x-stackQL-method: allclusterlibrarystatuses - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - statuses: - type: array - items: - type: object - properties: - cluster_id: - type: string - library_statuses: - type: array - items: - type: object - properties: - is_library_for_all_clusters: - type: boolean - library: - type: object - properties: - pypi: - type: object - properties: - package: - type: string - repo: - type: string - status: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - x-stackQL-objectKey: $.statuses - description: Get the status of all libraries on all clusters. A status is returned - for all libraries installed on this cluster via the API or the libraries UI. - /api/2.0/libraries/cluster-status: - get: - operationId: libraries-clusterstatus - externalDocs: - url: https://docs.databricks.com/api/workspace/libraries/clusterstatus - x-stackQL-resource: libraries - x-stackQL-method: clusterstatus - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: cluster_id - required: true - description: string - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: string - library_statuses: - type: array - items: - type: object - properties: - is_library_for_all_clusters: - type: boolean - library: - type: object - properties: - pypi: - type: object - properties: - package: - type: string - repo: - type: string - status: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: 'Get the status of libraries on a cluster. A status is returned - for all libraries installed on this cluster via the API or the libraries UI. - The order of returned libraries is as follows:' - /api/2.0/libraries/install: - post: - operationId: libraries-install - externalDocs: - url: https://docs.databricks.com/api/workspace/libraries/install - x-stackQL-resource: libraries - x-stackQL-method: install - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - libraries: - type: required - example: - cluster_id: 1234-56789-abcde - libraries: - - pypi: - package: numpy - repo: http://my-pypi-repo.com - - jar: /Workspace/path/to/library.jar - - whl: /Workspace/path/to/library.whl - - maven: - coordinates: com.databricks:spark-csv_2.11:1.5.0 - exclusions: - - org.slf4j:slf4j-log4j12 - repo: http://my-maven-repo.com - - cran: - package: ggplot2 - repo: http://cran.us.r-project.org - - requirements: /Workspace/path/to/requirements.txt - description: Add libraries to install on a cluster. The installation is asynchronous; - it happens in the background after the completion of this request. - /api/2.0/libraries/uninstall: - post: - operationId: libraries-uninstall - externalDocs: - url: https://docs.databricks.com/api/workspace/libraries/uninstall - x-stackQL-resource: libraries - x-stackQL-method: uninstall - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - cluster_id: - type: required - libraries: - type: required - example: - cluster_id: 1234-56789-abcde - libraries: - - pypi: - package: numpy - repo: http://my-pypi-repo.com - - jar: /Workspace/path/to/library.jar - - whl: /Workspace/path/to/library.whl - - maven: - coordinates: com.databricks:spark-csv_2.11:1.5.0 - exclusions: - - org.slf4j:slf4j-log4j12 - repo: http://my-maven-repo.com - - cran: - package: ggplot2 - repo: http://cran.us.r-project.org - - requirements: /Workspace/path/to/requirements.txt - description: Set libraries to uninstall from a cluster. The libraries won't - be uninstalled until the cluster is restarted. A request to uninstall a library - that is not currently installed is ignored. - /api/2.1/clusters/list-node-types: - get: - operationId: node-types-listnodetypes - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/listnodetypes - x-stackQL-resource: node_types - x-stackQL-method: listnodetypes - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - node_types: - type: array - items: - type: object - properties: - node_type_id: - type: string - memory_mb: - type: integer - num_cores: - type: integer - description: - type: string - instance_type_id: - type: string - is_deprecated: - type: boolean - category: - type: string - support_ebs_volumes: - type: boolean - support_cluster_tags: - type: boolean - num_gpus: - type: integer - node_instance_type: - type: object - properties: - instance_type_id: - type: string - local_disks: - type: integer - local_disk_size_gb: - type: integer - instance_family: - type: string - swap_size: - type: string - is_hidden: - type: boolean - support_port_forwarding: - type: boolean - supports_elastic_disk: - type: boolean - display_order: - type: integer - is_io_cache_enabled: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.node_types - description: Returns a list of supported Spark node types. These node types - can be used to launch a cluster. - /api/2.0/policy-families/{policy_family_id}: - get: - operationId: policy-families-get - externalDocs: - url: https://docs.databricks.com/api/workspace/policyfamilies/get - x-stackQL-resource: policy_families - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: policy_family_id - required: true - description: string - in: path - - name: version - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - policy_family_id: - type: string - name: - type: string - description: - type: string - definition: - type: object - properties: - custom_tags.test_tag: - type: object - properties: - type: - type: string - value: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieve the information for an policy family based on its identifier - and version - /api/2.0/policy-families: - get: - operationId: policy-families-list - externalDocs: - url: https://docs.databricks.com/api/workspace/policyfamilies/list - x-stackQL-resource: policy_families - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - policy_families: - type: array - items: - type: object - properties: - policy_family_id: - type: string - name: - type: string - description: - type: string - definition: - type: object - properties: - custom_tags.test_tag: - type: object - properties: - type: - type: string - value: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.policy_families - description: Returns the list of policy definition types available to use at - their latest version. This API is paginated. - /api/2.0/permissions/cluster-policies/{cluster_policy_id}/permissionLevels: - get: - operationId: policy-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/clusterpolicies/getpermissionlevels - x-stackQL-resource: policy_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: cluster_policy_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.1/clusters/spark-versions: - get: - operationId: spark-versions-sparkversions - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/sparkversions - x-stackQL-resource: spark_versions - x-stackQL-method: sparkversions - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - versions: - type: array - items: - type: object - properties: - key: - type: string - name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.versions - description: Returns the list of available Spark versions. These versions can - be used to launch a cluster. - /api/2.1/clusters/list-zones: - get: - operationId: zones-listzones - externalDocs: - url: https://docs.databricks.com/api/workspace/clusters/listzones - x-stackQL-resource: zones - x-stackQL-method: listzones - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - zones: - type: array - items: - type: string - default_zone: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - description: Returns a list of availability zones where clusters can be created - in (For example, us-west-2a). These zones can be used to launch a cluster. -components: - x-stackQL-resources: - all_clusters_policy_compliance: - id: databricks_workspace.compute.all_clusters_policy_compliance - name: all_clusters_policy_compliance - title: All_clusters_policy_compliance - methods: - listcompliance: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1list-compliance/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.clusters - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/all_clusters_policy_compliance/methods/listcompliance' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - clusters: - id: databricks_workspace.compute.clusters - name: clusters - title: Clusters - methods: - changeowner: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1change-owner/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - create: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - edit: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1edit/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.clusters - permanentdelete: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1permanent-delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - pin: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1pin/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - resize: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1resize/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - restart: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1restart/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - start: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1start/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - unpin: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1unpin/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1update/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/clusters/methods/list' - - $ref: '#/components/x-stackQL-resources/clusters/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/clusters/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/clusters/methods/update' - replace: - - $ref: '#/components/x-stackQL-resources/clusters/methods/edit' - delete: - - $ref: '#/components/x-stackQL-resources/clusters/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/clusters/methods/unpin' - - $ref: '#/components/x-stackQL-resources/clusters/methods/start' - - $ref: '#/components/x-stackQL-resources/clusters/methods/restart' - - $ref: '#/components/x-stackQL-resources/clusters/methods/resize' - - $ref: '#/components/x-stackQL-resources/clusters/methods/pin' - - $ref: '#/components/x-stackQL-resources/clusters/methods/permanentdelete' - - $ref: '#/components/x-stackQL-resources/clusters/methods/changeowner' - cluster_events: - id: databricks_workspace.compute.cluster_events - name: cluster_events - title: Cluster_events - methods: - events: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1events/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/cluster_events/methods/events' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - cluster_permissions: - id: databricks_workspace.compute.cluster_permissions - name: cluster_permissions - title: Cluster_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/cluster_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/cluster_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/cluster_permissions/methods/setpermissions' - delete: [] - exec: [] - cluster_permission_levels: - id: databricks_workspace.compute.cluster_permission_levels - name: cluster_permission_levels - title: Cluster_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/cluster_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - cluster_policies: - id: databricks_workspace.compute.cluster_policies - name: cluster_policies - title: Cluster_policies - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - edit: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1edit/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.policies - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/list' - - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/edit' - delete: - - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/delete' - exec: [] - cluster_policies_permissions: - id: databricks_workspace.compute.cluster_policies_permissions - name: cluster_policies_permissions - title: Cluster_policies_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/cluster_policies_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/cluster_policies_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/cluster_policies_permissions/methods/setpermissions' - delete: [] - exec: [] - cluster_policy_compliance: - id: databricks_workspace.compute.cluster_policy_compliance - name: cluster_policy_compliance - title: Cluster_policy_compliance - methods: - enforcecompliance: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1enforce-compliance/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getcompliance: - operation: - $ref: '#/paths/~1api~12.0~1policies~1clusters~1get-compliance/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/cluster_policy_compliance/methods/getcompliance' - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/cluster_policy_compliance/methods/enforcecompliance' - delete: [] - exec: [] - commands: - id: databricks_workspace.compute.commands - name: commands - title: Commands - methods: - cancel: - operation: - $ref: '#/paths/~1api~11.2~1commands~1cancel/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - commandstatus: - operation: - $ref: '#/paths/~1api~11.2~1commands~1status/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - create: - operation: - $ref: '#/paths/~1api~11.2~1contexts~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - destroy: - operation: - $ref: '#/paths/~1api~11.2~1contexts~1destroy/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - execute: - operation: - $ref: '#/paths/~1api~11.2~1commands~1execute/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/commands/methods/commandstatus' - insert: - - $ref: '#/components/x-stackQL-resources/commands/methods/create' - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/commands/methods/execute' - - $ref: '#/components/x-stackQL-resources/commands/methods/destroy' - - $ref: '#/components/x-stackQL-resources/commands/methods/cancel' - command_status: - id: databricks_workspace.compute.command_status - name: command_status - title: Command_status - methods: - contextstatus: - operation: - $ref: '#/paths/~1api~11.2~1contexts~1status/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/command_status/methods/contextstatus' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - global_init_scripts: - id: databricks_workspace.compute.global_init_scripts - name: global_init_scripts - title: Global_init_scripts - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1global-init-scripts/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1global-init-scripts/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.scripts - delete: - operation: - $ref: '#/paths/~1api~12.0~1global-init-scripts~1{script_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1global-init-scripts~1{script_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1global-init-scripts~1{script_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/global_init_scripts/methods/get' - - $ref: '#/components/x-stackQL-resources/global_init_scripts/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/global_init_scripts/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/global_init_scripts/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/global_init_scripts/methods/delete' - exec: [] - instance_pools: - id: databricks_workspace.compute.instance_pools - name: instance_pools - title: Instance_pools - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1instance-pools~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1instance-pools~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - edit: - operation: - $ref: '#/paths/~1api~12.0~1instance-pools~1edit/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1instance-pools~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1instance-pools~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.instance_pools - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/instance_pools/methods/list' - - $ref: '#/components/x-stackQL-resources/instance_pools/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/instance_pools/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/instance_pools/methods/edit' - delete: - - $ref: '#/components/x-stackQL-resources/instance_pools/methods/delete' - exec: [] - instance_pool_permissions: - id: databricks_workspace.compute.instance_pool_permissions - name: instance_pool_permissions - title: Instance_pool_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/instance_pool_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/instance_pool_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/instance_pool_permissions/methods/setpermissions' - delete: [] - exec: [] - instance_pool_permission_levels: - id: databricks_workspace.compute.instance_pool_permission_levels - name: instance_pool_permission_levels - title: Instance_pool_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/instance_pool_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - instance_profiles: - id: databricks_workspace.compute.instance_profiles - name: instance_profiles - title: Instance_profiles - methods: - add: - operation: - $ref: '#/paths/~1api~12.0~1instance-profiles~1add/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - edit: - operation: - $ref: '#/paths/~1api~12.0~1instance-profiles~1edit/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1instance-profiles~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.instance_profiles - remove: - operation: - $ref: '#/paths/~1api~12.0~1instance-profiles~1remove/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/add' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/edit' - delete: - - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/remove' - exec: [] - libraries: - id: databricks_workspace.compute.libraries - name: libraries - title: Libraries - methods: - allclusterlibrarystatuses: - operation: - $ref: '#/paths/~1api~12.0~1libraries~1all-cluster-statuses/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.statuses - clusterstatus: - operation: - $ref: '#/paths/~1api~12.0~1libraries~1cluster-status/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - install: - operation: - $ref: '#/paths/~1api~12.0~1libraries~1install/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - uninstall: - operation: - $ref: '#/paths/~1api~12.0~1libraries~1uninstall/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/libraries/methods/clusterstatus' - - $ref: '#/components/x-stackQL-resources/libraries/methods/allclusterlibrarystatuses' - insert: - - $ref: '#/components/x-stackQL-resources/libraries/methods/install' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/libraries/methods/uninstall' - exec: [] - node_types: - id: databricks_workspace.compute.node_types - name: node_types - title: Node_types - methods: - listnodetypes: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1list-node-types/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.node_types - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/node_types/methods/listnodetypes' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - policy_families: - id: databricks_workspace.compute.policy_families - name: policy_families - title: Policy_families - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1policy-families~1{policy_family_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1policy-families/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.policy_families - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/policy_families/methods/list' - - $ref: '#/components/x-stackQL-resources/policy_families/methods/get' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - policy_permission_levels: - id: databricks_workspace.compute.policy_permission_levels - name: policy_permission_levels - title: Policy_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/policy_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - spark_versions: - id: databricks_workspace.compute.spark_versions - name: spark_versions - title: Spark_versions - methods: - sparkversions: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1spark-versions/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.versions - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/spark_versions/methods/sparkversions' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - zones: - id: databricks_workspace.compute.zones - name: zones - title: Zones - methods: - listzones: - operation: - $ref: '#/paths/~1api~12.1~1clusters~1list-zones/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/zones/methods/listzones' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] +openapi: 3.0.0 +info: + title: Databricks Compute API (workspace) + description: >- + OpenAPI specification for the Databricks compute service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.compute +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/policies/clusters/create: + post: + operationId: cluster_policies_create + summary: Creates a new policy with prescribed settings. + tags: + - compute + - cluster_policies + description: |- + Creates a new policy with prescribed settings. + + :param definition: str (optional) + Policy definition document expressed in [Databricks Cluster Policy Definition Language]. + + [Databricks Cluster Policy Definition Language]: https://docs.databricks.com/administration-guide/clusters/policy-definition.html + :param description: str (optional) + Additional human-readable description of the cluster policy. + :param libraries: List[:class:`Library`] (optional) + A list of libraries to be installed on the next cluster restart that uses this policy. The maximum + number of libraries is 500. + :param max_clusters_per_user: int (optional) + Max number of clusters per user that can be active using this policy. If not present, there is no + max limit. + :param name: str (optional) + Cluster Policy name requested by the user. This has to be unique. Length must be between 1 and 100 + characters. + :param policy_family_definition_overrides: str (optional) + Policy definition JSON document expressed in [Databricks Policy Definition Language]. The JSON + document must be passed as a string and cannot be embedded in the requests. + + You can use this to customize the policy definition inherited from the policy family. Policy rules + specified here are merged into the inherited policy definition. + + [Databricks Policy Definition Language]: https://docs.databricks.com/administration-guide/clusters/policy-definition.html + :param policy_family_id: str (optional) + ID of the policy family. The cluster policy's policy definition inherits the policy family's policy + definition. + + Cannot be used with `definition`. Use `policy_family_definition_overrides` instead to customize the + policy definition. + + :returns: :class:`CreatePolicyResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + definition: + type: string + description: >- + Policy definition document expressed in [Databricks Cluster + Policy Definition Language]. [Databricks Cluster Policy + Definition Language]: + https://docs.databricks.com/administration-guide/clusters/policy-definition.html + description: + type: string + description: Additional human-readable description of the cluster policy. + libraries: + type: string + description: >- + A list of libraries to be installed on the next cluster + restart that uses this policy. The maximum number of + libraries is 500. + max_clusters_per_user: + type: string + description: >- + Max number of clusters per user that can be active using + this policy. If not present, there is no max limit. + name: + type: string + description: >- + Cluster Policy name requested by the user. This has to be + unique. Length must be between 1 and 100 characters. + policy_family_definition_overrides: + type: string + description: >- + Policy definition JSON document expressed in [Databricks + Policy Definition Language]. The JSON document must be + passed as a string and cannot be embedded in the requests. + You can use this to customize the policy definition + inherited from the policy family. Policy rules specified + here are merged into the inherited policy definition. + [Databricks Policy Definition Language]: + https://docs.databricks.com/administration-guide/clusters/policy-definition.html + policy_family_id: + type: string + description: >- + ID of the policy family. The cluster policy's policy + definition inherits the policy family's policy definition. + Cannot be used with `definition`. Use + `policy_family_definition_overrides` instead to customize + the policy definition. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreatePolicyResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + /api/2.0/policies/clusters/delete: + post: + operationId: cluster_policies_delete + summary: >- + Delete a policy for a cluster. Clusters governed by this policy can + still run, but cannot be edited. + tags: + - compute + - cluster_policies + description: >- + Delete a policy for a cluster. Clusters governed by this policy can + still run, but cannot be edited. + + + :param policy_id: str + The ID of the policy to delete. + requestBody: + content: + application/json: + schema: + type: object + properties: + policy_id: + type: string + description: The ID of the policy to delete. + required: + - policy_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + /api/2.0/policies/clusters/edit: + post: + operationId: cluster_policies_edit + summary: >- + Update an existing policy for cluster. This operation may make some + clusters governed by the previous + tags: + - compute + - cluster_policies + description: >- + Update an existing policy for cluster. This operation may make some + clusters governed by the previous + + policy invalid. + + + :param policy_id: str + The ID of the policy to update. + :param definition: str (optional) + Policy definition document expressed in [Databricks Cluster Policy Definition Language]. + + [Databricks Cluster Policy Definition Language]: https://docs.databricks.com/administration-guide/clusters/policy-definition.html + :param description: str (optional) + Additional human-readable description of the cluster policy. + :param libraries: List[:class:`Library`] (optional) + A list of libraries to be installed on the next cluster restart that uses this policy. The maximum + number of libraries is 500. + :param max_clusters_per_user: int (optional) + Max number of clusters per user that can be active using this policy. If not present, there is no + max limit. + :param name: str (optional) + Cluster Policy name requested by the user. This has to be unique. Length must be between 1 and 100 + characters. + :param policy_family_definition_overrides: str (optional) + Policy definition JSON document expressed in [Databricks Policy Definition Language]. The JSON + document must be passed as a string and cannot be embedded in the requests. + + You can use this to customize the policy definition inherited from the policy family. Policy rules + specified here are merged into the inherited policy definition. + + [Databricks Policy Definition Language]: https://docs.databricks.com/administration-guide/clusters/policy-definition.html + :param policy_family_id: str (optional) + ID of the policy family. The cluster policy's policy definition inherits the policy family's policy + definition. + + Cannot be used with `definition`. Use `policy_family_definition_overrides` instead to customize the + policy definition. + requestBody: + content: + application/json: + schema: + type: object + properties: + policy_id: + type: string + description: The ID of the policy to update. + definition: + type: string + description: >- + Policy definition document expressed in [Databricks Cluster + Policy Definition Language]. [Databricks Cluster Policy + Definition Language]: + https://docs.databricks.com/administration-guide/clusters/policy-definition.html + description: + type: string + description: Additional human-readable description of the cluster policy. + libraries: + type: string + description: >- + A list of libraries to be installed on the next cluster + restart that uses this policy. The maximum number of + libraries is 500. + max_clusters_per_user: + type: string + description: >- + Max number of clusters per user that can be active using + this policy. If not present, there is no max limit. + name: + type: string + description: >- + Cluster Policy name requested by the user. This has to be + unique. Length must be between 1 and 100 characters. + policy_family_definition_overrides: + type: string + description: >- + Policy definition JSON document expressed in [Databricks + Policy Definition Language]. The JSON document must be + passed as a string and cannot be embedded in the requests. + You can use this to customize the policy definition + inherited from the policy family. Policy rules specified + here are merged into the inherited policy definition. + [Databricks Policy Definition Language]: + https://docs.databricks.com/administration-guide/clusters/policy-definition.html + policy_family_id: + type: string + description: >- + ID of the policy family. The cluster policy's policy + definition inherits the policy family's policy definition. + Cannot be used with `definition`. Use + `policy_family_definition_overrides` instead to customize + the policy definition. + required: + - policy_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + /api/2.0/policies/clusters/get: + get: + operationId: cluster_policies_get + summary: >- + Get a cluster policy entity. Creation and editing is available to admins + only. + tags: + - compute + - cluster_policies + description: >- + Get a cluster policy entity. Creation and editing is available to admins + only. + + + :param policy_id: str + Canonical unique identifier for the Cluster Policy. + + :returns: :class:`Policy` + parameters: + - name: policy_id + in: query + required: true + schema: + type: string + description: Canonical unique identifier for the Cluster Policy. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Policy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + /api/2.0/permissions/cluster-policies/{cluster_policy_id}/permissionLevels: + get: + operationId: cluster_policies_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - compute + - cluster_policies + description: |- + Gets the permission levels that a user can have on an object. + + :param cluster_policy_id: str + The cluster policy for which to get or manage permissions. + + :returns: :class:`GetClusterPolicyPermissionLevelsResponse` + parameters: + - name: cluster_policy_id + in: path + required: true + schema: + type: string + description: The cluster policy for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetClusterPolicyPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + /api/2.0/permissions/cluster-policies/{cluster_policy_id}: + get: + operationId: cluster_policies_get_permissions + summary: >- + Gets the permissions of a cluster policy. Cluster policies can inherit + permissions from their root + tags: + - compute + - cluster_policies + description: >- + Gets the permissions of a cluster policy. Cluster policies can inherit + permissions from their root + + object. + + + :param cluster_policy_id: str + The cluster policy for which to get or manage permissions. + + :returns: :class:`ClusterPolicyPermissions` + parameters: + - name: cluster_policy_id + in: path + required: true + schema: + type: string + description: The cluster policy for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterPolicyPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + put: + operationId: cluster_policies_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - compute + - cluster_policies + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param cluster_policy_id: str + The cluster policy for which to get or manage permissions. + :param access_control_list: + List[:class:`ClusterPolicyAccessControlRequest`] (optional) + + + :returns: :class:`ClusterPolicyPermissions` + parameters: + - name: cluster_policy_id + in: path + required: true + schema: + type: string + description: The cluster policy for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ClusterPolicyPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterPolicyPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + patch: + operationId: cluster_policies_update_permissions + summary: >- + Updates the permissions on a cluster policy. Cluster policies can + inherit permissions from their root + tags: + - compute + - cluster_policies + description: >- + Updates the permissions on a cluster policy. Cluster policies can + inherit permissions from their root + + object. + + + :param cluster_policy_id: str + The cluster policy for which to get or manage permissions. + :param access_control_list: + List[:class:`ClusterPolicyAccessControlRequest`] (optional) + + + :returns: :class:`ClusterPolicyPermissions` + parameters: + - name: cluster_policy_id + in: path + required: true + schema: + type: string + description: The cluster policy for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ClusterPolicyPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterPolicyPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + /api/2.0/policies/clusters/list: + get: + operationId: cluster_policies_list + summary: Returns a list of policies accessible by the requesting user. + tags: + - compute + - cluster_policies + description: |- + Returns a list of policies accessible by the requesting user. + + :param sort_column: :class:`ListSortColumn` (optional) + The cluster policy attribute to sort by. * `POLICY_CREATION_TIME` - Sort result list by policy + creation time. * `POLICY_NAME` - Sort result list by policy name. + :param sort_order: :class:`ListSortOrder` (optional) + The order in which the policies get listed. * `DESC` - Sort result list in descending order. * `ASC` + - Sort result list in ascending order. + + :returns: Iterator over :class:`Policy` + parameters: + - name: sort_column + in: query + required: false + schema: + type: string + description: >- + The cluster policy attribute to sort by. * `POLICY_CREATION_TIME` - + Sort result list by policy creation time. * `POLICY_NAME` - Sort + result list by policy name. + - name: sort_order + in: query + required: false + schema: + type: string + description: >- + The order in which the policies get listed. * `DESC` - Sort result + list in descending order. * `ASC` - Sort result list in ascending + order. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClusterPoliciesAPI + /api/2.1/clusters/change-owner: + post: + operationId: clusters_change_owner + summary: >- + Change the owner of the cluster. You must be an admin and the cluster + must be terminated to perform + tags: + - compute + - clusters + description: >- + Change the owner of the cluster. You must be an admin and the cluster + must be terminated to perform + + this operation. The service principal application ID can be supplied as + an argument to + + `owner_username`. + + + :param cluster_id: str + + :param owner_username: str + New owner of the cluster_id after this RPC. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: >- + :param owner_username: str New owner of the cluster_id after + this RPC. + owner_username: + type: string + required: + - cluster_id + - owner_username + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/create: + post: + operationId: clusters_create + summary: >- + Creates a new Spark cluster. This method will acquire new instances from + the cloud provider if + tags: + - compute + - clusters + description: >- + Creates a new Spark cluster. This method will acquire new instances from + the cloud provider if + + necessary. This method is asynchronous; the returned ``cluster_id`` can + be used to poll the cluster + + status. When this method returns, the cluster will be in a ``PENDING`` + state. The cluster will be + + usable once it enters a ``RUNNING`` state. Note: Databricks may not be + able to acquire some of the + + requested nodes, due to cloud provider limitations (account limits, spot + price, etc.) or transient + + network issues. + + + If Databricks acquires at least 85% of the requested on-demand nodes, + cluster creation will succeed. + + Otherwise the cluster will terminate with an informative error message. + + + Rather than authoring the cluster's JSON definition from scratch, + Databricks recommends filling out + + the [create compute UI] and then copying the generated JSON definition + from the UI. + + + [create compute UI]: https://docs.databricks.com/compute/configure.html + + + :param spark_version: str + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of available Spark versions can be + retrieved by using the :method:clusters/sparkVersions API call. + :param apply_policy_default_values: bool (optional) + When set to true, fixed and default values from the policy will be used for fields that are omitted. + When set to false, only fixed values from the policy will be applied. + :param autoscale: :class:`AutoScale` (optional) + Parameters needed in order to automatically scale clusters up and down based on load. Note: + autoscaling works best with DB runtime versions 3.0 or later. + :param autotermination_minutes: int (optional) + Automatically terminates the cluster after it is inactive for this time in minutes. If not set, this + cluster will not be automatically terminated. If specified, the threshold must be between 10 and + 10000 minutes. Users can also set this value to 0 to explicitly disable automatic termination. + :param aws_attributes: :class:`AwsAttributes` (optional) + Attributes related to clusters running on Amazon Web Services. If not specified at cluster creation, + a set of default values will be used. + :param azure_attributes: :class:`AzureAttributes` (optional) + Attributes related to clusters running on Microsoft Azure. If not specified at cluster creation, a + set of default values will be used. + :param clone_from: :class:`CloneCluster` (optional) + When specified, this clones libraries from a source cluster during the creation of a new cluster. + :param cluster_log_conf: :class:`ClusterLogConf` (optional) + The configuration for delivering spark logs to a long-term storage destination. Three kinds of + destinations (DBFS, S3 and Unity Catalog volumes) are supported. Only one destination can be + specified for one cluster. If the conf is given, the logs will be delivered to the destination every + `5 mins`. The destination of driver logs is `$destination/$clusterId/driver`, while the destination + of executor logs is `$destination/$clusterId/executor`. + :param cluster_name: str (optional) + Cluster name requested by the user. This doesn't have to be unique. If not specified at creation, + the cluster name will be an empty string. For job clusters, the cluster name is automatically set + based on the job and job run IDs. + :param custom_tags: Dict[str,str] (optional) + Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS + instances and EBS volumes) with these tags in addition to `default_tags`. Notes: + + - Currently, Databricks allows at most 45 custom tags + + - Clusters can only reuse cloud resources if the resources' tags are a subset of the cluster tags + :param data_security_mode: :class:`DataSecurityMode` (optional) + + :param docker_image: :class:`DockerImage` (optional) + Custom docker image BYOC + :param driver_instance_pool_id: str (optional) + The optional ID of the instance pool for the driver of the cluster belongs. The pool cluster uses + the instance pool with id (instance_pool_id) if the driver pool is not assigned. + :param driver_node_type_flexibility: :class:`NodeTypeFlexibility` + (optional) + Flexible node type configuration for the driver node. + :param driver_node_type_id: str (optional) + The node type of the Spark driver. Note that this field is optional; if unset, the driver node type + will be set as the same value as `node_type_id` defined above. + + This field, along with node_type_id, should not be set if virtual_cluster_size is set. If both + driver_node_type_id, node_type_id, and virtual_cluster_size are specified, driver_node_type_id and + node_type_id take precedence. + :param enable_elastic_disk: bool (optional) + Autoscaling Local Storage: when enabled, this cluster will dynamically acquire additional disk space + when its Spark workers are running low on disk space. + :param enable_local_disk_encryption: bool (optional) + Whether to enable LUKS on cluster VMs' local disks + :param gcp_attributes: :class:`GcpAttributes` (optional) + Attributes related to clusters running on Google Cloud Platform. If not specified at cluster + creation, a set of default values will be used. + :param init_scripts: List[:class:`InitScriptInfo`] (optional) + The configuration for storing init scripts. Any number of destinations can be specified. The scripts + are executed sequentially in the order provided. If `cluster_log_conf` is specified, init script + logs are sent to `//init_scripts`. + :param instance_pool_id: str (optional) + The optional ID of the instance pool to which the cluster belongs. + :param is_single_node: bool (optional) + This field can only be used when `kind = CLASSIC_PREVIEW`. + + When set to true, Databricks will automatically set single node related `custom_tags`, `spark_conf`, + and `num_workers` + :param kind: :class:`Kind` (optional) + + :param node_type_id: str (optional) + This field encodes, through a single value, the resources available to each of the Spark nodes in + this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved by using the + :method:clusters/listNodeTypes API call. + :param num_workers: int (optional) + Number of worker nodes that this cluster should have. A cluster has one Spark Driver and + `num_workers` Executors for a total of `num_workers` + 1 Spark nodes. + + Note: When reading the properties of a cluster, this field reflects the desired number of workers + rather than the actual current number of workers. For instance, if a cluster is resized from 5 to 10 + workers, this field will immediately be updated to reflect the target size of 10 workers, whereas + the workers listed in `spark_info` will gradually increase from 5 to 10 as the new nodes are + provisioned. + :param policy_id: str (optional) + The ID of the cluster policy used to create the cluster if applicable. + :param remote_disk_throughput: int (optional) + If set, what the configurable throughput (in Mb/s) for the remote disk is. Currently only supported + for GCP HYPERDISK_BALANCED disks. + :param runtime_engine: :class:`RuntimeEngine` (optional) + Determines the cluster's runtime engine, either standard or Photon. + + This field is not compatible with legacy `spark_version` values that contain `-photon-`. Remove + `-photon-` from the `spark_version` and set `runtime_engine` to `PHOTON`. + + If left unspecified, the runtime engine defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + :param single_user_name: str (optional) + Single user name if data_security_mode is `SINGLE_USER` + :param spark_conf: Dict[str,str] (optional) + An object containing a set of optional, user-specified Spark configuration key-value pairs. Users + can also pass in a string of extra JVM options to the driver and the executors via + `spark.driver.extraJavaOptions` and `spark.executor.extraJavaOptions` respectively. + :param spark_env_vars: Dict[str,str] (optional) + An object containing a set of optional, user-specified environment variable key-value pairs. Please + note that key-value pair of the form (X,Y) will be exported as is (i.e., `export X='Y'`) while + launching the driver and workers. + + In order to specify an additional set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This ensures that all default databricks + managed environmental variables are included as well. + + Example Spark environment variables: `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": + "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": "$SPARK_DAEMON_JAVA_OPTS + -Dspark.shuffle.service.enabled=true"}` + :param ssh_public_keys: List[str] (optional) + SSH public key contents that will be added to each Spark node in this cluster. The corresponding + private keys can be used to login with the user name `ubuntu` on port `2200`. Up to 10 keys can be + specified. + :param total_initial_remote_disk_size: int (optional) + If set, what the total initial volume size (in GB) of the remote disks should be. Currently only + supported for GCP HYPERDISK_BALANCED disks. + :param use_ml_runtime: bool (optional) + This field can only be used when `kind = CLASSIC_PREVIEW`. + + `effective_spark_version` is determined by `spark_version` (DBR release), this field + `use_ml_runtime`, and whether `node_type_id` is gpu node or not. + :param worker_node_type_flexibility: :class:`NodeTypeFlexibility` + (optional) + Flexible node type configuration for worker nodes. + :param workload_type: :class:`WorkloadType` (optional) + + + :returns: + Long-running operation waiter for :class:`ClusterDetails`. + See :method:wait_get_cluster_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + spark_version: + type: string + description: >- + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A + list of available Spark versions can be retrieved by using + the :method:clusters/sparkVersions API call. + apply_policy_default_values: + type: string + description: >- + When set to true, fixed and default values from the policy + will be used for fields that are omitted. When set to false, + only fixed values from the policy will be applied. + autoscale: + type: string + description: >- + Parameters needed in order to automatically scale clusters + up and down based on load. Note: autoscaling works best with + DB runtime versions 3.0 or later. + autotermination_minutes: + type: string + description: >- + Automatically terminates the cluster after it is inactive + for this time in minutes. If not set, this cluster will not + be automatically terminated. If specified, the threshold + must be between 10 and 10000 minutes. Users can also set + this value to 0 to explicitly disable automatic termination. + aws_attributes: + type: string + description: >- + Attributes related to clusters running on Amazon Web + Services. If not specified at cluster creation, a set of + default values will be used. + azure_attributes: + type: string + description: >- + Attributes related to clusters running on Microsoft Azure. + If not specified at cluster creation, a set of default + values will be used. + clone_from: + type: string + description: >- + When specified, this clones libraries from a source cluster + during the creation of a new cluster. + cluster_log_conf: + type: string + description: >- + The configuration for delivering spark logs to a long-term + storage destination. Three kinds of destinations (DBFS, S3 + and Unity Catalog volumes) are supported. Only one + destination can be specified for one cluster. If the conf is + given, the logs will be delivered to the destination every + `5 mins`. The destination of driver logs is + `$destination/$clusterId/driver`, while the destination of + executor logs is `$destination/$clusterId/executor`. + cluster_name: + type: string + description: >- + Cluster name requested by the user. This doesn't have to be + unique. If not specified at creation, the cluster name will + be an empty string. For job clusters, the cluster name is + automatically set based on the job and job run IDs. + custom_tags: + type: string + description: >- + Additional tags for cluster resources. Databricks will tag + all cluster resources (e.g., AWS instances and EBS volumes) + with these tags in addition to `default_tags`. Notes: - + Currently, Databricks allows at most 45 custom tags - + Clusters can only reuse cloud resources if the resources' + tags are a subset of the cluster tags + data_security_mode: + type: string + description: >- + :param docker_image: :class:`DockerImage` (optional) Custom + docker image BYOC + docker_image: + type: string + driver_instance_pool_id: + type: string + description: >- + The optional ID of the instance pool for the driver of the + cluster belongs. The pool cluster uses the instance pool + with id (instance_pool_id) if the driver pool is not + assigned. + driver_node_type_flexibility: + type: string + description: Flexible node type configuration for the driver node. + driver_node_type_id: + type: string + description: >- + The node type of the Spark driver. Note that this field is + optional; if unset, the driver node type will be set as the + same value as `node_type_id` defined above. This field, + along with node_type_id, should not be set if + virtual_cluster_size is set. If both driver_node_type_id, + node_type_id, and virtual_cluster_size are specified, + driver_node_type_id and node_type_id take precedence. + enable_elastic_disk: + type: string + description: >- + Autoscaling Local Storage: when enabled, this cluster will + dynamically acquire additional disk space when its Spark + workers are running low on disk space. + enable_local_disk_encryption: + type: string + description: Whether to enable LUKS on cluster VMs' local disks + gcp_attributes: + type: string + description: >- + Attributes related to clusters running on Google Cloud + Platform. If not specified at cluster creation, a set of + default values will be used. + init_scripts: + type: string + description: >- + The configuration for storing init scripts. Any number of + destinations can be specified. The scripts are executed + sequentially in the order provided. If `cluster_log_conf` is + specified, init script logs are sent to + `//init_scripts`. + instance_pool_id: + type: string + description: >- + The optional ID of the instance pool to which the cluster + belongs. + is_single_node: + type: string + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + When set to true, Databricks will automatically set single + node related `custom_tags`, `spark_conf`, and `num_workers` + kind: + type: string + description: >- + :param node_type_id: str (optional) This field encodes, + through a single value, the resources available to each of + the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be + retrieved by using the :method:clusters/listNodeTypes API + call. + node_type_id: + type: string + num_workers: + type: string + description: >- + Number of worker nodes that this cluster should have. A + cluster has one Spark Driver and `num_workers` Executors for + a total of `num_workers` + 1 Spark nodes. Note: When reading + the properties of a cluster, this field reflects the desired + number of workers rather than the actual current number of + workers. For instance, if a cluster is resized from 5 to 10 + workers, this field will immediately be updated to reflect + the target size of 10 workers, whereas the workers listed in + `spark_info` will gradually increase from 5 to 10 as the new + nodes are provisioned. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + remote_disk_throughput: + type: string + description: >- + If set, what the configurable throughput (in Mb/s) for the + remote disk is. Currently only supported for GCP + HYPERDISK_BALANCED disks. + runtime_engine: + type: string + description: >- + Determines the cluster's runtime engine, either standard or + Photon. This field is not compatible with legacy + `spark_version` values that contain `-photon-`. Remove + `-photon-` from the `spark_version` and set `runtime_engine` + to `PHOTON`. If left unspecified, the runtime engine + defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + single_user_name: + type: string + description: Single user name if data_security_mode is `SINGLE_USER` + spark_conf: + type: string + description: >- + An object containing a set of optional, user-specified Spark + configuration key-value pairs. Users can also pass in a + string of extra JVM options to the driver and the executors + via `spark.driver.extraJavaOptions` and + `spark.executor.extraJavaOptions` respectively. + spark_env_vars: + type: string + description: >- + An object containing a set of optional, user-specified + environment variable key-value pairs. Please note that + key-value pair of the form (X,Y) will be exported as is + (i.e., `export X='Y'`) while launching the driver and + workers. In order to specify an additional set of + `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. + This ensures that all default databricks managed + environmental variables are included as well. Example Spark + environment variables: `{"SPARK_WORKER_MEMORY": "28000m", + "SPARK_LOCAL_DIRS": "/local_disk0"}` or + `{"SPARK_DAEMON_JAVA_OPTS": "$SPARK_DAEMON_JAVA_OPTS + -Dspark.shuffle.service.enabled=true"}` + ssh_public_keys: + type: string + description: >- + SSH public key contents that will be added to each Spark + node in this cluster. The corresponding private keys can be + used to login with the user name `ubuntu` on port `2200`. Up + to 10 keys can be specified. + total_initial_remote_disk_size: + type: string + description: >- + If set, what the total initial volume size (in GB) of the + remote disks should be. Currently only supported for GCP + HYPERDISK_BALANCED disks. + use_ml_runtime: + type: string + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + `effective_spark_version` is determined by `spark_version` + (DBR release), this field `use_ml_runtime`, and whether + `node_type_id` is gpu node or not. + worker_node_type_flexibility: + type: string + description: Flexible node type configuration for worker nodes. + workload_type: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`ClusterDetails`. See + :method:wait_get_cluster_running for more details. + required: + - spark_version + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/delete: + post: + operationId: clusters_delete + summary: >- + Terminates the Spark cluster with the specified ID. The cluster is + removed asynchronously. Once the + tags: + - compute + - clusters + description: >- + Terminates the Spark cluster with the specified ID. The cluster is + removed asynchronously. Once the + + termination has completed, the cluster will be in a `TERMINATED` state. + If the cluster is already in a + + `TERMINATING` or `TERMINATED` state, nothing will happen. + + + :param cluster_id: str + The cluster to be terminated. + + :returns: + Long-running operation waiter for :class:`ClusterDetails`. + See :method:wait_get_cluster_terminated for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: The cluster to be terminated. + required: + - cluster_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/edit: + post: + operationId: clusters_edit + summary: >- + Updates the configuration of a cluster to match the provided attributes + and size. A cluster can be + tags: + - compute + - clusters + description: >- + Updates the configuration of a cluster to match the provided attributes + and size. A cluster can be + + updated if it is in a `RUNNING` or `TERMINATED` state. + + + If a cluster is updated while in a `RUNNING` state, it will be restarted + so that the new attributes + + can take effect. + + + If a cluster is updated while in a `TERMINATED` state, it will remain + `TERMINATED`. The next time it + + is started using the `clusters/start` API, the new attributes will take + effect. Any attempt to update + + a cluster in any other state will be rejected with an `INVALID_STATE` + error code. + + + Clusters created by the Databricks Jobs service cannot be edited. + + + :param cluster_id: str + ID of the cluster + :param spark_version: str + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of available Spark versions can be + retrieved by using the :method:clusters/sparkVersions API call. + :param apply_policy_default_values: bool (optional) + When set to true, fixed and default values from the policy will be used for fields that are omitted. + When set to false, only fixed values from the policy will be applied. + :param autoscale: :class:`AutoScale` (optional) + Parameters needed in order to automatically scale clusters up and down based on load. Note: + autoscaling works best with DB runtime versions 3.0 or later. + :param autotermination_minutes: int (optional) + Automatically terminates the cluster after it is inactive for this time in minutes. If not set, this + cluster will not be automatically terminated. If specified, the threshold must be between 10 and + 10000 minutes. Users can also set this value to 0 to explicitly disable automatic termination. + :param aws_attributes: :class:`AwsAttributes` (optional) + Attributes related to clusters running on Amazon Web Services. If not specified at cluster creation, + a set of default values will be used. + :param azure_attributes: :class:`AzureAttributes` (optional) + Attributes related to clusters running on Microsoft Azure. If not specified at cluster creation, a + set of default values will be used. + :param cluster_log_conf: :class:`ClusterLogConf` (optional) + The configuration for delivering spark logs to a long-term storage destination. Three kinds of + destinations (DBFS, S3 and Unity Catalog volumes) are supported. Only one destination can be + specified for one cluster. If the conf is given, the logs will be delivered to the destination every + `5 mins`. The destination of driver logs is `$destination/$clusterId/driver`, while the destination + of executor logs is `$destination/$clusterId/executor`. + :param cluster_name: str (optional) + Cluster name requested by the user. This doesn't have to be unique. If not specified at creation, + the cluster name will be an empty string. For job clusters, the cluster name is automatically set + based on the job and job run IDs. + :param custom_tags: Dict[str,str] (optional) + Additional tags for cluster resources. Databricks will tag all cluster resources (e.g., AWS + instances and EBS volumes) with these tags in addition to `default_tags`. Notes: + + - Currently, Databricks allows at most 45 custom tags + + - Clusters can only reuse cloud resources if the resources' tags are a subset of the cluster tags + :param data_security_mode: :class:`DataSecurityMode` (optional) + + :param docker_image: :class:`DockerImage` (optional) + Custom docker image BYOC + :param driver_instance_pool_id: str (optional) + The optional ID of the instance pool for the driver of the cluster belongs. The pool cluster uses + the instance pool with id (instance_pool_id) if the driver pool is not assigned. + :param driver_node_type_flexibility: :class:`NodeTypeFlexibility` + (optional) + Flexible node type configuration for the driver node. + :param driver_node_type_id: str (optional) + The node type of the Spark driver. Note that this field is optional; if unset, the driver node type + will be set as the same value as `node_type_id` defined above. + + This field, along with node_type_id, should not be set if virtual_cluster_size is set. If both + driver_node_type_id, node_type_id, and virtual_cluster_size are specified, driver_node_type_id and + node_type_id take precedence. + :param enable_elastic_disk: bool (optional) + Autoscaling Local Storage: when enabled, this cluster will dynamically acquire additional disk space + when its Spark workers are running low on disk space. + :param enable_local_disk_encryption: bool (optional) + Whether to enable LUKS on cluster VMs' local disks + :param gcp_attributes: :class:`GcpAttributes` (optional) + Attributes related to clusters running on Google Cloud Platform. If not specified at cluster + creation, a set of default values will be used. + :param init_scripts: List[:class:`InitScriptInfo`] (optional) + The configuration for storing init scripts. Any number of destinations can be specified. The scripts + are executed sequentially in the order provided. If `cluster_log_conf` is specified, init script + logs are sent to `//init_scripts`. + :param instance_pool_id: str (optional) + The optional ID of the instance pool to which the cluster belongs. + :param is_single_node: bool (optional) + This field can only be used when `kind = CLASSIC_PREVIEW`. + + When set to true, Databricks will automatically set single node related `custom_tags`, `spark_conf`, + and `num_workers` + :param kind: :class:`Kind` (optional) + + :param node_type_id: str (optional) + This field encodes, through a single value, the resources available to each of the Spark nodes in + this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved by using the + :method:clusters/listNodeTypes API call. + :param num_workers: int (optional) + Number of worker nodes that this cluster should have. A cluster has one Spark Driver and + `num_workers` Executors for a total of `num_workers` + 1 Spark nodes. + + Note: When reading the properties of a cluster, this field reflects the desired number of workers + rather than the actual current number of workers. For instance, if a cluster is resized from 5 to 10 + workers, this field will immediately be updated to reflect the target size of 10 workers, whereas + the workers listed in `spark_info` will gradually increase from 5 to 10 as the new nodes are + provisioned. + :param policy_id: str (optional) + The ID of the cluster policy used to create the cluster if applicable. + :param remote_disk_throughput: int (optional) + If set, what the configurable throughput (in Mb/s) for the remote disk is. Currently only supported + for GCP HYPERDISK_BALANCED disks. + :param runtime_engine: :class:`RuntimeEngine` (optional) + Determines the cluster's runtime engine, either standard or Photon. + + This field is not compatible with legacy `spark_version` values that contain `-photon-`. Remove + `-photon-` from the `spark_version` and set `runtime_engine` to `PHOTON`. + + If left unspecified, the runtime engine defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + :param single_user_name: str (optional) + Single user name if data_security_mode is `SINGLE_USER` + :param spark_conf: Dict[str,str] (optional) + An object containing a set of optional, user-specified Spark configuration key-value pairs. Users + can also pass in a string of extra JVM options to the driver and the executors via + `spark.driver.extraJavaOptions` and `spark.executor.extraJavaOptions` respectively. + :param spark_env_vars: Dict[str,str] (optional) + An object containing a set of optional, user-specified environment variable key-value pairs. Please + note that key-value pair of the form (X,Y) will be exported as is (i.e., `export X='Y'`) while + launching the driver and workers. + + In order to specify an additional set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This ensures that all default databricks + managed environmental variables are included as well. + + Example Spark environment variables: `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": + "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": "$SPARK_DAEMON_JAVA_OPTS + -Dspark.shuffle.service.enabled=true"}` + :param ssh_public_keys: List[str] (optional) + SSH public key contents that will be added to each Spark node in this cluster. The corresponding + private keys can be used to login with the user name `ubuntu` on port `2200`. Up to 10 keys can be + specified. + :param total_initial_remote_disk_size: int (optional) + If set, what the total initial volume size (in GB) of the remote disks should be. Currently only + supported for GCP HYPERDISK_BALANCED disks. + :param use_ml_runtime: bool (optional) + This field can only be used when `kind = CLASSIC_PREVIEW`. + + `effective_spark_version` is determined by `spark_version` (DBR release), this field + `use_ml_runtime`, and whether `node_type_id` is gpu node or not. + :param worker_node_type_flexibility: :class:`NodeTypeFlexibility` + (optional) + Flexible node type configuration for worker nodes. + :param workload_type: :class:`WorkloadType` (optional) + + + :returns: + Long-running operation waiter for :class:`ClusterDetails`. + See :method:wait_get_cluster_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: ID of the cluster + spark_version: + type: string + description: >- + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A + list of available Spark versions can be retrieved by using + the :method:clusters/sparkVersions API call. + apply_policy_default_values: + type: string + description: >- + When set to true, fixed and default values from the policy + will be used for fields that are omitted. When set to false, + only fixed values from the policy will be applied. + autoscale: + type: string + description: >- + Parameters needed in order to automatically scale clusters + up and down based on load. Note: autoscaling works best with + DB runtime versions 3.0 or later. + autotermination_minutes: + type: string + description: >- + Automatically terminates the cluster after it is inactive + for this time in minutes. If not set, this cluster will not + be automatically terminated. If specified, the threshold + must be between 10 and 10000 minutes. Users can also set + this value to 0 to explicitly disable automatic termination. + aws_attributes: + type: string + description: >- + Attributes related to clusters running on Amazon Web + Services. If not specified at cluster creation, a set of + default values will be used. + azure_attributes: + type: string + description: >- + Attributes related to clusters running on Microsoft Azure. + If not specified at cluster creation, a set of default + values will be used. + cluster_log_conf: + type: string + description: >- + The configuration for delivering spark logs to a long-term + storage destination. Three kinds of destinations (DBFS, S3 + and Unity Catalog volumes) are supported. Only one + destination can be specified for one cluster. If the conf is + given, the logs will be delivered to the destination every + `5 mins`. The destination of driver logs is + `$destination/$clusterId/driver`, while the destination of + executor logs is `$destination/$clusterId/executor`. + cluster_name: + type: string + description: >- + Cluster name requested by the user. This doesn't have to be + unique. If not specified at creation, the cluster name will + be an empty string. For job clusters, the cluster name is + automatically set based on the job and job run IDs. + custom_tags: + type: string + description: >- + Additional tags for cluster resources. Databricks will tag + all cluster resources (e.g., AWS instances and EBS volumes) + with these tags in addition to `default_tags`. Notes: - + Currently, Databricks allows at most 45 custom tags - + Clusters can only reuse cloud resources if the resources' + tags are a subset of the cluster tags + data_security_mode: + type: string + description: >- + :param docker_image: :class:`DockerImage` (optional) Custom + docker image BYOC + docker_image: + type: string + driver_instance_pool_id: + type: string + description: >- + The optional ID of the instance pool for the driver of the + cluster belongs. The pool cluster uses the instance pool + with id (instance_pool_id) if the driver pool is not + assigned. + driver_node_type_flexibility: + type: string + description: Flexible node type configuration for the driver node. + driver_node_type_id: + type: string + description: >- + The node type of the Spark driver. Note that this field is + optional; if unset, the driver node type will be set as the + same value as `node_type_id` defined above. This field, + along with node_type_id, should not be set if + virtual_cluster_size is set. If both driver_node_type_id, + node_type_id, and virtual_cluster_size are specified, + driver_node_type_id and node_type_id take precedence. + enable_elastic_disk: + type: string + description: >- + Autoscaling Local Storage: when enabled, this cluster will + dynamically acquire additional disk space when its Spark + workers are running low on disk space. + enable_local_disk_encryption: + type: string + description: Whether to enable LUKS on cluster VMs' local disks + gcp_attributes: + type: string + description: >- + Attributes related to clusters running on Google Cloud + Platform. If not specified at cluster creation, a set of + default values will be used. + init_scripts: + type: string + description: >- + The configuration for storing init scripts. Any number of + destinations can be specified. The scripts are executed + sequentially in the order provided. If `cluster_log_conf` is + specified, init script logs are sent to + `//init_scripts`. + instance_pool_id: + type: string + description: >- + The optional ID of the instance pool to which the cluster + belongs. + is_single_node: + type: string + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + When set to true, Databricks will automatically set single + node related `custom_tags`, `spark_conf`, and `num_workers` + kind: + type: string + description: >- + :param node_type_id: str (optional) This field encodes, + through a single value, the resources available to each of + the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be + retrieved by using the :method:clusters/listNodeTypes API + call. + node_type_id: + type: string + num_workers: + type: string + description: >- + Number of worker nodes that this cluster should have. A + cluster has one Spark Driver and `num_workers` Executors for + a total of `num_workers` + 1 Spark nodes. Note: When reading + the properties of a cluster, this field reflects the desired + number of workers rather than the actual current number of + workers. For instance, if a cluster is resized from 5 to 10 + workers, this field will immediately be updated to reflect + the target size of 10 workers, whereas the workers listed in + `spark_info` will gradually increase from 5 to 10 as the new + nodes are provisioned. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + remote_disk_throughput: + type: string + description: >- + If set, what the configurable throughput (in Mb/s) for the + remote disk is. Currently only supported for GCP + HYPERDISK_BALANCED disks. + runtime_engine: + type: string + description: >- + Determines the cluster's runtime engine, either standard or + Photon. This field is not compatible with legacy + `spark_version` values that contain `-photon-`. Remove + `-photon-` from the `spark_version` and set `runtime_engine` + to `PHOTON`. If left unspecified, the runtime engine + defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + single_user_name: + type: string + description: Single user name if data_security_mode is `SINGLE_USER` + spark_conf: + type: string + description: >- + An object containing a set of optional, user-specified Spark + configuration key-value pairs. Users can also pass in a + string of extra JVM options to the driver and the executors + via `spark.driver.extraJavaOptions` and + `spark.executor.extraJavaOptions` respectively. + spark_env_vars: + type: string + description: >- + An object containing a set of optional, user-specified + environment variable key-value pairs. Please note that + key-value pair of the form (X,Y) will be exported as is + (i.e., `export X='Y'`) while launching the driver and + workers. In order to specify an additional set of + `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. + This ensures that all default databricks managed + environmental variables are included as well. Example Spark + environment variables: `{"SPARK_WORKER_MEMORY": "28000m", + "SPARK_LOCAL_DIRS": "/local_disk0"}` or + `{"SPARK_DAEMON_JAVA_OPTS": "$SPARK_DAEMON_JAVA_OPTS + -Dspark.shuffle.service.enabled=true"}` + ssh_public_keys: + type: string + description: >- + SSH public key contents that will be added to each Spark + node in this cluster. The corresponding private keys can be + used to login with the user name `ubuntu` on port `2200`. Up + to 10 keys can be specified. + total_initial_remote_disk_size: + type: string + description: >- + If set, what the total initial volume size (in GB) of the + remote disks should be. Currently only supported for GCP + HYPERDISK_BALANCED disks. + use_ml_runtime: + type: string + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + `effective_spark_version` is determined by `spark_version` + (DBR release), this field `use_ml_runtime`, and whether + `node_type_id` is gpu node or not. + worker_node_type_flexibility: + type: string + description: Flexible node type configuration for worker nodes. + workload_type: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`ClusterDetails`. See + :method:wait_get_cluster_running for more details. + required: + - cluster_id + - spark_version + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/events: + post: + operationId: clusters_events + summary: >- + Retrieves a list of events about the activity of a cluster. This API is + paginated. If there are more + tags: + - compute + - clusters + description: >- + Retrieves a list of events about the activity of a cluster. This API is + paginated. If there are more + + events to read, the response includes all the parameters necessary to + request the next page of events. + + + :param cluster_id: str + The ID of the cluster to retrieve events about. + :param end_time: int (optional) + The end time in epoch milliseconds. If empty, returns events up to the current time. + :param event_types: List[:class:`EventType`] (optional) + An optional set of event types to filter on. If empty, all event types are returned. + :param limit: int (optional) + Deprecated: use page_token in combination with page_size instead. + + The maximum number of events to include in a page of events. Defaults to 50, and maximum allowed + value is 500. + :param offset: int (optional) + Deprecated: use page_token in combination with page_size instead. + + The offset in the result set. Defaults to 0 (no offset). When an offset is specified and the results + are requested in descending order, the end_time field is required. + :param order: :class:`GetEventsOrder` (optional) + The order to list events in; either "ASC" or "DESC". Defaults to "DESC". + :param page_size: int (optional) + The maximum number of events to include in a page of events. The server may further constrain the + maximum number of results returned in a single page. If the page_size is empty or 0, the server will + decide the number of results to be returned. The field has to be in the range [0,500]. If the value + is outside the range, the server enforces 0 or 500. + :param page_token: str (optional) + Use next_page_token or prev_page_token returned from the previous request to list the next or + previous page of events respectively. If page_token is empty, the first page is returned. + :param start_time: int (optional) + The start time in epoch milliseconds. If empty, returns events starting from the beginning of time. + + :returns: Iterator over :class:`ClusterEvent` + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: The ID of the cluster to retrieve events about. + end_time: + type: string + description: >- + The end time in epoch milliseconds. If empty, returns events + up to the current time. + event_types: + type: string + description: >- + An optional set of event types to filter on. If empty, all + event types are returned. + limit: + type: string + description: >- + Deprecated: use page_token in combination with page_size + instead. The maximum number of events to include in a page + of events. Defaults to 50, and maximum allowed value is 500. + offset: + type: string + description: >- + Deprecated: use page_token in combination with page_size + instead. The offset in the result set. Defaults to 0 (no + offset). When an offset is specified and the results are + requested in descending order, the end_time field is + required. + order: + type: string + description: >- + The order to list events in; either "ASC" or "DESC". + Defaults to "DESC". + page_size: + type: string + description: >- + The maximum number of events to include in a page of events. + The server may further constrain the maximum number of + results returned in a single page. If the page_size is empty + or 0, the server will decide the number of results to be + returned. The field has to be in the range [0,500]. If the + value is outside the range, the server enforces 0 or 500. + page_token: + type: string + description: >- + Use next_page_token or prev_page_token returned from the + previous request to list the next or previous page of events + respectively. If page_token is empty, the first page is + returned. + start_time: + type: string + description: >- + The start time in epoch milliseconds. If empty, returns + events starting from the beginning of time. + required: + - cluster_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetEventsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/get: + get: + operationId: clusters_get + summary: >- + Retrieves the information for a cluster given its identifier. Clusters + can be described while they are + tags: + - compute + - clusters + description: >- + Retrieves the information for a cluster given its identifier. Clusters + can be described while they are + + running, or up to 60 days after they are terminated. + + + :param cluster_id: str + The cluster about which to retrieve information. + + :returns: :class:`ClusterDetails` + parameters: + - name: cluster_id + in: query + required: true + schema: + type: string + description: The cluster about which to retrieve information. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.0/permissions/clusters/{cluster_id}/permissionLevels: + get: + operationId: clusters_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - compute + - clusters + description: |- + Gets the permission levels that a user can have on an object. + + :param cluster_id: str + The cluster for which to get or manage permissions. + + :returns: :class:`GetClusterPermissionLevelsResponse` + parameters: + - name: cluster_id + in: path + required: true + schema: + type: string + description: The cluster for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetClusterPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.0/permissions/clusters/{cluster_id}: + get: + operationId: clusters_get_permissions + summary: >- + Gets the permissions of a cluster. Clusters can inherit permissions from + their root object. + tags: + - compute + - clusters + description: >- + Gets the permissions of a cluster. Clusters can inherit permissions from + their root object. + + + :param cluster_id: str + The cluster for which to get or manage permissions. + + :returns: :class:`ClusterPermissions` + parameters: + - name: cluster_id + in: path + required: true + schema: + type: string + description: The cluster for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + put: + operationId: clusters_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - compute + - clusters + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param cluster_id: str + The cluster for which to get or manage permissions. + :param access_control_list: List[:class:`ClusterAccessControlRequest`] + (optional) + + + :returns: :class:`ClusterPermissions` + parameters: + - name: cluster_id + in: path + required: true + schema: + type: string + description: The cluster for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ClusterPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + patch: + operationId: clusters_update_permissions + summary: >- + Updates the permissions on a cluster. Clusters can inherit permissions + from their root object. + tags: + - compute + - clusters + description: >- + Updates the permissions on a cluster. Clusters can inherit permissions + from their root object. + + + :param cluster_id: str + The cluster for which to get or manage permissions. + :param access_control_list: List[:class:`ClusterAccessControlRequest`] + (optional) + + + :returns: :class:`ClusterPermissions` + parameters: + - name: cluster_id + in: path + required: true + schema: + type: string + description: The cluster for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ClusterPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/list: + get: + operationId: clusters_list + summary: >- + Return information about all pinned and active clusters, and all + clusters terminated within the last + tags: + - compute + - clusters + description: >- + Return information about all pinned and active clusters, and all + clusters terminated within the last + + 30 days. Clusters terminated prior to this period are not included. + + + :param filter_by: :class:`ListClustersFilterBy` (optional) + Filters to apply to the list of clusters. + :param page_size: int (optional) + Use this field to specify the maximum number of results to be returned by the server. The server may + further constrain the maximum number of results returned in a single page. + :param page_token: str (optional) + Use next_page_token or prev_page_token returned from the previous request to list the next or + previous page of clusters respectively. + :param sort_by: :class:`ListClustersSortBy` (optional) + Sort the list of clusters by a specific criteria. + + :returns: Iterator over :class:`ClusterDetails` + parameters: + - name: filter_by + in: query + required: false + schema: + type: string + description: Filters to apply to the list of clusters. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + Use this field to specify the maximum number of results to be + returned by the server. The server may further constrain the maximum + number of results returned in a single page. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Use next_page_token or prev_page_token returned from the previous + request to list the next or previous page of clusters respectively. + - name: sort_by + in: query + required: false + schema: + type: string + description: Sort the list of clusters by a specific criteria. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListClustersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/list-node-types: + get: + operationId: clusters_list_node_types + summary: >- + Returns a list of supported Spark node types. These node types can be + used to launch a cluster. + tags: + - compute + - clusters + description: >- + Returns a list of supported Spark node types. These node types can be + used to launch a cluster. + + + + :returns: :class:`ListNodeTypesResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListNodeTypesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/list-zones: + get: + operationId: clusters_list_zones + summary: >- + Returns a list of availability zones where clusters can be created in + (For example, us-west-2a). These + tags: + - compute + - clusters + description: >- + Returns a list of availability zones where clusters can be created in + (For example, us-west-2a). These + + zones can be used to launch a cluster. + + + + :returns: :class:`ListAvailableZonesResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAvailableZonesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/permanent-delete: + post: + operationId: clusters_permanent_delete + summary: >- + Permanently deletes a Spark cluster. This cluster is terminated and + resources are asynchronously + tags: + - compute + - clusters + description: >- + Permanently deletes a Spark cluster. This cluster is terminated and + resources are asynchronously + + removed. + + + In addition, users will no longer see permanently deleted clusters in + the cluster list, and API users + + can no longer perform any action on permanently deleted clusters. + + + :param cluster_id: str + The cluster to be deleted. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: The cluster to be deleted. + required: + - cluster_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/pin: + post: + operationId: clusters_pin + summary: >- + Pinning a cluster ensures that the cluster will always be returned by + the ListClusters API. Pinning a + tags: + - compute + - clusters + description: >- + Pinning a cluster ensures that the cluster will always be returned by + the ListClusters API. Pinning a + + cluster that is already pinned will have no effect. This API can only be + called by workspace admins. + + + :param cluster_id: str + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: str + required: + - cluster_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/resize: + post: + operationId: clusters_resize + summary: >- + Resizes a cluster to have a desired number of workers. This will fail + unless the cluster is in a + tags: + - compute + - clusters + description: >- + Resizes a cluster to have a desired number of workers. This will fail + unless the cluster is in a + + `RUNNING` state. + + + :param cluster_id: str + The cluster to be resized. + :param autoscale: :class:`AutoScale` (optional) + Parameters needed in order to automatically scale clusters up and down based on load. Note: + autoscaling works best with DB runtime versions 3.0 or later. + :param num_workers: int (optional) + Number of worker nodes that this cluster should have. A cluster has one Spark Driver and + `num_workers` Executors for a total of `num_workers` + 1 Spark nodes. + + Note: When reading the properties of a cluster, this field reflects the desired number of workers + rather than the actual current number of workers. For instance, if a cluster is resized from 5 to 10 + workers, this field will immediately be updated to reflect the target size of 10 workers, whereas + the workers listed in `spark_info` will gradually increase from 5 to 10 as the new nodes are + provisioned. + + :returns: + Long-running operation waiter for :class:`ClusterDetails`. + See :method:wait_get_cluster_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: The cluster to be resized. + autoscale: + type: string + description: >- + Parameters needed in order to automatically scale clusters + up and down based on load. Note: autoscaling works best with + DB runtime versions 3.0 or later. + num_workers: + type: string + description: >- + Number of worker nodes that this cluster should have. A + cluster has one Spark Driver and `num_workers` Executors for + a total of `num_workers` + 1 Spark nodes. Note: When reading + the properties of a cluster, this field reflects the desired + number of workers rather than the actual current number of + workers. For instance, if a cluster is resized from 5 to 10 + workers, this field will immediately be updated to reflect + the target size of 10 workers, whereas the workers listed in + `spark_info` will gradually increase from 5 to 10 as the new + nodes are provisioned. + required: + - cluster_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/restart: + post: + operationId: clusters_restart + summary: >- + Restarts a Spark cluster with the supplied ID. If the cluster is not + currently in a `RUNNING` state, + tags: + - compute + - clusters + description: >- + Restarts a Spark cluster with the supplied ID. If the cluster is not + currently in a `RUNNING` state, + + nothing will happen. + + + :param cluster_id: str + The cluster to be started. + :param restart_user: str (optional) + + + :returns: + Long-running operation waiter for :class:`ClusterDetails`. + See :method:wait_get_cluster_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: The cluster to be started. + restart_user: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`ClusterDetails`. See + :method:wait_get_cluster_running for more details. + required: + - cluster_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/spark-versions: + get: + operationId: clusters_spark_versions + summary: >- + Returns the list of available Spark versions. These versions can be used + to launch a cluster. + tags: + - compute + - clusters + description: >- + Returns the list of available Spark versions. These versions can be used + to launch a cluster. + + + + :returns: :class:`GetSparkVersionsResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetSparkVersionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/start: + post: + operationId: clusters_start + summary: >- + Starts a terminated Spark cluster with the supplied ID. This works + similar to `createCluster` except: + tags: + - compute + - clusters + description: >- + Starts a terminated Spark cluster with the supplied ID. This works + similar to `createCluster` except: + + - The previous cluster id and attributes are preserved. - The cluster + starts with the last specified + + cluster size. - If the previous cluster was an autoscaling cluster, the + current cluster starts with + + the minimum number of nodes. - If the cluster is not currently in a + ``TERMINATED`` state, nothing will + + happen. - Clusters launched to run a job cannot be started. + + + :param cluster_id: str + The cluster to be started. + + :returns: + Long-running operation waiter for :class:`ClusterDetails`. + See :method:wait_get_cluster_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: The cluster to be started. + required: + - cluster_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/unpin: + post: + operationId: clusters_unpin + summary: >- + Unpinning a cluster will allow the cluster to eventually be removed from + the ListClusters API. + tags: + - compute + - clusters + description: >- + Unpinning a cluster will allow the cluster to eventually be removed from + the ListClusters API. + + Unpinning a cluster that is not pinned will have no effect. This API can + only be called by workspace + + admins. + + + :param cluster_id: str + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: str + required: + - cluster_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/2.1/clusters/update: + post: + operationId: clusters_update + summary: >- + Updates the configuration of a cluster to match the partial set of + attributes and size. Denote which + tags: + - compute + - clusters + description: >- + Updates the configuration of a cluster to match the partial set of + attributes and size. Denote which + + fields to update using the `update_mask` field in the request body. A + cluster can be updated if it is + + in a `RUNNING` or `TERMINATED` state. If a cluster is updated while in a + `RUNNING` state, it will be + + restarted so that the new attributes can take effect. If a cluster is + updated while in a `TERMINATED` + + state, it will remain `TERMINATED`. The updated attributes will take + effect the next time the cluster + + is started using the `clusters/start` API. Attempts to update a cluster + in any other state will be + + rejected with an `INVALID_STATE` error code. Clusters created by the + Databricks Jobs service cannot be + + updated. + + + :param cluster_id: str + ID of the cluster. + :param update_mask: str + Used to specify which cluster attributes and size fields to update. See https://google.aip.dev/161 + for more details. + + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + :param cluster: :class:`UpdateClusterResource` (optional) + The cluster to be updated. + + :returns: + Long-running operation waiter for :class:`ClusterDetails`. + See :method:wait_get_cluster_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: ID of the cluster. + update_mask: + type: string + description: >- + Used to specify which cluster attributes and size fields to + update. See https://google.aip.dev/161 for more details. The + field mask must be a single string, with multiple fields + separated by commas (no spaces). The field path is relative + to the resource object, using a dot (`.`) to navigate + sub-fields (e.g., `author.given_name`). Specification of + elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names + must exactly match the resource field names. A field mask of + `*` indicates full replacement. It’s recommended to always + explicitly list the fields being updated and avoid using `*` + wildcards, as it can lead to unintended results if the API + changes in the future. + cluster: + type: string + description: The cluster to be updated. + required: + - cluster_id + - update_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterDetails' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ClustersAPI + /api/1.2/commands/cancel: + post: + operationId: command_execution_cancel + summary: Cancels a currently running command within an execution context. + tags: + - compute + - command_execution + description: |- + Cancels a currently running command within an execution context. + + The command ID is obtained from a prior successful call to __execute__. + + :param cluster_id: str (optional) + :param command_id: str (optional) + :param context_id: str (optional) + + :returns: + Long-running operation waiter for :class:`CommandStatusResponse`. + See :method:wait_command_status_command_execution_cancelled for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: ':param command_id: str (optional)' + command_id: + type: string + context_id: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`CommandStatusResponse`. See + :method:wait_command_status_command_execution_cancelled for + more details. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CommandStatusResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CommandExecutionAPI + /api/1.2/commands/status: + get: + operationId: command_execution_command_status + summary: >- + Gets the status of and, if available, the results from a currently + executing command. + tags: + - compute + - command_execution + description: >- + Gets the status of and, if available, the results from a currently + executing command. + + + The command ID is obtained from a prior successful call to __execute__. + + + :param cluster_id: str + + :param context_id: str + + :param command_id: str + + + :returns: :class:`CommandStatusResponse` + parameters: + - name: cluster_id + in: query + required: true + schema: + type: string + description: ':param context_id: str' + - name: context_id + in: query + required: true + schema: + type: string + - name: command_id + in: query + required: true + schema: + type: string + description: ':returns: :class:`CommandStatusResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CommandStatusResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CommandExecutionAPI + /api/1.2/contexts/status: + get: + operationId: command_execution_context_status + summary: Gets the status for an execution context. + tags: + - compute + - command_execution + description: |- + Gets the status for an execution context. + + :param cluster_id: str + :param context_id: str + + :returns: :class:`ContextStatusResponse` + parameters: + - name: cluster_id + in: query + required: true + schema: + type: string + description: ':param context_id: str' + - name: context_id + in: query + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ContextStatusResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CommandExecutionAPI + /api/1.2/contexts/create: + post: + operationId: command_execution_create + summary: Creates an execution context for running cluster commands. + tags: + - compute + - command_execution + description: |- + Creates an execution context for running cluster commands. + + If successful, this method returns the ID of the new execution context. + + :param cluster_id: str (optional) + Running cluster id + :param language: :class:`Language` (optional) + + :returns: + Long-running operation waiter for :class:`ContextStatusResponse`. + See :method:wait_context_status_command_execution_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: Running cluster id + language: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`ContextStatusResponse`. See + :method:wait_context_status_command_execution_running for + more details. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ContextStatusResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CommandExecutionAPI + /api/1.2/contexts/destroy: + post: + operationId: command_execution_destroy + summary: Deletes an execution context. + tags: + - compute + - command_execution + description: |- + Deletes an execution context. + + :param cluster_id: str + :param context_id: str + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: ':param context_id: str' + context_id: + type: string + required: + - cluster_id + - context_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CommandExecutionAPI + /api/1.2/commands/execute: + post: + operationId: command_execution_execute + summary: >- + Runs a cluster command in the given execution context, using the + provided language. + tags: + - compute + - command_execution + description: >- + Runs a cluster command in the given execution context, using the + provided language. + + + If successful, it returns an ID for tracking the status of the command's + execution. + + + :param cluster_id: str (optional) + Running cluster id + :param command: str (optional) + Executable code + :param context_id: str (optional) + Running context id + :param language: :class:`Language` (optional) + + + :returns: + Long-running operation waiter for :class:`CommandStatusResponse`. + See :method:wait_command_status_command_execution_finished_or_error for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: Running cluster id + command: + type: string + description: Executable code + context_id: + type: string + description: Running context id + language: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`CommandStatusResponse`. See + :method:wait_command_status_command_execution_finished_or_error + for more details. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CommandStatusResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CommandExecutionAPI + /api/2.0/global-init-scripts: + post: + operationId: global_init_scripts_create + summary: Creates a new global init script in this workspace. + tags: + - compute + - global_init_scripts + description: |- + Creates a new global init script in this workspace. + + :param name: str + The name of the script + :param script: str + The Base64-encoded content of the script. + :param enabled: bool (optional) + Specifies whether the script is enabled. The script runs only if enabled. + :param position: int (optional) + The position of a global init script, where 0 represents the first script to run, 1 is the second + script to run, in ascending order. + + If you omit the numeric position for a new global init script, it defaults to last position. It will + run after all current scripts. Setting any value greater than the position of the last script is + equivalent to the last position. Example: Take three existing scripts with positions 0, 1, and 2. + Any position of (3) or greater puts the script in the last position. If an explicit position value + conflicts with an existing script value, your request succeeds, but the original script at that + position and all later scripts have their positions incremented by 1. + + :returns: :class:`CreateResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: The name of the script + script: + type: string + description: The Base64-encoded content of the script. + enabled: + type: string + description: >- + Specifies whether the script is enabled. The script runs + only if enabled. + position: + type: string + description: >- + The position of a global init script, where 0 represents the + first script to run, 1 is the second script to run, in + ascending order. If you omit the numeric position for a new + global init script, it defaults to last position. It will + run after all current scripts. Setting any value greater + than the position of the last script is equivalent to the + last position. Example: Take three existing scripts with + positions 0, 1, and 2. Any position of (3) or greater puts + the script in the last position. If an explicit position + value conflicts with an existing script value, your request + succeeds, but the original script at that position and all + later scripts have their positions incremented by 1. + required: + - name + - script + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GlobalInitScriptsAPI + get: + operationId: global_init_scripts_list + summary: >- + Get a list of all global init scripts for this workspace. This returns + all properties for each script + tags: + - compute + - global_init_scripts + description: >- + Get a list of all global init scripts for this workspace. This returns + all properties for each script + + but **not** the script contents. To retrieve the contents of a script, + use the [get a global init + + script](:method:globalinitscripts/get) operation. + + + + :returns: Iterator over :class:`GlobalInitScriptDetails` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListGlobalInitScriptsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GlobalInitScriptsAPI + /api/2.0/global-init-scripts/{script_id}: + delete: + operationId: global_init_scripts_delete + summary: Deletes a global init script. + tags: + - compute + - global_init_scripts + description: |- + Deletes a global init script. + + :param script_id: str + The ID of the global init script. + parameters: + - name: script_id + in: path + required: true + schema: + type: string + description: The ID of the global init script. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GlobalInitScriptsAPI + get: + operationId: global_init_scripts_get + summary: Gets all the details of a script, including its Base64-encoded contents. + tags: + - compute + - global_init_scripts + description: |- + Gets all the details of a script, including its Base64-encoded contents. + + :param script_id: str + The ID of the global init script. + + :returns: :class:`GlobalInitScriptDetailsWithContent` + parameters: + - name: script_id + in: path + required: true + schema: + type: string + description: The ID of the global init script. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GlobalInitScriptDetailsWithContent' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GlobalInitScriptsAPI + patch: + operationId: global_init_scripts_update + summary: >- + Updates a global init script, specifying only the fields to change. All + fields are optional. + tags: + - compute + - global_init_scripts + description: >- + Updates a global init script, specifying only the fields to change. All + fields are optional. + + Unspecified fields retain their current value. + + + :param script_id: str + The ID of the global init script. + :param name: str + The name of the script + :param script: str + The Base64-encoded content of the script. + :param enabled: bool (optional) + Specifies whether the script is enabled. The script runs only if enabled. + :param position: int (optional) + The position of a script, where 0 represents the first script to run, 1 is the second script to run, + in ascending order. To move the script to run first, set its position to 0. + + To move the script to the end, set its position to any value greater or equal to the position of the + last script. Example, three existing scripts with positions 0, 1, and 2. Any position value of 2 or + greater puts the script in the last position (2). + + If an explicit position value conflicts with an existing script, your request succeeds, but the + original script at that position and all later scripts have their positions incremented by 1. + parameters: + - name: script_id + in: path + required: true + schema: + type: string + description: The ID of the global init script. + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: The name of the script + script: + type: string + description: The Base64-encoded content of the script. + enabled: + type: string + description: >- + Specifies whether the script is enabled. The script runs + only if enabled. + position: + type: string + description: >- + The position of a script, where 0 represents the first + script to run, 1 is the second script to run, in ascending + order. To move the script to run first, set its position to + 0. To move the script to the end, set its position to any + value greater or equal to the position of the last script. + Example, three existing scripts with positions 0, 1, and 2. + Any position value of 2 or greater puts the script in the + last position (2). If an explicit position value conflicts + with an existing script, your request succeeds, but the + original script at that position and all later scripts have + their positions incremented by 1. + required: + - name + - script + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GlobalInitScriptsAPI + /api/2.0/instance-pools/create: + post: + operationId: instance_pools_create + summary: Creates a new instance pool using idle and ready-to-use cloud instances. + tags: + - compute + - instance_pools + description: |- + Creates a new instance pool using idle and ready-to-use cloud instances. + + :param instance_pool_name: str + Pool name requested by the user. Pool name must be unique. Length must be between 1 and 100 + characters. + :param node_type_id: str + This field encodes, through a single value, the resources available to each of the Spark nodes in + this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved by using the + :method:clusters/listNodeTypes API call. + :param aws_attributes: :class:`InstancePoolAwsAttributes` (optional) + Attributes related to instance pools running on Amazon Web Services. If not specified at pool + creation, a set of default values will be used. + :param azure_attributes: :class:`InstancePoolAzureAttributes` (optional) + Attributes related to instance pools running on Azure. If not specified at pool creation, a set of + default values will be used. + :param custom_tags: Dict[str,str] (optional) + Additional tags for pool resources. Databricks will tag all pool resources (e.g., AWS instances and + EBS volumes) with these tags in addition to `default_tags`. Notes: + + - Currently, Databricks allows at most 45 custom tags + :param disk_spec: :class:`DiskSpec` (optional) + Defines the specification of the disks that will be attached to all spark containers. + :param enable_elastic_disk: bool (optional) + Autoscaling Local Storage: when enabled, this instances in this pool will dynamically acquire + additional disk space when its Spark workers are running low on disk space. In AWS, this feature + requires specific AWS permissions to function correctly - refer to the User Guide for more details. + :param gcp_attributes: :class:`InstancePoolGcpAttributes` (optional) + Attributes related to instance pools running on Google Cloud Platform. If not specified at pool + creation, a set of default values will be used. + :param idle_instance_autotermination_minutes: int (optional) + Automatically terminates the extra instances in the pool cache after they are inactive for this time + in minutes if min_idle_instances requirement is already met. If not set, the extra pool instances + will be automatically terminated after a default timeout. If specified, the threshold must be + between 0 and 10000 minutes. Users can also set this value to 0 to instantly remove idle instances + from the cache if min cache size could still hold. + :param max_capacity: int (optional) + Maximum number of outstanding instances to keep in the pool, including both instances used by + clusters and idle instances. Clusters that require further instance provisioning will fail during + upsize requests. + :param min_idle_instances: int (optional) + Minimum number of idle instances to keep in the instance pool + :param node_type_flexibility: :class:`NodeTypeFlexibility` (optional) + Flexible node type configuration for the pool. + :param preloaded_docker_images: List[:class:`DockerImage`] (optional) + Custom Docker Image BYOC + :param preloaded_spark_versions: List[str] (optional) + A list containing at most one preloaded Spark image version for the pool. Pool-backed clusters + started with the preloaded Spark version will start faster. A list of available Spark versions can + be retrieved by using the :method:clusters/sparkVersions API call. + :param remote_disk_throughput: int (optional) + If set, what the configurable throughput (in Mb/s) for the remote disk is. Currently only supported + for GCP HYPERDISK_BALANCED types. + :param total_initial_remote_disk_size: int (optional) + If set, what the total initial volume size (in GB) of the remote disks should be. Currently only + supported for GCP HYPERDISK_BALANCED types. + + :returns: :class:`CreateInstancePoolResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + instance_pool_name: + type: string + description: >- + Pool name requested by the user. Pool name must be unique. + Length must be between 1 and 100 characters. + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources + available to each of the Spark nodes in this cluster. For + example, the Spark nodes can be provisioned and optimized + for memory or compute intensive workloads. A list of + available node types can be retrieved by using the + :method:clusters/listNodeTypes API call. + aws_attributes: + type: string + description: >- + Attributes related to instance pools running on Amazon Web + Services. If not specified at pool creation, a set of + default values will be used. + azure_attributes: + type: string + description: >- + Attributes related to instance pools running on Azure. If + not specified at pool creation, a set of default values will + be used. + custom_tags: + type: string + description: >- + Additional tags for pool resources. Databricks will tag all + pool resources (e.g., AWS instances and EBS volumes) with + these tags in addition to `default_tags`. Notes: - + Currently, Databricks allows at most 45 custom tags + disk_spec: + type: string + description: >- + Defines the specification of the disks that will be attached + to all spark containers. + enable_elastic_disk: + type: string + description: >- + Autoscaling Local Storage: when enabled, this instances in + this pool will dynamically acquire additional disk space + when its Spark workers are running low on disk space. In + AWS, this feature requires specific AWS permissions to + function correctly - refer to the User Guide for more + details. + gcp_attributes: + type: string + description: >- + Attributes related to instance pools running on Google Cloud + Platform. If not specified at pool creation, a set of + default values will be used. + idle_instance_autotermination_minutes: + type: string + description: >- + Automatically terminates the extra instances in the pool + cache after they are inactive for this time in minutes if + min_idle_instances requirement is already met. If not set, + the extra pool instances will be automatically terminated + after a default timeout. If specified, the threshold must be + between 0 and 10000 minutes. Users can also set this value + to 0 to instantly remove idle instances from the cache if + min cache size could still hold. + max_capacity: + type: string + description: >- + Maximum number of outstanding instances to keep in the pool, + including both instances used by clusters and idle + instances. Clusters that require further instance + provisioning will fail during upsize requests. + min_idle_instances: + type: string + description: >- + Minimum number of idle instances to keep in the instance + pool + node_type_flexibility: + type: string + description: Flexible node type configuration for the pool. + preloaded_docker_images: + type: string + description: Custom Docker Image BYOC + preloaded_spark_versions: + type: string + description: >- + A list containing at most one preloaded Spark image version + for the pool. Pool-backed clusters started with the + preloaded Spark version will start faster. A list of + available Spark versions can be retrieved by using the + :method:clusters/sparkVersions API call. + remote_disk_throughput: + type: string + description: >- + If set, what the configurable throughput (in Mb/s) for the + remote disk is. Currently only supported for GCP + HYPERDISK_BALANCED types. + total_initial_remote_disk_size: + type: string + description: >- + If set, what the total initial volume size (in GB) of the + remote disks should be. Currently only supported for GCP + HYPERDISK_BALANCED types. + required: + - instance_pool_name + - node_type_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateInstancePoolResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + /api/2.0/instance-pools/delete: + post: + operationId: instance_pools_delete + summary: >- + Deletes the instance pool permanently. The idle instances in the pool + are terminated asynchronously. + tags: + - compute + - instance_pools + description: >- + Deletes the instance pool permanently. The idle instances in the pool + are terminated asynchronously. + + + :param instance_pool_id: str + The instance pool to be terminated. + requestBody: + content: + application/json: + schema: + type: object + properties: + instance_pool_id: + type: string + description: The instance pool to be terminated. + required: + - instance_pool_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + /api/2.0/instance-pools/edit: + post: + operationId: instance_pools_edit + summary: Modifies the configuration of an existing instance pool. + tags: + - compute + - instance_pools + description: |- + Modifies the configuration of an existing instance pool. + + :param instance_pool_id: str + Instance pool ID + :param instance_pool_name: str + Pool name requested by the user. Pool name must be unique. Length must be between 1 and 100 + characters. + :param node_type_id: str + This field encodes, through a single value, the resources available to each of the Spark nodes in + this cluster. For example, the Spark nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved by using the + :method:clusters/listNodeTypes API call. + :param custom_tags: Dict[str,str] (optional) + Additional tags for pool resources. Databricks will tag all pool resources (e.g., AWS instances and + EBS volumes) with these tags in addition to `default_tags`. Notes: + + - Currently, Databricks allows at most 45 custom tags + :param idle_instance_autotermination_minutes: int (optional) + Automatically terminates the extra instances in the pool cache after they are inactive for this time + in minutes if min_idle_instances requirement is already met. If not set, the extra pool instances + will be automatically terminated after a default timeout. If specified, the threshold must be + between 0 and 10000 minutes. Users can also set this value to 0 to instantly remove idle instances + from the cache if min cache size could still hold. + :param max_capacity: int (optional) + Maximum number of outstanding instances to keep in the pool, including both instances used by + clusters and idle instances. Clusters that require further instance provisioning will fail during + upsize requests. + :param min_idle_instances: int (optional) + Minimum number of idle instances to keep in the instance pool + :param node_type_flexibility: :class:`NodeTypeFlexibility` (optional) + Flexible node type configuration for the pool. + :param remote_disk_throughput: int (optional) + If set, what the configurable throughput (in Mb/s) for the remote disk is. Currently only supported + for GCP HYPERDISK_BALANCED types. + :param total_initial_remote_disk_size: int (optional) + If set, what the total initial volume size (in GB) of the remote disks should be. Currently only + supported for GCP HYPERDISK_BALANCED types. + requestBody: + content: + application/json: + schema: + type: object + properties: + instance_pool_id: + type: string + description: Instance pool ID + instance_pool_name: + type: string + description: >- + Pool name requested by the user. Pool name must be unique. + Length must be between 1 and 100 characters. + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources + available to each of the Spark nodes in this cluster. For + example, the Spark nodes can be provisioned and optimized + for memory or compute intensive workloads. A list of + available node types can be retrieved by using the + :method:clusters/listNodeTypes API call. + custom_tags: + type: string + description: >- + Additional tags for pool resources. Databricks will tag all + pool resources (e.g., AWS instances and EBS volumes) with + these tags in addition to `default_tags`. Notes: - + Currently, Databricks allows at most 45 custom tags + idle_instance_autotermination_minutes: + type: string + description: >- + Automatically terminates the extra instances in the pool + cache after they are inactive for this time in minutes if + min_idle_instances requirement is already met. If not set, + the extra pool instances will be automatically terminated + after a default timeout. If specified, the threshold must be + between 0 and 10000 minutes. Users can also set this value + to 0 to instantly remove idle instances from the cache if + min cache size could still hold. + max_capacity: + type: string + description: >- + Maximum number of outstanding instances to keep in the pool, + including both instances used by clusters and idle + instances. Clusters that require further instance + provisioning will fail during upsize requests. + min_idle_instances: + type: string + description: >- + Minimum number of idle instances to keep in the instance + pool + node_type_flexibility: + type: string + description: Flexible node type configuration for the pool. + remote_disk_throughput: + type: string + description: >- + If set, what the configurable throughput (in Mb/s) for the + remote disk is. Currently only supported for GCP + HYPERDISK_BALANCED types. + total_initial_remote_disk_size: + type: string + description: >- + If set, what the total initial volume size (in GB) of the + remote disks should be. Currently only supported for GCP + HYPERDISK_BALANCED types. + required: + - instance_pool_id + - instance_pool_name + - node_type_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + /api/2.0/instance-pools/get: + get: + operationId: instance_pools_get + summary: Retrieve the information for an instance pool based on its identifier. + tags: + - compute + - instance_pools + description: |- + Retrieve the information for an instance pool based on its identifier. + + :param instance_pool_id: str + The canonical unique identifier for the instance pool. + + :returns: :class:`GetInstancePool` + parameters: + - name: instance_pool_id + in: query + required: true + schema: + type: string + description: The canonical unique identifier for the instance pool. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetInstancePool' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + /api/2.0/permissions/instance-pools/{instance_pool_id}/permissionLevels: + get: + operationId: instance_pools_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - compute + - instance_pools + description: |- + Gets the permission levels that a user can have on an object. + + :param instance_pool_id: str + The instance pool for which to get or manage permissions. + + :returns: :class:`GetInstancePoolPermissionLevelsResponse` + parameters: + - name: instance_pool_id + in: path + required: true + schema: + type: string + description: The instance pool for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetInstancePoolPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + /api/2.0/permissions/instance-pools/{instance_pool_id}: + get: + operationId: instance_pools_get_permissions + summary: >- + Gets the permissions of an instance pool. Instance pools can inherit + permissions from their root + tags: + - compute + - instance_pools + description: >- + Gets the permissions of an instance pool. Instance pools can inherit + permissions from their root + + object. + + + :param instance_pool_id: str + The instance pool for which to get or manage permissions. + + :returns: :class:`InstancePoolPermissions` + parameters: + - name: instance_pool_id + in: path + required: true + schema: + type: string + description: The instance pool for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/InstancePoolPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + put: + operationId: instance_pools_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - compute + - instance_pools + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param instance_pool_id: str + The instance pool for which to get or manage permissions. + :param access_control_list: + List[:class:`InstancePoolAccessControlRequest`] (optional) + + + :returns: :class:`InstancePoolPermissions` + parameters: + - name: instance_pool_id + in: path + required: true + schema: + type: string + description: The instance pool for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`InstancePoolPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/InstancePoolPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + patch: + operationId: instance_pools_update_permissions + summary: >- + Updates the permissions on an instance pool. Instance pools can inherit + permissions from their root + tags: + - compute + - instance_pools + description: >- + Updates the permissions on an instance pool. Instance pools can inherit + permissions from their root + + object. + + + :param instance_pool_id: str + The instance pool for which to get or manage permissions. + :param access_control_list: + List[:class:`InstancePoolAccessControlRequest`] (optional) + + + :returns: :class:`InstancePoolPermissions` + parameters: + - name: instance_pool_id + in: path + required: true + schema: + type: string + description: The instance pool for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`InstancePoolPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/InstancePoolPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + /api/2.0/instance-pools/list: + get: + operationId: instance_pools_list + summary: Gets a list of instance pools with their statistics. + tags: + - compute + - instance_pools + description: |- + Gets a list of instance pools with their statistics. + + + :returns: Iterator over :class:`InstancePoolAndStats` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListInstancePools' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstancePoolsAPI + /api/2.0/instance-profiles/add: + post: + operationId: instance_profiles_add + summary: >- + Registers an instance profile in Databricks. In the UI, you can then + give users the permission to use + tags: + - compute + - instance_profiles + description: >- + Registers an instance profile in Databricks. In the UI, you can then + give users the permission to use + + this instance profile when launching clusters. + + + This API is only available to admin users. + + + :param instance_profile_arn: str + The AWS ARN of the instance profile to register with Databricks. This field is required. + :param iam_role_arn: str (optional) + The AWS IAM role ARN of the role associated with the instance profile. This field is required if + your role name and instance profile name do not match and you want to use the instance profile with + [Databricks SQL Serverless]. + + Otherwise, this field is optional. + + [Databricks SQL Serverless]: https://docs.databricks.com/sql/admin/serverless.html + :param is_meta_instance_profile: bool (optional) + Boolean flag indicating whether the instance profile should only be used in credential passthrough + scenarios. If true, it means the instance profile contains an meta IAM role which could assume a + wide range of roles. Therefore it should always be used with authorization. This field is optional, + the default value is `false`. + :param skip_validation: bool (optional) + By default, Databricks validates that it has sufficient permissions to launch instances with the + instance profile. This validation uses AWS dry-run mode for the RunInstances API. If validation + fails with an error message that does not indicate an IAM related permission issue, (e.g. “Your + requested instance type is not supported in your requested availability zone”), you can pass this + flag to skip the validation and forcibly add the instance profile. + requestBody: + content: + application/json: + schema: + type: object + properties: + instance_profile_arn: + type: string + description: >- + The AWS ARN of the instance profile to register with + Databricks. This field is required. + iam_role_arn: + type: string + description: >- + The AWS IAM role ARN of the role associated with the + instance profile. This field is required if your role name + and instance profile name do not match and you want to use + the instance profile with [Databricks SQL Serverless]. + Otherwise, this field is optional. [Databricks SQL + Serverless]: + https://docs.databricks.com/sql/admin/serverless.html + is_meta_instance_profile: + type: string + description: >- + Boolean flag indicating whether the instance profile should + only be used in credential passthrough scenarios. If true, + it means the instance profile contains an meta IAM role + which could assume a wide range of roles. Therefore it + should always be used with authorization. This field is + optional, the default value is `false`. + skip_validation: + type: string + description: >- + By default, Databricks validates that it has sufficient + permissions to launch instances with the instance profile. + This validation uses AWS dry-run mode for the RunInstances + API. If validation fails with an error message that does not + indicate an IAM related permission issue, (e.g. “Your + requested instance type is not supported in your requested + availability zone”), you can pass this flag to skip the + validation and forcibly add the instance profile. + required: + - instance_profile_arn + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstanceProfilesAPI + /api/2.0/instance-profiles/edit: + post: + operationId: instance_profiles_edit + summary: >- + The only supported field to change is the optional IAM role ARN + associated with the instance profile. + tags: + - compute + - instance_profiles + description: >- + The only supported field to change is the optional IAM role ARN + associated with the instance profile. + + It is required to specify the IAM role ARN if both of the following are + true: + + + * Your role name and instance profile name do not match. The name is the + part after the last slash in + + each ARN. * You want to use the instance profile with [Databricks SQL + Serverless]. + + + To understand where these fields are in the AWS console, see [Enable + serverless SQL warehouses]. + + + This API is only available to admin users. + + + [Databricks SQL Serverless]: + https://docs.databricks.com/sql/admin/serverless.html + + [Enable serverless SQL warehouses]: + https://docs.databricks.com/sql/admin/serverless.html + + + :param instance_profile_arn: str + The AWS ARN of the instance profile to register with Databricks. This field is required. + :param iam_role_arn: str (optional) + The AWS IAM role ARN of the role associated with the instance profile. This field is required if + your role name and instance profile name do not match and you want to use the instance profile with + [Databricks SQL Serverless]. + + Otherwise, this field is optional. + + [Databricks SQL Serverless]: https://docs.databricks.com/sql/admin/serverless.html + :param is_meta_instance_profile: bool (optional) + Boolean flag indicating whether the instance profile should only be used in credential passthrough + scenarios. If true, it means the instance profile contains an meta IAM role which could assume a + wide range of roles. Therefore it should always be used with authorization. This field is optional, + the default value is `false`. + requestBody: + content: + application/json: + schema: + type: object + properties: + instance_profile_arn: + type: string + description: >- + The AWS ARN of the instance profile to register with + Databricks. This field is required. + iam_role_arn: + type: string + description: >- + The AWS IAM role ARN of the role associated with the + instance profile. This field is required if your role name + and instance profile name do not match and you want to use + the instance profile with [Databricks SQL Serverless]. + Otherwise, this field is optional. [Databricks SQL + Serverless]: + https://docs.databricks.com/sql/admin/serverless.html + is_meta_instance_profile: + type: string + description: >- + Boolean flag indicating whether the instance profile should + only be used in credential passthrough scenarios. If true, + it means the instance profile contains an meta IAM role + which could assume a wide range of roles. Therefore it + should always be used with authorization. This field is + optional, the default value is `false`. + required: + - instance_profile_arn + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstanceProfilesAPI + /api/2.0/instance-profiles/list: + get: + operationId: instance_profiles_list + summary: >- + List the instance profiles that the calling user can use to launch a + cluster. + tags: + - compute + - instance_profiles + description: >- + List the instance profiles that the calling user can use to launch a + cluster. + + + This API is available to all users. + + + + :returns: Iterator over :class:`InstanceProfile` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListInstanceProfilesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstanceProfilesAPI + /api/2.0/instance-profiles/remove: + post: + operationId: instance_profiles_remove + summary: >- + Remove the instance profile with the provided ARN. Existing clusters + with this instance profile will + tags: + - compute + - instance_profiles + description: >- + Remove the instance profile with the provided ARN. Existing clusters + with this instance profile will + + continue to function. + + + This API is only accessible to admin users. + + + :param instance_profile_arn: str + The ARN of the instance profile to remove. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + instance_profile_arn: + type: string + description: >- + The ARN of the instance profile to remove. This field is + required. + required: + - instance_profile_arn + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: InstanceProfilesAPI + /api/2.0/libraries/all-cluster-statuses: + get: + operationId: libraries_all_cluster_statuses + summary: >- + Get the status of all libraries on all clusters. A status is returned + for all libraries installed on + tags: + - compute + - libraries + description: >- + Get the status of all libraries on all clusters. A status is returned + for all libraries installed on + + this cluster via the API or the libraries UI. + + + + :returns: Iterator over :class:`ClusterLibraryStatuses` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAllClusterLibraryStatusesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LibrariesAPI + /api/2.0/libraries/cluster-status: + get: + operationId: libraries_cluster_status + summary: >- + Get the status of libraries on a cluster. A status is returned for all + libraries installed on this + tags: + - compute + - libraries + description: >- + Get the status of libraries on a cluster. A status is returned for all + libraries installed on this + + cluster via the API or the libraries UI. The order of returned libraries + is as follows: 1. Libraries + + set to be installed on this cluster, in the order that the libraries + were added to the cluster, are + + returned first. 2. Libraries that were previously requested to be + installed on this cluster or, but + + are now marked for removal, in no particular order, are returned last. + + + :param cluster_id: str + Unique identifier of the cluster whose status should be retrieved. + + :returns: Iterator over :class:`LibraryFullStatus` + parameters: + - name: cluster_id + in: query + required: true + schema: + type: string + description: Unique identifier of the cluster whose status should be retrieved. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClusterLibraryStatuses' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LibrariesAPI + /api/2.0/libraries/install: + post: + operationId: libraries_install + summary: >- + Add libraries to install on a cluster. The installation is asynchronous; + it happens in the background + tags: + - compute + - libraries + description: >- + Add libraries to install on a cluster. The installation is asynchronous; + it happens in the background + + after the completion of this request. + + + :param cluster_id: str + Unique identifier for the cluster on which to install these libraries. + :param libraries: List[:class:`Library`] + The libraries to install. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: >- + Unique identifier for the cluster on which to install these + libraries. + libraries: + type: string + description: The libraries to install. + required: + - cluster_id + - libraries + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LibrariesAPI + /api/2.0/libraries/uninstall: + post: + operationId: libraries_uninstall + summary: >- + Set libraries to uninstall from a cluster. The libraries won't be + uninstalled until the cluster is + tags: + - compute + - libraries + description: >- + Set libraries to uninstall from a cluster. The libraries won't be + uninstalled until the cluster is + + restarted. A request to uninstall a library that is not currently + installed is ignored. + + + :param cluster_id: str + Unique identifier for the cluster on which to uninstall these libraries. + :param libraries: List[:class:`Library`] + The libraries to uninstall. + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: >- + Unique identifier for the cluster on which to uninstall + these libraries. + libraries: + type: string + description: The libraries to uninstall. + required: + - cluster_id + - libraries + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LibrariesAPI + /api/2.0/policies/clusters/enforce-compliance: + post: + operationId: policy_compliance_for_clusters_enforce_compliance + summary: >- + Updates a cluster to be compliant with the current version of its + policy. A cluster can be updated if + tags: + - compute + - policy_compliance_for_clusters + description: >- + Updates a cluster to be compliant with the current version of its + policy. A cluster can be updated if + + it is in a `RUNNING` or `TERMINATED` state. + + + If a cluster is updated while in a `RUNNING` state, it will be restarted + so that the new attributes + + can take effect. + + + If a cluster is updated while in a `TERMINATED` state, it will remain + `TERMINATED`. The next time the + + cluster is started, the new attributes will take effect. + + + Clusters created by the Databricks Jobs, DLT, or Models services cannot + be enforced by this API. + + Instead, use the "Enforce job policy compliance" API to enforce policy + compliance on jobs. + + + :param cluster_id: str + The ID of the cluster you want to enforce policy compliance on. + :param validate_only: bool (optional) + If set, previews the changes that would be made to a cluster to enforce compliance but does not + update the cluster. + + :returns: :class:`EnforceClusterComplianceResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + cluster_id: + type: string + description: >- + The ID of the cluster you want to enforce policy compliance + on. + validate_only: + type: string + description: >- + If set, previews the changes that would be made to a cluster + to enforce compliance but does not update the cluster. + required: + - cluster_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnforceClusterComplianceResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyComplianceForClustersAPI + /api/2.0/policies/clusters/get-compliance: + get: + operationId: policy_compliance_for_clusters_get_compliance + summary: >- + Returns the policy compliance status of a cluster. Clusters could be out + of compliance if their policy + tags: + - compute + - policy_compliance_for_clusters + description: >- + Returns the policy compliance status of a cluster. Clusters could be out + of compliance if their policy + + was updated after the cluster was last edited. + + + :param cluster_id: str + The ID of the cluster to get the compliance status + + :returns: :class:`GetClusterComplianceResponse` + parameters: + - name: cluster_id + in: query + required: true + schema: + type: string + description: The ID of the cluster to get the compliance status + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetClusterComplianceResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyComplianceForClustersAPI + /api/2.0/policies/clusters/list-compliance: + get: + operationId: policy_compliance_for_clusters_list_compliance + summary: >- + Returns the policy compliance status of all clusters that use a given + policy. Clusters could be out of + tags: + - compute + - policy_compliance_for_clusters + description: >- + Returns the policy compliance status of all clusters that use a given + policy. Clusters could be out of + + compliance if their policy was updated after the cluster was last + edited. + + + :param policy_id: str + Canonical unique identifier for the cluster policy. + :param page_size: int (optional) + Use this field to specify the maximum number of results to be returned by the server. The server may + further constrain the maximum number of results returned in a single page. + :param page_token: str (optional) + A page token that can be used to navigate to the next page or previous page as returned by + `next_page_token` or `prev_page_token`. + + :returns: Iterator over :class:`ClusterCompliance` + parameters: + - name: policy_id + in: query + required: true + schema: + type: string + description: Canonical unique identifier for the cluster policy. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + Use this field to specify the maximum number of results to be + returned by the server. The server may further constrain the maximum + number of results returned in a single page. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token that can be used to navigate to the next page or + previous page as returned by `next_page_token` or `prev_page_token`. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListClusterCompliancesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyComplianceForClustersAPI + /api/2.0/policy-families/{policy_family_id}: + get: + operationId: policy_families_get + summary: >- + Retrieve the information for an policy family based on its identifier + and version + tags: + - compute + - policy_families + description: >- + Retrieve the information for an policy family based on its identifier + and version + + + :param policy_family_id: str + The family ID about which to retrieve information. + :param version: int (optional) + The version number for the family to fetch. Defaults to the latest version. + + :returns: :class:`PolicyFamily` + parameters: + - name: policy_family_id + in: path + required: true + schema: + type: string + description: The family ID about which to retrieve information. + - name: version + in: query + required: false + schema: + type: string + description: >- + The version number for the family to fetch. Defaults to the latest + version. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PolicyFamily' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyFamiliesAPI + /api/2.0/policy-families: + get: + operationId: policy_families_list + summary: >- + Returns the list of policy definition types available to use at their + latest version. This API is + tags: + - compute + - policy_families + description: >- + Returns the list of policy definition types available to use at their + latest version. This API is + + paginated. + + + :param max_results: int (optional) + Maximum number of policy families to return. + :param page_token: str (optional) + A token that can be used to get the next page of results. + + :returns: Iterator over :class:`PolicyFamily` + parameters: + - name: max_results + in: query + required: false + schema: + type: string + description: Maximum number of policy families to return. + - name: page_token + in: query + required: false + schema: + type: string + description: A token that can be used to get the next page of results. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListPolicyFamiliesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyFamiliesAPI +components: + schemas: + AddResponse: + type: object + properties: {} + Adlsgen2Info: + type: object + properties: + destination: + type: string + description: >- + abfss destination, e.g. + `abfss://@.dfs.core.windows.net/`. + required: + - destination + description: A storage location in Adls Gen2 + AutoScale: + type: object + properties: + max_workers: + type: integer + min_workers: + type: integer + description: >- + The minimum number of workers to which the cluster can scale down + when underutilized. It is also the initial number of workers the + cluster will have after creation. + AwsAttributes: + type: object + properties: + availability: + $ref: '#/components/schemas/AwsAvailability' + ebs_volume_count: + type: integer + description: >- + The number of volumes launched for each instance. Users can choose + up to 10 volumes. This feature is only enabled for supported node + types. Legacy node types cannot specify custom EBS volumes. For node + types with no instance store, at least one EBS volume needs to be + specified; otherwise, cluster creation will fail. These EBS volumes + will be mounted at `/ebs0`, `/ebs1`, and etc. Instance store volumes + will be mounted at `/local_disk0`, `/local_disk1`, and etc. If EBS + volumes are attached, Databricks will configure Spark to use only + the EBS volumes for scratch storage because heterogenously sized + scratch devices can lead to inefficient disk utilization. If no EBS + volumes are attached, Databricks will configure Spark to use + instance store volumes. Please note that if EBS volumes are + specified, then the Spark configuration `spark.local.dir` will be + overridden. + ebs_volume_iops: + type: integer + description: >- + If using gp3 volumes, what IOPS to use for the disk. If this is not + set, the maximum performance of a gp2 volume with the same volume + size will be used. + ebs_volume_size: + type: integer + description: >- + The size of each EBS volume (in GiB) launched for each instance. For + general purpose SSD, this value must be within the range 100 - 4096. + For throughput optimized HDD, this value must be within the range + 500 - 4096. + ebs_volume_throughput: + type: integer + description: >- + If using gp3 volumes, what throughput to use for the disk. If this + is not set, the maximum performance of a gp2 volume with the same + volume size will be used. + ebs_volume_type: + $ref: '#/components/schemas/EbsVolumeType' + description: The type of EBS volumes that will be launched with this cluster. + first_on_demand: + type: integer + description: >- + The first `first_on_demand` nodes of the cluster will be placed on + on-demand instances. If this value is greater than 0, the cluster + driver node in particular will be placed on an on-demand instance. + If this value is greater than or equal to the current cluster size, + all nodes will be placed on on-demand instances. If this value is + less than the current cluster size, `first_on_demand` nodes will be + placed on on-demand instances and the remainder will be placed on + `availability` instances. Note that this value does not affect + cluster size and cannot currently be mutated over the lifetime of a + cluster. + instance_profile_arn: + type: string + description: >- + Nodes for this cluster will only be placed on AWS instances with + this instance profile. If ommitted, nodes will be placed on + instances without an IAM instance profile. The instance profile must + have previously been added to the Databricks environment by an + account administrator. This feature may only be available to certain + customer plans. + spot_bid_price_percent: + type: integer + description: >- + The bid price for AWS spot instances, as a percentage of the + corresponding instance type's on-demand price. For example, if this + field is set to 50, and the cluster needs a new `r3.xlarge` spot + instance, then the bid price is half of the price of on-demand + `r3.xlarge` instances. Similarly, if this field is set to 200, the + bid price is twice the price of on-demand `r3.xlarge` instances. If + not specified, the default value is 100. When spot instances are + requested for this cluster, only spot instances whose bid price + percentage matches this field will be considered. Note that, for + safety, we enforce this field to be no more than 10000. + zone_id: + type: string + description: >- + Identifier for the availability zone/datacenter in which the cluster + resides. This string will be of a form like "us-west-2a". The + provided availability zone must be in the same region as the + Databricks deployment. For example, "us-west-2a" is not a valid zone + id if the Databricks deployment resides in the "us-east-1" region. + This is an optional field at cluster creation, and if not specified, + the zone "auto" will be used. If the zone specified is "auto", will + try to place cluster in a zone with high availability, and will + retry placement in a different AZ if there is not enough capacity. + The list of available zones as well as the default value can be + found by using the `List Zones` method. + description: >- + Attributes set during cluster creation which are related to Amazon Web + Services. + AzureAttributes: + type: object + properties: + availability: + $ref: '#/components/schemas/AzureAvailability' + description: >- + Availability type used for all subsequent nodes past the + `first_on_demand` ones. Note: If `first_on_demand` is zero, this + availability type will be used for the entire cluster. + first_on_demand: + type: integer + description: >- + The first `first_on_demand` nodes of the cluster will be placed on + on-demand instances. This value should be greater than 0, to make + sure the cluster driver node is placed on an on-demand instance. If + this value is greater than or equal to the current cluster size, all + nodes will be placed on on-demand instances. If this value is less + than the current cluster size, `first_on_demand` nodes will be + placed on on-demand instances and the remainder will be placed on + `availability` instances. Note that this value does not affect + cluster size and cannot currently be mutated over the lifetime of a + cluster. + log_analytics_info: + $ref: '#/components/schemas/LogAnalyticsInfo' + description: >- + Defines values necessary to configure and run Azure Log Analytics + agent + spot_bid_max_price: + type: number + description: >- + The max bid price to be used for Azure spot instances. The Max price + for the bid cannot be higher than the on-demand price of the + instance. If not specified, the default value is -1, which specifies + that the instance cannot be evicted on the basis of price, and only + on the basis of availability. Further, the value should > 0 or -1. + description: >- + Attributes set during cluster creation which are related to Microsoft + Azure. + CancelResponse: + type: object + properties: {} + ChangeClusterOwnerResponse: + type: object + properties: {} + ClientsTypes: + type: object + properties: + jobs: + type: boolean + notebooks: + type: boolean + description: With notebooks set, this cluster can be used for notebooks + CloneCluster: + type: object + properties: + source_cluster_id: + type: string + required: + - source_cluster_id + CloudProviderNodeInfo: + type: object + properties: + status: + type: array + items: + $ref: '#/components/schemas/CloudProviderNodeStatus' + ClusterAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/ClusterPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + ClusterAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/ClusterPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + ClusterAttributes: + type: object + properties: + spark_version: + type: string + description: >- + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of + available Spark versions can be retrieved by using the + :method:clusters/sparkVersions API call. + autotermination_minutes: + type: integer + description: >- + Automatically terminates the cluster after it is inactive for this + time in minutes. If not set, this cluster will not be automatically + terminated. If specified, the threshold must be between 10 and 10000 + minutes. Users can also set this value to 0 to explicitly disable + automatic termination. + aws_attributes: + $ref: '#/components/schemas/AwsAttributes' + description: >- + Attributes related to clusters running on Amazon Web Services. If + not specified at cluster creation, a set of default values will be + used. + azure_attributes: + $ref: '#/components/schemas/AzureAttributes' + description: >- + Attributes related to clusters running on Microsoft Azure. If not + specified at cluster creation, a set of default values will be used. + cluster_log_conf: + $ref: '#/components/schemas/ClusterLogConf' + description: >- + The configuration for delivering spark logs to a long-term storage + destination. Three kinds of destinations (DBFS, S3 and Unity Catalog + volumes) are supported. Only one destination can be specified for + one cluster. If the conf is given, the logs will be delivered to the + destination every `5 mins`. The destination of driver logs is + `$destination/$clusterId/driver`, while the destination of executor + logs is `$destination/$clusterId/executor`. + cluster_name: + type: string + description: >- + Cluster name requested by the user. This doesn't have to be unique. + If not specified at creation, the cluster name will be an empty + string. For job clusters, the cluster name is automatically set + based on the job and job run IDs. + custom_tags: + type: object + description: >- + Additional tags for cluster resources. Databricks will tag all + cluster resources (e.g., AWS instances and EBS volumes) with these + tags in addition to `default_tags`. Notes: - Currently, Databricks + allows at most 45 custom tags - Clusters can only reuse cloud + resources if the resources' tags are a subset of the cluster tags + data_security_mode: + $ref: '#/components/schemas/DataSecurityMode' + docker_image: + $ref: '#/components/schemas/DockerImage' + description: Custom docker image BYOC + driver_instance_pool_id: + type: string + description: >- + The optional ID of the instance pool for the driver of the cluster + belongs. The pool cluster uses the instance pool with id + (instance_pool_id) if the driver pool is not assigned. + driver_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for the driver node. + driver_node_type_id: + type: string + description: >- + The node type of the Spark driver. Note that this field is optional; + if unset, the driver node type will be set as the same value as + `node_type_id` defined above. This field, along with node_type_id, + should not be set if virtual_cluster_size is set. If both + driver_node_type_id, node_type_id, and virtual_cluster_size are + specified, driver_node_type_id and node_type_id take precedence. + enable_elastic_disk: + type: boolean + description: >- + Autoscaling Local Storage: when enabled, this cluster will + dynamically acquire additional disk space when its Spark workers are + running low on disk space. + enable_local_disk_encryption: + type: boolean + description: Whether to enable LUKS on cluster VMs' local disks + gcp_attributes: + $ref: '#/components/schemas/GcpAttributes' + description: >- + Attributes related to clusters running on Google Cloud Platform. If + not specified at cluster creation, a set of default values will be + used. + init_scripts: + type: array + items: + $ref: '#/components/schemas/InitScriptInfo' + description: >- + The configuration for storing init scripts. Any number of + destinations can be specified. The scripts are executed sequentially + in the order provided. If `cluster_log_conf` is specified, init + script logs are sent to `//init_scripts`. + instance_pool_id: + type: string + description: The optional ID of the instance pool to which the cluster belongs. + is_single_node: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. When set + to true, Databricks will automatically set single node related + `custom_tags`, `spark_conf`, and `num_workers` + kind: + $ref: '#/components/schemas/Kind' + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources available + to each of the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved + by using the :method:clusters/listNodeTypes API call. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + remote_disk_throughput: + type: integer + description: >- + If set, what the configurable throughput (in Mb/s) for the remote + disk is. Currently only supported for GCP HYPERDISK_BALANCED disks. + runtime_engine: + $ref: '#/components/schemas/RuntimeEngine' + description: >- + Determines the cluster's runtime engine, either standard or Photon. + This field is not compatible with legacy `spark_version` values that + contain `-photon-`. Remove `-photon-` from the `spark_version` and + set `runtime_engine` to `PHOTON`. If left unspecified, the runtime + engine defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + single_user_name: + type: string + description: Single user name if data_security_mode is `SINGLE_USER` + spark_conf: + type: object + description: >- + An object containing a set of optional, user-specified Spark + configuration key-value pairs. Users can also pass in a string of + extra JVM options to the driver and the executors via + `spark.driver.extraJavaOptions` and + `spark.executor.extraJavaOptions` respectively. + spark_env_vars: + type: object + description: >- + An object containing a set of optional, user-specified environment + variable key-value pairs. Please note that key-value pair of the + form (X,Y) will be exported as is (i.e., `export X='Y'`) while + launching the driver and workers. In order to specify an additional + set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This + ensures that all default databricks managed environmental variables + are included as well. Example Spark environment variables: + `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": + "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": + "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}` + ssh_public_keys: + type: array + items: + type: string + description: >- + SSH public key contents that will be added to each Spark node in + this cluster. The corresponding private keys can be used to login + with the user name `ubuntu` on port `2200`. Up to 10 keys can be + specified. + total_initial_remote_disk_size: + type: integer + description: >- + If set, what the total initial volume size (in GB) of the remote + disks should be. Currently only supported for GCP HYPERDISK_BALANCED + disks. + use_ml_runtime: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + `effective_spark_version` is determined by `spark_version` (DBR + release), this field `use_ml_runtime`, and whether `node_type_id` is + gpu node or not. + worker_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for worker nodes. + workload_type: + $ref: '#/components/schemas/WorkloadType' + required: + - spark_version + description: >- + Common set of attributes set during cluster creation. These attributes + cannot be changed over + the lifetime of a cluster. + ClusterCompliance: + type: object + properties: + cluster_id: + type: string + is_compliant: + type: boolean + description: >- + Whether this cluster is in compliance with the latest version of its + policy. + violations: + type: object + description: >- + An object containing key-value mappings representing the first 200 + policy validation errors. The keys indicate the path where the + policy validation error is occurring. The values indicate an error + message describing the policy validation error. + required: + - cluster_id + ClusterDetails: + type: object + properties: + autoscale: + $ref: '#/components/schemas/AutoScale' + description: >- + Parameters needed in order to automatically scale clusters up and + down based on load. Note: autoscaling works best with DB runtime + versions 3.0 or later. + autotermination_minutes: + type: integer + description: >- + Automatically terminates the cluster after it is inactive for this + time in minutes. If not set, this cluster will not be automatically + terminated. If specified, the threshold must be between 10 and 10000 + minutes. Users can also set this value to 0 to explicitly disable + automatic termination. + aws_attributes: + $ref: '#/components/schemas/AwsAttributes' + description: >- + Attributes related to clusters running on Amazon Web Services. If + not specified at cluster creation, a set of default values will be + used. + azure_attributes: + $ref: '#/components/schemas/AzureAttributes' + description: >- + Attributes related to clusters running on Microsoft Azure. If not + specified at cluster creation, a set of default values will be used. + cluster_cores: + type: number + description: >- + Number of CPU cores available for this cluster. Note that this can + be fractional, e.g. 7.5 cores, since certain node types are + configured to share cores between Spark nodes on the same instance. + cluster_id: + type: string + description: >- + Canonical identifier for the cluster. This id is retained during + cluster restarts and resizes, while each new cluster has a globally + unique id. + cluster_log_conf: + $ref: '#/components/schemas/ClusterLogConf' + description: >- + The configuration for delivering spark logs to a long-term storage + destination. Three kinds of destinations (DBFS, S3 and Unity Catalog + volumes) are supported. Only one destination can be specified for + one cluster. If the conf is given, the logs will be delivered to the + destination every `5 mins`. The destination of driver logs is + `$destination/$clusterId/driver`, while the destination of executor + logs is `$destination/$clusterId/executor`. + cluster_log_status: + $ref: '#/components/schemas/LogSyncStatus' + description: Cluster log delivery status. + cluster_memory_mb: + type: integer + description: Total amount of cluster memory, in megabytes + cluster_name: + type: string + description: >- + Cluster name requested by the user. This doesn't have to be unique. + If not specified at creation, the cluster name will be an empty + string. For job clusters, the cluster name is automatically set + based on the job and job run IDs. + cluster_source: + $ref: '#/components/schemas/ClusterSource' + description: >- + Determines whether the cluster was created by a user through the UI, + created by the Databricks Jobs Scheduler, or through an API request. + creator_user_name: + type: string + description: >- + Creator user name. The field won't be included in the response if + the user has already been deleted. + custom_tags: + type: object + description: >- + Additional tags for cluster resources. Databricks will tag all + cluster resources (e.g., AWS instances and EBS volumes) with these + tags in addition to `default_tags`. Notes: - Currently, Databricks + allows at most 45 custom tags - Clusters can only reuse cloud + resources if the resources' tags are a subset of the cluster tags + data_security_mode: + $ref: '#/components/schemas/DataSecurityMode' + default_tags: + type: object + description: >- + Tags that are added by Databricks regardless of any `custom_tags`, + including: - Vendor: Databricks - Creator: - + ClusterName: - ClusterId: - Name: + + docker_image: + $ref: '#/components/schemas/DockerImage' + description: Custom docker image BYOC + driver: + $ref: '#/components/schemas/SparkNode' + description: >- + Node on which the Spark driver resides. The driver node contains the + Spark master and the Databricks application that manages the + per-notebook Spark REPLs. + driver_instance_pool_id: + type: string + description: >- + The optional ID of the instance pool for the driver of the cluster + belongs. The pool cluster uses the instance pool with id + (instance_pool_id) if the driver pool is not assigned. + driver_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for the driver node. + driver_node_type_id: + type: string + description: >- + The node type of the Spark driver. Note that this field is optional; + if unset, the driver node type will be set as the same value as + `node_type_id` defined above. This field, along with node_type_id, + should not be set if virtual_cluster_size is set. If both + driver_node_type_id, node_type_id, and virtual_cluster_size are + specified, driver_node_type_id and node_type_id take precedence. + enable_elastic_disk: + type: boolean + description: >- + Autoscaling Local Storage: when enabled, this cluster will + dynamically acquire additional disk space when its Spark workers are + running low on disk space. + enable_local_disk_encryption: + type: boolean + description: Whether to enable LUKS on cluster VMs' local disks + executors: + type: array + items: + $ref: '#/components/schemas/SparkNode' + description: Nodes on which the Spark executors reside. + gcp_attributes: + $ref: '#/components/schemas/GcpAttributes' + description: >- + Attributes related to clusters running on Google Cloud Platform. If + not specified at cluster creation, a set of default values will be + used. + init_scripts: + type: array + items: + $ref: '#/components/schemas/InitScriptInfo' + description: >- + The configuration for storing init scripts. Any number of + destinations can be specified. The scripts are executed sequentially + in the order provided. If `cluster_log_conf` is specified, init + script logs are sent to `//init_scripts`. + instance_pool_id: + type: string + description: The optional ID of the instance pool to which the cluster belongs. + is_single_node: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. When set + to true, Databricks will automatically set single node related + `custom_tags`, `spark_conf`, and `num_workers` + jdbc_port: + type: integer + description: >- + Port on which Spark JDBC server is listening, in the driver nod. No + service will be listeningon on this port in executor nodes. + kind: + $ref: '#/components/schemas/Kind' + last_restarted_time: + type: integer + description: the timestamp that the cluster was started/restarted + last_state_loss_time: + type: integer + description: >- + Time when the cluster driver last lost its state (due to a restart + or driver failure). + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources available + to each of the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved + by using the :method:clusters/listNodeTypes API call. + num_workers: + type: integer + description: >- + Number of worker nodes that this cluster should have. A cluster has + one Spark Driver and `num_workers` Executors for a total of + `num_workers` + 1 Spark nodes. Note: When reading the properties of + a cluster, this field reflects the desired number of workers rather + than the actual current number of workers. For instance, if a + cluster is resized from 5 to 10 workers, this field will immediately + be updated to reflect the target size of 10 workers, whereas the + workers listed in `spark_info` will gradually increase from 5 to 10 + as the new nodes are provisioned. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + remote_disk_throughput: + type: integer + description: >- + If set, what the configurable throughput (in Mb/s) for the remote + disk is. Currently only supported for GCP HYPERDISK_BALANCED disks. + runtime_engine: + $ref: '#/components/schemas/RuntimeEngine' + description: >- + Determines the cluster's runtime engine, either standard or Photon. + This field is not compatible with legacy `spark_version` values that + contain `-photon-`. Remove `-photon-` from the `spark_version` and + set `runtime_engine` to `PHOTON`. If left unspecified, the runtime + engine defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + single_user_name: + type: string + description: Single user name if data_security_mode is `SINGLE_USER` + spark_conf: + type: object + description: >- + An object containing a set of optional, user-specified Spark + configuration key-value pairs. Users can also pass in a string of + extra JVM options to the driver and the executors via + `spark.driver.extraJavaOptions` and + `spark.executor.extraJavaOptions` respectively. + spark_context_id: + type: integer + description: >- + A canonical SparkContext identifier. This value *does* change when + the Spark driver restarts. The pair `(cluster_id, spark_context_id)` + is a globally unique identifier over all Spark contexts. + spark_env_vars: + type: object + description: >- + An object containing a set of optional, user-specified environment + variable key-value pairs. Please note that key-value pair of the + form (X,Y) will be exported as is (i.e., `export X='Y'`) while + launching the driver and workers. In order to specify an additional + set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This + ensures that all default databricks managed environmental variables + are included as well. Example Spark environment variables: + `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": + "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": + "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}` + spark_version: + type: string + description: >- + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of + available Spark versions can be retrieved by using the + :method:clusters/sparkVersions API call. + spec: + $ref: '#/components/schemas/ClusterSpec' + description: >- + The spec contains a snapshot of the latest user specified settings + that were used to create/edit the cluster. Note: not included in the + response of the ListClusters API. + ssh_public_keys: + type: array + items: + type: string + description: >- + SSH public key contents that will be added to each Spark node in + this cluster. The corresponding private keys can be used to login + with the user name `ubuntu` on port `2200`. Up to 10 keys can be + specified. + start_time: + type: integer + description: >- + Time (in epoch milliseconds) when the cluster creation request was + received (when the cluster entered a `PENDING` state). + state: + $ref: '#/components/schemas/State' + description: Current state of the cluster. + state_message: + type: string + description: >- + A message associated with the most recent state transition (e.g., + the reason why the cluster entered a `TERMINATED` state). + terminated_time: + type: integer + description: >- + Time (in epoch milliseconds) when the cluster was terminated, if + applicable. + termination_reason: + $ref: '#/components/schemas/TerminationReason' + description: >- + Information about why the cluster was terminated. This field only + appears when the cluster is in a `TERMINATING` or `TERMINATED` + state. + total_initial_remote_disk_size: + type: integer + description: >- + If set, what the total initial volume size (in GB) of the remote + disks should be. Currently only supported for GCP HYPERDISK_BALANCED + disks. + use_ml_runtime: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + `effective_spark_version` is determined by `spark_version` (DBR + release), this field `use_ml_runtime`, and whether `node_type_id` is + gpu node or not. + worker_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for worker nodes. + workload_type: + $ref: '#/components/schemas/WorkloadType' + description: >- + Describes all of the metadata about a single Spark cluster in + Databricks. + ClusterEvent: + type: object + properties: + cluster_id: + type: string + data_plane_event_details: + $ref: '#/components/schemas/DataPlaneEventDetails' + details: + $ref: '#/components/schemas/EventDetails' + timestamp: + type: integer + description: >- + The timestamp when the event occurred, stored as the number of + milliseconds since the Unix epoch. If not provided, this will be + assigned by the Timeline service. + type: + $ref: '#/components/schemas/EventType' + required: + - cluster_id + ClusterLibraryStatuses: + type: object + properties: + cluster_id: + type: string + library_statuses: + type: array + items: + $ref: '#/components/schemas/LibraryFullStatus' + description: Status of all libraries on the cluster. + ClusterLogConf: + type: object + properties: + dbfs: + $ref: '#/components/schemas/DbfsStorageInfo' + description: >- + destination needs to be provided. e.g. `{ "dbfs" : { "destination" : + "dbfs:/home/cluster_log" } }` + s3: + $ref: '#/components/schemas/S3StorageInfo' + description: >- + destination and either the region or endpoint need to be provided. + e.g. `{ "s3": { "destination" : "s3://cluster_log_bucket/prefix", + "region" : "us-west-2" } }` Cluster iam role is used to access s3, + please make sure the cluster iam role in `instance_profile_arn` has + permission to write data to the s3 destination. + volumes: + $ref: '#/components/schemas/VolumesStorageInfo' + description: >- + destination needs to be provided, e.g. `{ "volumes": { + "destination": "/Volumes/catalog/schema/volume/cluster_log" } }` + description: Cluster log delivery config + ClusterPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/ClusterPermissionLevel' + ClusterPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/ClusterAccessControlResponse' + object_id: + type: string + object_type: + type: string + ClusterPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/ClusterPermissionLevel' + ClusterPolicyAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/ClusterPolicyPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + ClusterPolicyAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/ClusterPolicyPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + ClusterPolicyPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/ClusterPolicyPermissionLevel' + ClusterPolicyPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/ClusterPolicyAccessControlResponse' + object_id: + type: string + object_type: + type: string + ClusterPolicyPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/ClusterPolicyPermissionLevel' + ClusterSettingsChange: + type: object + properties: + field: + type: string + description: The field where this change would be made. + new_value: + type: string + description: >- + The new value of this field after enforcing policy compliance + (either a number, a boolean, or a string) converted to a string. + This is intended to be read by a human. The typed new value of this + field can be retrieved by reading the settings field in the API + response. + previous_value: + type: string + description: >- + The previous value of this field before enforcing policy compliance + (either a number, a boolean, or a string) converted to a string. + This is intended to be read by a human. The type of the field can be + retrieved by reading the settings field in the API response. + description: >- + Represents a change to the cluster settings required for the cluster to + become compliant with + its policy. + ClusterSize: + type: object + properties: + autoscale: + $ref: '#/components/schemas/AutoScale' + num_workers: + type: integer + description: >- + Number of worker nodes that this cluster should have. A cluster has + one Spark Driver and `num_workers` Executors for a total of + `num_workers` + 1 Spark nodes. Note: When reading the properties of + a cluster, this field reflects the desired number of workers rather + than the actual current number of workers. For instance, if a + cluster is resized from 5 to 10 workers, this field will immediately + be updated to reflect the target size of 10 workers, whereas the + workers listed in `spark_info` will gradually increase from 5 to 10 + as the new nodes are provisioned. + ClusterSpec: + type: object + properties: + apply_policy_default_values: + type: boolean + description: >- + When set to true, fixed and default values from the policy will be + used for fields that are omitted. When set to false, only fixed + values from the policy will be applied. + autoscale: + $ref: '#/components/schemas/AutoScale' + description: >- + Parameters needed in order to automatically scale clusters up and + down based on load. Note: autoscaling works best with DB runtime + versions 3.0 or later. + autotermination_minutes: + type: integer + description: >- + Automatically terminates the cluster after it is inactive for this + time in minutes. If not set, this cluster will not be automatically + terminated. If specified, the threshold must be between 10 and 10000 + minutes. Users can also set this value to 0 to explicitly disable + automatic termination. + aws_attributes: + $ref: '#/components/schemas/AwsAttributes' + description: >- + Attributes related to clusters running on Amazon Web Services. If + not specified at cluster creation, a set of default values will be + used. + azure_attributes: + $ref: '#/components/schemas/AzureAttributes' + description: >- + Attributes related to clusters running on Microsoft Azure. If not + specified at cluster creation, a set of default values will be used. + cluster_log_conf: + $ref: '#/components/schemas/ClusterLogConf' + description: >- + The configuration for delivering spark logs to a long-term storage + destination. Three kinds of destinations (DBFS, S3 and Unity Catalog + volumes) are supported. Only one destination can be specified for + one cluster. If the conf is given, the logs will be delivered to the + destination every `5 mins`. The destination of driver logs is + `$destination/$clusterId/driver`, while the destination of executor + logs is `$destination/$clusterId/executor`. + cluster_name: + type: string + description: >- + Cluster name requested by the user. This doesn't have to be unique. + If not specified at creation, the cluster name will be an empty + string. For job clusters, the cluster name is automatically set + based on the job and job run IDs. + custom_tags: + type: object + description: >- + Additional tags for cluster resources. Databricks will tag all + cluster resources (e.g., AWS instances and EBS volumes) with these + tags in addition to `default_tags`. Notes: - Currently, Databricks + allows at most 45 custom tags - Clusters can only reuse cloud + resources if the resources' tags are a subset of the cluster tags + data_security_mode: + $ref: '#/components/schemas/DataSecurityMode' + docker_image: + $ref: '#/components/schemas/DockerImage' + description: Custom docker image BYOC + driver_instance_pool_id: + type: string + description: >- + The optional ID of the instance pool for the driver of the cluster + belongs. The pool cluster uses the instance pool with id + (instance_pool_id) if the driver pool is not assigned. + driver_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for the driver node. + driver_node_type_id: + type: string + description: >- + The node type of the Spark driver. Note that this field is optional; + if unset, the driver node type will be set as the same value as + `node_type_id` defined above. This field, along with node_type_id, + should not be set if virtual_cluster_size is set. If both + driver_node_type_id, node_type_id, and virtual_cluster_size are + specified, driver_node_type_id and node_type_id take precedence. + enable_elastic_disk: + type: boolean + description: >- + Autoscaling Local Storage: when enabled, this cluster will + dynamically acquire additional disk space when its Spark workers are + running low on disk space. + enable_local_disk_encryption: + type: boolean + description: Whether to enable LUKS on cluster VMs' local disks + gcp_attributes: + $ref: '#/components/schemas/GcpAttributes' + description: >- + Attributes related to clusters running on Google Cloud Platform. If + not specified at cluster creation, a set of default values will be + used. + init_scripts: + type: array + items: + $ref: '#/components/schemas/InitScriptInfo' + description: >- + The configuration for storing init scripts. Any number of + destinations can be specified. The scripts are executed sequentially + in the order provided. If `cluster_log_conf` is specified, init + script logs are sent to `//init_scripts`. + instance_pool_id: + type: string + description: The optional ID of the instance pool to which the cluster belongs. + is_single_node: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. When set + to true, Databricks will automatically set single node related + `custom_tags`, `spark_conf`, and `num_workers` + kind: + $ref: '#/components/schemas/Kind' + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources available + to each of the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved + by using the :method:clusters/listNodeTypes API call. + num_workers: + type: integer + description: >- + Number of worker nodes that this cluster should have. A cluster has + one Spark Driver and `num_workers` Executors for a total of + `num_workers` + 1 Spark nodes. Note: When reading the properties of + a cluster, this field reflects the desired number of workers rather + than the actual current number of workers. For instance, if a + cluster is resized from 5 to 10 workers, this field will immediately + be updated to reflect the target size of 10 workers, whereas the + workers listed in `spark_info` will gradually increase from 5 to 10 + as the new nodes are provisioned. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + remote_disk_throughput: + type: integer + description: >- + If set, what the configurable throughput (in Mb/s) for the remote + disk is. Currently only supported for GCP HYPERDISK_BALANCED disks. + runtime_engine: + $ref: '#/components/schemas/RuntimeEngine' + description: >- + Determines the cluster's runtime engine, either standard or Photon. + This field is not compatible with legacy `spark_version` values that + contain `-photon-`. Remove `-photon-` from the `spark_version` and + set `runtime_engine` to `PHOTON`. If left unspecified, the runtime + engine defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + single_user_name: + type: string + description: Single user name if data_security_mode is `SINGLE_USER` + spark_conf: + type: object + description: >- + An object containing a set of optional, user-specified Spark + configuration key-value pairs. Users can also pass in a string of + extra JVM options to the driver and the executors via + `spark.driver.extraJavaOptions` and + `spark.executor.extraJavaOptions` respectively. + spark_env_vars: + type: object + description: >- + An object containing a set of optional, user-specified environment + variable key-value pairs. Please note that key-value pair of the + form (X,Y) will be exported as is (i.e., `export X='Y'`) while + launching the driver and workers. In order to specify an additional + set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This + ensures that all default databricks managed environmental variables + are included as well. Example Spark environment variables: + `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": + "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": + "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}` + spark_version: + type: string + description: >- + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of + available Spark versions can be retrieved by using the + :method:clusters/sparkVersions API call. + ssh_public_keys: + type: array + items: + type: string + description: >- + SSH public key contents that will be added to each Spark node in + this cluster. The corresponding private keys can be used to login + with the user name `ubuntu` on port `2200`. Up to 10 keys can be + specified. + total_initial_remote_disk_size: + type: integer + description: >- + If set, what the total initial volume size (in GB) of the remote + disks should be. Currently only supported for GCP HYPERDISK_BALANCED + disks. + use_ml_runtime: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + `effective_spark_version` is determined by `spark_version` (DBR + release), this field `use_ml_runtime`, and whether `node_type_id` is + gpu node or not. + worker_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for worker nodes. + workload_type: + $ref: '#/components/schemas/WorkloadType' + description: >- + Contains a snapshot of the latest user specified settings that were used + to create/edit the + cluster. + CommandStatusResponse: + type: object + properties: + id: + type: string + results: + $ref: '#/components/schemas/Results' + status: + $ref: '#/components/schemas/CommandStatus' + ContextStatusResponse: + type: object + properties: + id: + type: string + status: + $ref: '#/components/schemas/ContextStatus' + CreateClusterResponse: + type: object + properties: + cluster_id: + type: string + CreateInstancePoolResponse: + type: object + properties: + instance_pool_id: + type: string + CreatePolicyResponse: + type: object + properties: + policy_id: + type: string + CreateResponse: + type: object + properties: + script_id: + type: string + Created: + type: object + properties: + id: + type: string + CustomPolicyTag: + type: object + properties: + key: + type: string + value: + type: string + description: The value of the tag. + required: + - key + DataPlaneEventDetails: + type: object + properties: + event_type: + $ref: '#/components/schemas/DataPlaneEventDetailsEventType' + executor_failures: + type: integer + host_id: + type: string + timestamp: + type: integer + DbfsStorageInfo: + type: object + properties: + destination: + type: string + description: dbfs destination, e.g. `dbfs:/my/path` + required: + - destination + description: A storage location in DBFS + DeleteClusterResponse: + type: object + properties: {} + DeleteInstancePoolResponse: + type: object + properties: {} + DeletePolicyResponse: + type: object + properties: {} + DeleteResponse: + type: object + properties: {} + DestroyResponse: + type: object + properties: {} + DiskSpec: + type: object + properties: + disk_count: + type: integer + description: >- + The number of disks launched for each instance: - This feature is + only enabled for supported node types. - Users can choose up to the + limit of the disks supported by the node type. - For node types with + no OS disk, at least one disk must be specified; otherwise, cluster + creation will fail. If disks are attached, Databricks will configure + Spark to use only the disks for scratch storage, because + heterogenously sized scratch devices can lead to inefficient disk + utilization. If no disks are attached, Databricks will configure + Spark to use instance store disks. Note: If disks are specified, + then the Spark configuration `spark.local.dir` will be overridden. + Disks will be mounted at: - For AWS: `/ebs0`, `/ebs1`, and etc. - + For Azure: `/remote_volume0`, `/remote_volume1`, and etc. + disk_iops: + type: integer + disk_size: + type: integer + description: >- + The size of each disk (in GiB) launched for each instance. Values + must fall into the supported range for a particular instance type. + For AWS: - General Purpose SSD: 100 - 4096 GiB - Throughput + Optimized HDD: 500 - 4096 GiB For Azure: - Premium LRS (SSD): 1 - + 1023 GiB - Standard LRS (HDD): 1- 1023 GiB + disk_throughput: + type: integer + disk_type: + $ref: '#/components/schemas/DiskType' + description: The type of disks that will be launched with this cluster. + description: >- + Describes the disks that are launched for each instance in the spark + cluster. For example, if + the cluster has 3 instances, each instance is configured to launch 2 disks, 100 GiB each, then + Databricks will launch a total of 6 disks, 100 GiB each, for this cluster. + DiskType: + type: object + properties: + azure_disk_volume_type: + $ref: '#/components/schemas/DiskTypeAzureDiskVolumeType' + ebs_volume_type: + $ref: '#/components/schemas/DiskTypeEbsVolumeType' + description: Describes the disk type. + DockerBasicAuth: + type: object + properties: + password: + type: string + username: + type: string + description: Name of the user + DockerImage: + type: object + properties: + basic_auth: + $ref: '#/components/schemas/DockerBasicAuth' + url: + type: string + description: URL of the docker image. + EditClusterResponse: + type: object + properties: {} + EditInstancePoolResponse: + type: object + properties: {} + EditPolicyResponse: + type: object + properties: {} + EditResponse: + type: object + properties: {} + EnforceClusterComplianceResponse: + type: object + properties: + changes: + type: array + items: + $ref: '#/components/schemas/ClusterSettingsChange' + has_changes: + type: boolean + description: >- + Whether any changes have been made to the cluster settings for the + cluster to become compliant with its policy. + Environment: + type: object + properties: + base_environment: + type: string + description: >- + The `base_environment` key refers to an `env.yaml` file that + specifies an environment version and a collection of dependencies + required for the environment setup. This `env.yaml` file may itself + include a `base_environment` reference pointing to another + `env_1.yaml` file. However, when used as a base environment, + `env_1.yaml` (or further nested references) will not be processed or + included in the final environment, meaning that the resolution of + `base_environment` references is not recursive. + client: + type: string + description: Use `environment_version` instead. + dependencies: + type: array + items: + type: string + description: >- + List of pip dependencies, as supported by the version of pip in this + environment. Each dependency is a valid pip requirements file line + per + https://pip.pypa.io/en/stable/reference/requirements-file-format/. + Allowed dependencies include a requirement specifier, an archive + URL, a local project path (such as WSFS or UC Volumes in + Databricks), or a VCS project URL. + environment_version: + type: string + description: >- + Required. Environment version used by the environment. Each version + comes with a specific Python version and a set of Python packages. + The version is a string, consisting of an integer. + java_dependencies: + type: array + items: + type: string + description: >- + List of java dependencies. Each dependency is a string representing + a java library path. For example: `/Volumes/path/to/test.jar`. + description: >- + The environment entity used to preserve serverless environment side + panel, jobs' environment for + non-notebook task, and DLT's environment for classic and serverless pipelines. In this minimal + environment spec, only pip dependencies are supported. + EventDetails: + type: object + properties: + attributes: + $ref: '#/components/schemas/ClusterAttributes' + cause: + $ref: '#/components/schemas/EventDetailsCause' + description: The cause of a change in target size. + cluster_size: + $ref: '#/components/schemas/ClusterSize' + description: >- + The actual cluster size that was set in the cluster creation or + edit. + current_num_vcpus: + type: integer + description: The current number of vCPUs in the cluster. + current_num_workers: + type: integer + description: The current number of nodes in the cluster. + did_not_expand_reason: + type: string + disk_size: + type: integer + description: Current disk size in bytes + driver_state_message: + type: string + description: More details about the change in driver's state + enable_termination_for_node_blocklisted: + type: boolean + description: >- + Whether or not a blocklisted node should be terminated. For + ClusterEventType NODE_BLACKLISTED. + free_space: + type: integer + init_scripts: + $ref: '#/components/schemas/InitScriptEventDetails' + description: >- + List of global and cluster init scripts associated with this cluster + event. + instance_id: + type: string + description: Instance Id where the event originated from + job_run_name: + type: string + description: >- + Unique identifier of the specific job run associated with this + cluster event * For clusters created for jobs, this will be the same + as the cluster name + previous_attributes: + $ref: '#/components/schemas/ClusterAttributes' + description: The cluster attributes before a cluster was edited. + previous_cluster_size: + $ref: '#/components/schemas/ClusterSize' + description: The size of the cluster before an edit or resize. + previous_disk_size: + type: integer + description: Previous disk size in bytes + reason: + $ref: '#/components/schemas/TerminationReason' + description: >- + A termination reason: * On a TERMINATED event, this is the reason of + the termination. * On a RESIZE_COMPLETE event, this indicates the + reason that we failed to acquire some nodes. + target_num_vcpus: + type: integer + description: The targeted number of vCPUs in the cluster. + target_num_workers: + type: integer + description: The targeted number of nodes in the cluster. + user: + type: string + description: >- + The user that caused the event to occur. (Empty if it was done by + the control plane.) + GcpAttributes: + type: object + properties: + availability: + $ref: '#/components/schemas/GcpAvailability' + description: >- + This field determines whether the spark executors will be scheduled + to run on preemptible VMs, on-demand VMs, or preemptible VMs with a + fallback to on-demand VMs if the former is unavailable. + boot_disk_size: + type: integer + description: Boot disk size in GB + first_on_demand: + type: integer + description: >- + The first `first_on_demand` nodes of the cluster will be placed on + on-demand instances. This value should be greater than 0, to make + sure the cluster driver node is placed on an on-demand instance. If + this value is greater than or equal to the current cluster size, all + nodes will be placed on on-demand instances. If this value is less + than the current cluster size, `first_on_demand` nodes will be + placed on on-demand instances and the remainder will be placed on + `availability` instances. Note that this value does not affect + cluster size and cannot currently be mutated over the lifetime of a + cluster. + google_service_account: + type: string + description: >- + If provided, the cluster will impersonate the google service account + when accessing gcloud services (like GCS). The google service + account must have previously been added to the Databricks + environment by an account administrator. + local_ssd_count: + type: integer + description: >- + If provided, each node (workers and driver) in the cluster will have + this number of local SSDs attached. Each local SSD is 375GB in size. + Refer to [GCP documentation] for the supported number of local SSDs + for each instance type. [GCP documentation]: + https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds + use_preemptible_executors: + type: boolean + description: >- + This field determines whether the spark executors will be scheduled + to run on preemptible VMs (when set to true) versus standard compute + engine VMs (when set to false; default). Note: Soon to be + deprecated, use the 'availability' field instead. + zone_id: + type: string + description: >- + Identifier for the availability zone in which the cluster resides. + This can be one of the following: - "HA" => High availability, + spread nodes across availability zones for a Databricks deployment + region [default]. - "AUTO" => Databricks picks an availability zone + to schedule the cluster on. - A GCP availability zone => Pick One of + the available zones for (machine type + region) from + https://cloud.google.com/compute/docs/regions-zones. + description: Attributes set during cluster creation which are related to GCP. + GcsStorageInfo: + type: object + properties: + destination: + type: string + description: GCS destination/URI, e.g. `gs://my-bucket/some-prefix` + required: + - destination + description: A storage location in Google Cloud Platform's GCS + GetClusterComplianceResponse: + type: object + properties: + is_compliant: + type: boolean + violations: + type: object + description: >- + An object containing key-value mappings representing the first 200 + policy validation errors. The keys indicate the path where the + policy validation error is occurring. The values indicate an error + message describing the policy validation error. + GetClusterPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/ClusterPermissionsDescription' + GetClusterPolicyPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/ClusterPolicyPermissionsDescription' + GetEvents: + type: object + properties: + cluster_id: + type: string + end_time: + type: integer + description: >- + The end time in epoch milliseconds. If empty, returns events up to + the current time. + event_types: + type: array + items: + $ref: '#/components/schemas/EventType' + description: >- + An optional set of event types to filter on. If empty, all event + types are returned. + limit: + type: integer + description: >- + Deprecated: use page_token in combination with page_size instead. + The maximum number of events to include in a page of events. + Defaults to 50, and maximum allowed value is 500. + offset: + type: integer + description: >- + Deprecated: use page_token in combination with page_size instead. + The offset in the result set. Defaults to 0 (no offset). When an + offset is specified and the results are requested in descending + order, the end_time field is required. + order: + $ref: '#/components/schemas/GetEventsOrder' + description: >- + The order to list events in; either "ASC" or "DESC". Defaults to + "DESC". + page_size: + type: integer + description: >- + The maximum number of events to include in a page of events. The + server may further constrain the maximum number of results returned + in a single page. If the page_size is empty or 0, the server will + decide the number of results to be returned. The field has to be in + the range [0,500]. If the value is outside the range, the server + enforces 0 or 500. + page_token: + type: string + description: >- + Use next_page_token or prev_page_token returned from the previous + request to list the next or previous page of events respectively. If + page_token is empty, the first page is returned. + start_time: + type: integer + description: >- + The start time in epoch milliseconds. If empty, returns events + starting from the beginning of time. + required: + - cluster_id + GetEventsResponse: + type: object + properties: + events: + type: array + items: + $ref: '#/components/schemas/ClusterEvent' + next_page: + $ref: '#/components/schemas/GetEvents' + description: >- + Deprecated: use next_page_token or prev_page_token instead. The + parameters required to retrieve the next page of events. Omitted if + there are no more events to read. + next_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the next page + of results. If the value is "", it means no further results for the + request. + prev_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the previous + page of results. If the value is "", it means no further results for + the request. + total_count: + type: integer + description: >- + Deprecated: Returns 0 when request uses page_token. Will start + returning zero when request uses offset/limit soon. The total number + of events filtered by the start_time, end_time, and event_types. + GetInstancePool: + type: object + properties: + instance_pool_id: + type: string + aws_attributes: + $ref: '#/components/schemas/InstancePoolAwsAttributes' + description: >- + Attributes related to instance pools running on Amazon Web Services. + If not specified at pool creation, a set of default values will be + used. + azure_attributes: + $ref: '#/components/schemas/InstancePoolAzureAttributes' + description: >- + Attributes related to instance pools running on Azure. If not + specified at pool creation, a set of default values will be used. + custom_tags: + type: object + description: >- + Additional tags for pool resources. Databricks will tag all pool + resources (e.g., AWS instances and EBS volumes) with these tags in + addition to `default_tags`. Notes: - Currently, Databricks allows at + most 45 custom tags + default_tags: + type: object + description: >- + Tags that are added by Databricks regardless of any ``custom_tags``, + including: - Vendor: Databricks - InstancePoolCreator: + - InstancePoolName: - + InstancePoolId: + disk_spec: + $ref: '#/components/schemas/DiskSpec' + description: >- + Defines the specification of the disks that will be attached to all + spark containers. + enable_elastic_disk: + type: boolean + description: >- + Autoscaling Local Storage: when enabled, this instances in this pool + will dynamically acquire additional disk space when its Spark + workers are running low on disk space. In AWS, this feature requires + specific AWS permissions to function correctly - refer to the User + Guide for more details. + gcp_attributes: + $ref: '#/components/schemas/InstancePoolGcpAttributes' + description: >- + Attributes related to instance pools running on Google Cloud + Platform. If not specified at pool creation, a set of default values + will be used. + idle_instance_autotermination_minutes: + type: integer + description: >- + Automatically terminates the extra instances in the pool cache after + they are inactive for this time in minutes if min_idle_instances + requirement is already met. If not set, the extra pool instances + will be automatically terminated after a default timeout. If + specified, the threshold must be between 0 and 10000 minutes. Users + can also set this value to 0 to instantly remove idle instances from + the cache if min cache size could still hold. + instance_pool_name: + type: string + description: >- + Pool name requested by the user. Pool name must be unique. Length + must be between 1 and 100 characters. + max_capacity: + type: integer + description: >- + Maximum number of outstanding instances to keep in the pool, + including both instances used by clusters and idle instances. + Clusters that require further instance provisioning will fail during + upsize requests. + min_idle_instances: + type: integer + description: Minimum number of idle instances to keep in the instance pool + node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for the pool. + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources available + to each of the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved + by using the :method:clusters/listNodeTypes API call. + preloaded_docker_images: + type: array + items: + $ref: '#/components/schemas/DockerImage' + description: Custom Docker Image BYOC + preloaded_spark_versions: + type: array + items: + type: string + description: >- + A list containing at most one preloaded Spark image version for the + pool. Pool-backed clusters started with the preloaded Spark version + will start faster. A list of available Spark versions can be + retrieved by using the :method:clusters/sparkVersions API call. + remote_disk_throughput: + type: integer + description: >- + If set, what the configurable throughput (in Mb/s) for the remote + disk is. Currently only supported for GCP HYPERDISK_BALANCED types. + state: + $ref: '#/components/schemas/InstancePoolState' + description: Current state of the instance pool. + stats: + $ref: '#/components/schemas/InstancePoolStats' + description: Usage statistics about the instance pool. + status: + $ref: '#/components/schemas/InstancePoolStatus' + description: Status of failed pending instances in the pool. + total_initial_remote_disk_size: + type: integer + description: >- + If set, what the total initial volume size (in GB) of the remote + disks should be. Currently only supported for GCP HYPERDISK_BALANCED + types. + required: + - instance_pool_id + GetInstancePoolPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/InstancePoolPermissionsDescription' + GetSparkVersionsResponse: + type: object + properties: + versions: + type: array + items: + $ref: '#/components/schemas/SparkVersion' + GlobalInitScriptDetails: + type: object + properties: + created_at: + type: integer + created_by: + type: string + description: The username of the user who created the script. + enabled: + type: boolean + description: >- + Specifies whether the script is enabled. The script runs only if + enabled. + name: + type: string + description: The name of the script + position: + type: integer + description: >- + The position of a script, where 0 represents the first script to + run, 1 is the second script to run, in ascending order. + script_id: + type: string + description: The global init script ID. + updated_at: + type: integer + description: >- + Time when the script was updated, represented as a Unix timestamp in + milliseconds. + updated_by: + type: string + description: The username of the user who last updated the script + GlobalInitScriptDetailsWithContent: + type: object + properties: + created_at: + type: integer + created_by: + type: string + description: The username of the user who created the script. + enabled: + type: boolean + description: >- + Specifies whether the script is enabled. The script runs only if + enabled. + name: + type: string + description: The name of the script + position: + type: integer + description: >- + The position of a script, where 0 represents the first script to + run, 1 is the second script to run, in ascending order. + script: + type: string + description: The Base64-encoded content of the script. + script_id: + type: string + description: The global init script ID. + updated_at: + type: integer + description: >- + Time when the script was updated, represented as a Unix timestamp in + milliseconds. + updated_by: + type: string + description: The username of the user who last updated the script + InitScriptEventDetails: + type: object + properties: + cluster: + type: array + items: + $ref: '#/components/schemas/InitScriptInfoAndExecutionDetails' + global: + type: array + items: + $ref: '#/components/schemas/InitScriptInfoAndExecutionDetails' + description: The global init scripts associated with this cluster event. + reported_for_node: + type: string + description: >- + The private ip of the node we are reporting init script execution + details for (we will select the execution details from only one node + rather than reporting the execution details from every node to keep + these event details small) This should only be defined for the + INIT_SCRIPTS_FINISHED event + InitScriptInfo: + type: object + properties: + abfss: + $ref: '#/components/schemas/Adlsgen2Info' + description: >- + destination needs to be provided, e.g. + `abfss://@.dfs.core.windows.net/` + dbfs: + $ref: '#/components/schemas/DbfsStorageInfo' + description: >- + destination needs to be provided. e.g. `{ "dbfs": { "destination" : + "dbfs:/home/cluster_log" } }` + file: + $ref: '#/components/schemas/LocalFileInfo' + description: >- + destination needs to be provided, e.g. `{ "file": { "destination": + "file:/my/local/file.sh" } }` + gcs: + $ref: '#/components/schemas/GcsStorageInfo' + description: >- + destination needs to be provided, e.g. `{ "gcs": { "destination": + "gs://my-bucket/file.sh" } }` + s3: + $ref: '#/components/schemas/S3StorageInfo' + description: >- + destination and either the region or endpoint need to be provided. + e.g. `{ \"s3\": { \"destination\": + \"s3://cluster_log_bucket/prefix\", \"region\": \"us-west-2\" } }` + Cluster iam role is used to access s3, please make sure the cluster + iam role in `instance_profile_arn` has permission to write data to + the s3 destination. + volumes: + $ref: '#/components/schemas/VolumesStorageInfo' + description: >- + destination needs to be provided. e.g. `{ \"volumes\" : { + \"destination\" : \"/Volumes/my-init.sh\" } }` + workspace: + $ref: '#/components/schemas/WorkspaceStorageInfo' + description: >- + destination needs to be provided, e.g. `{ "workspace": { + "destination": "/cluster-init-scripts/setup-datadog.sh" } }` + description: 'Config for an individual init script Next ID: 11' + InitScriptInfoAndExecutionDetails: + type: object + properties: + abfss: + $ref: '#/components/schemas/Adlsgen2Info' + dbfs: + $ref: '#/components/schemas/DbfsStorageInfo' + description: >- + destination needs to be provided. e.g. `{ "dbfs": { "destination" : + "dbfs:/home/cluster_log" } }` + error_message: + type: string + description: >- + Additional details regarding errors (such as a file not found + message if the status is FAILED_FETCH). This field should only be + used to provide *additional* information to the status field, not + duplicate it. + execution_duration_seconds: + type: integer + description: The number duration of the script execution in seconds + file: + $ref: '#/components/schemas/LocalFileInfo' + description: >- + destination needs to be provided, e.g. `{ "file": { "destination": + "file:/my/local/file.sh" } }` + gcs: + $ref: '#/components/schemas/GcsStorageInfo' + description: >- + destination needs to be provided, e.g. `{ "gcs": { "destination": + "gs://my-bucket/file.sh" } }` + s3: + $ref: '#/components/schemas/S3StorageInfo' + description: >- + destination and either the region or endpoint need to be provided. + e.g. `{ \"s3\": { \"destination\": + \"s3://cluster_log_bucket/prefix\", \"region\": \"us-west-2\" } }` + Cluster iam role is used to access s3, please make sure the cluster + iam role in `instance_profile_arn` has permission to write data to + the s3 destination. + status: + $ref: >- + #/components/schemas/InitScriptExecutionDetailsInitScriptExecutionStatus + description: The current status of the script + stderr: + type: string + description: >- + The stderr output from the init script execution. Only populated + when init scripts debug is enabled and script execution fails. + volumes: + $ref: '#/components/schemas/VolumesStorageInfo' + description: >- + destination needs to be provided. e.g. `{ \"volumes\" : { + \"destination\" : \"/Volumes/my-init.sh\" } }` + workspace: + $ref: '#/components/schemas/WorkspaceStorageInfo' + description: >- + destination needs to be provided, e.g. `{ "workspace": { + "destination": "/cluster-init-scripts/setup-datadog.sh" } }` + InstallLibrariesResponse: + type: object + properties: {} + InstancePoolAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/InstancePoolPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + InstancePoolAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/InstancePoolPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + InstancePoolAndStats: + type: object + properties: + aws_attributes: + $ref: '#/components/schemas/InstancePoolAwsAttributes' + azure_attributes: + $ref: '#/components/schemas/InstancePoolAzureAttributes' + description: >- + Attributes related to instance pools running on Azure. If not + specified at pool creation, a set of default values will be used. + custom_tags: + type: object + description: >- + Additional tags for pool resources. Databricks will tag all pool + resources (e.g., AWS instances and EBS volumes) with these tags in + addition to `default_tags`. Notes: - Currently, Databricks allows at + most 45 custom tags + default_tags: + type: object + description: >- + Tags that are added by Databricks regardless of any ``custom_tags``, + including: - Vendor: Databricks - InstancePoolCreator: + - InstancePoolName: - + InstancePoolId: + disk_spec: + $ref: '#/components/schemas/DiskSpec' + description: >- + Defines the specification of the disks that will be attached to all + spark containers. + enable_elastic_disk: + type: boolean + description: >- + Autoscaling Local Storage: when enabled, this instances in this pool + will dynamically acquire additional disk space when its Spark + workers are running low on disk space. In AWS, this feature requires + specific AWS permissions to function correctly - refer to the User + Guide for more details. + gcp_attributes: + $ref: '#/components/schemas/InstancePoolGcpAttributes' + description: >- + Attributes related to instance pools running on Google Cloud + Platform. If not specified at pool creation, a set of default values + will be used. + idle_instance_autotermination_minutes: + type: integer + description: >- + Automatically terminates the extra instances in the pool cache after + they are inactive for this time in minutes if min_idle_instances + requirement is already met. If not set, the extra pool instances + will be automatically terminated after a default timeout. If + specified, the threshold must be between 0 and 10000 minutes. Users + can also set this value to 0 to instantly remove idle instances from + the cache if min cache size could still hold. + instance_pool_id: + type: string + description: Canonical unique identifier for the pool. + instance_pool_name: + type: string + description: >- + Pool name requested by the user. Pool name must be unique. Length + must be between 1 and 100 characters. + max_capacity: + type: integer + description: >- + Maximum number of outstanding instances to keep in the pool, + including both instances used by clusters and idle instances. + Clusters that require further instance provisioning will fail during + upsize requests. + min_idle_instances: + type: integer + description: Minimum number of idle instances to keep in the instance pool + node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for the pool. + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources available + to each of the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved + by using the :method:clusters/listNodeTypes API call. + preloaded_docker_images: + type: array + items: + $ref: '#/components/schemas/DockerImage' + description: Custom Docker Image BYOC + preloaded_spark_versions: + type: array + items: + type: string + description: >- + A list containing at most one preloaded Spark image version for the + pool. Pool-backed clusters started with the preloaded Spark version + will start faster. A list of available Spark versions can be + retrieved by using the :method:clusters/sparkVersions API call. + remote_disk_throughput: + type: integer + description: >- + If set, what the configurable throughput (in Mb/s) for the remote + disk is. Currently only supported for GCP HYPERDISK_BALANCED types. + state: + $ref: '#/components/schemas/InstancePoolState' + description: Current state of the instance pool. + stats: + $ref: '#/components/schemas/InstancePoolStats' + description: Usage statistics about the instance pool. + status: + $ref: '#/components/schemas/InstancePoolStatus' + description: Status of failed pending instances in the pool. + total_initial_remote_disk_size: + type: integer + description: >- + If set, what the total initial volume size (in GB) of the remote + disks should be. Currently only supported for GCP HYPERDISK_BALANCED + types. + InstancePoolAwsAttributes: + type: object + properties: + availability: + $ref: '#/components/schemas/InstancePoolAwsAttributesAvailability' + description: Availability type used for the spot nodes. + instance_profile_arn: + type: string + description: >- + All AWS instances belonging to the instance pool will have this + instance profile. If omitted, instances will initially be launched + with the workspace's default instance profile. If defined, clusters + that use the pool will inherit the instance profile, and must not + specify their own instance profile on cluster creation or update. If + the pool does not specify an instance profile, clusters using the + pool may specify any instance profile. The instance profile must + have previously been added to the Databricks environment by an + account administrator. This feature may only be available to certain + customer plans. + spot_bid_price_percent: + type: integer + description: >- + Calculates the bid price for AWS spot instances, as a percentage of + the corresponding instance type's on-demand price. For example, if + this field is set to 50, and the cluster needs a new `r3.xlarge` + spot instance, then the bid price is half of the price of on-demand + `r3.xlarge` instances. Similarly, if this field is set to 200, the + bid price is twice the price of on-demand `r3.xlarge` instances. If + not specified, the default value is 100. When spot instances are + requested for this cluster, only spot instances whose bid price + percentage matches this field will be considered. Note that, for + safety, we enforce this field to be no more than 10000. + zone_id: + type: string + description: >- + Identifier for the availability zone/datacenter in which the cluster + resides. This string will be of a form like "us-west-2a". The + provided availability zone must be in the same region as the + Databricks deployment. For example, "us-west-2a" is not a valid zone + id if the Databricks deployment resides in the "us-east-1" region. + This is an optional field at cluster creation, and if not specified, + a default zone will be used. The list of available zones as well as + the default value can be found by using the `List Zones` method. + description: >- + Attributes set during instance pool creation which are related to Amazon + Web Services. + InstancePoolAzureAttributes: + type: object + properties: + availability: + $ref: '#/components/schemas/InstancePoolAzureAttributesAvailability' + description: Availability type used for the spot nodes. + spot_bid_max_price: + type: number + description: >- + With variable pricing, you have option to set a max price, in US + dollars (USD) For example, the value 2 would be a max price of $2.00 + USD per hour. If you set the max price to be -1, the VM won't be + evicted based on price. The price for the VM will be the current + price for spot or the price for a standard VM, which ever is less, + as long as there is capacity and quota available. + description: Attributes set during instance pool creation which are related to Azure. + InstancePoolGcpAttributes: + type: object + properties: + gcp_availability: + $ref: '#/components/schemas/GcpAvailability' + local_ssd_count: + type: integer + description: >- + If provided, each node in the instance pool will have this number of + local SSDs attached. Each local SSD is 375GB in size. Refer to [GCP + documentation] for the supported number of local SSDs for each + instance type. [GCP documentation]: + https://cloud.google.com/compute/docs/disks/local-ssd#choose_number_local_ssds + zone_id: + type: string + description: >- + Identifier for the availability zone/datacenter in which the cluster + resides. This string will be of a form like "us-west1-a". The + provided availability zone must be in the same region as the + Databricks workspace. For example, "us-west1-a" is not a valid zone + id if the Databricks workspace resides in the "us-east1" region. + This is an optional field at instance pool creation, and if not + specified, a default zone will be used. This field can be one of the + following: - "HA" => High availability, spread nodes across + availability zones for a Databricks deployment region - A GCP + availability zone => Pick One of the available zones for (machine + type + region) from + https://cloud.google.com/compute/docs/regions-zones (e.g. + "us-west1-a"). If empty, Databricks picks an availability zone to + schedule the cluster on. + description: Attributes set during instance pool creation which are related to GCP. + InstancePoolPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/InstancePoolPermissionLevel' + InstancePoolPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/InstancePoolAccessControlResponse' + object_id: + type: string + object_type: + type: string + InstancePoolPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/InstancePoolPermissionLevel' + InstancePoolStats: + type: object + properties: + idle_count: + type: integer + pending_idle_count: + type: integer + description: >- + Number of pending instances in the pool that are NOT part of a + cluster. + pending_used_count: + type: integer + description: Number of pending instances in the pool that are part of a cluster. + used_count: + type: integer + description: Number of active instances in the pool that are part of a cluster. + InstancePoolStatus: + type: object + properties: + pending_instance_errors: + type: array + items: + $ref: '#/components/schemas/PendingInstanceError' + InstanceProfile: + type: object + properties: + instance_profile_arn: + type: string + iam_role_arn: + type: string + description: >- + The AWS IAM role ARN of the role associated with the instance + profile. This field is required if your role name and instance + profile name do not match and you want to use the instance profile + with [Databricks SQL Serverless]. Otherwise, this field is optional. + [Databricks SQL Serverless]: + https://docs.databricks.com/sql/admin/serverless.html + is_meta_instance_profile: + type: boolean + description: >- + Boolean flag indicating whether the instance profile should only be + used in credential passthrough scenarios. If true, it means the + instance profile contains an meta IAM role which could assume a wide + range of roles. Therefore it should always be used with + authorization. This field is optional, the default value is `false`. + required: + - instance_profile_arn + Library: + type: object + properties: + cran: + $ref: '#/components/schemas/RCranLibrary' + egg: + type: string + description: >- + Deprecated. URI of the egg library to install. Installing Python egg + files is deprecated and is not supported in Databricks Runtime 14.0 + and above. + jar: + type: string + description: >- + URI of the JAR library to install. Supported URIs include Workspace + paths, Unity Catalog Volumes paths, and S3 URIs. For example: `{ + "jar": "/Workspace/path/to/library.jar" }`, `{ "jar" : + "/Volumes/path/to/library.jar" }` or `{ "jar": + "s3://my-bucket/library.jar" }`. If S3 is used, please make sure the + cluster has read access on the library. You may need to launch the + cluster with an IAM role to access the S3 URI. + maven: + $ref: '#/components/schemas/MavenLibrary' + description: >- + Specification of a maven library to be installed. For example: `{ + "coordinates": "org.jsoup:jsoup:1.7.2" }` + pypi: + $ref: '#/components/schemas/PythonPyPiLibrary' + description: >- + Specification of a PyPi library to be installed. For example: `{ + "package": "simplejson" }` + requirements: + type: string + description: >- + URI of the requirements.txt file to install. Only Workspace paths + and Unity Catalog Volumes paths are supported. For example: `{ + "requirements": "/Workspace/path/to/requirements.txt" }` or `{ + "requirements" : "/Volumes/path/to/requirements.txt" }` + whl: + type: string + description: >- + URI of the wheel library to install. Supported URIs include + Workspace paths, Unity Catalog Volumes paths, and S3 URIs. For + example: `{ "whl": "/Workspace/path/to/library.whl" }`, `{ "whl" : + "/Volumes/path/to/library.whl" }` or `{ "whl": + "s3://my-bucket/library.whl" }`. If S3 is used, please make sure the + cluster has read access on the library. You may need to launch the + cluster with an IAM role to access the S3 URI. + LibraryFullStatus: + type: object + properties: + is_library_for_all_clusters: + type: boolean + description: >- + Whether the library was set to be installed on all clusters via the + libraries UI. + library: + $ref: '#/components/schemas/Library' + description: Unique identifier for the library. + messages: + type: array + items: + type: string + description: >- + All the info and warning messages that have occurred so far for this + library. + status: + $ref: '#/components/schemas/LibraryInstallStatus' + description: Status of installing the library on the cluster. + description: The status of the library on a specific cluster. + ListAllClusterLibraryStatusesResponse: + type: object + properties: + statuses: + type: array + items: + $ref: '#/components/schemas/ClusterLibraryStatuses' + ListAvailableZonesResponse: + type: object + properties: + default_zone: + type: string + zones: + type: array + items: + type: string + description: The list of available zones (e.g., ['us-west-2c', 'us-east-2']). + ListClusterCompliancesResponse: + type: object + properties: + clusters: + type: array + items: + $ref: '#/components/schemas/ClusterCompliance' + next_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the next page + of results. If the value is "", it means no further results for the + request. + prev_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the previous + page of results. If the value is "", it means no further results for + the request. + ListClustersFilterBy: + type: object + properties: + cluster_sources: + type: array + items: + $ref: '#/components/schemas/ClusterSource' + cluster_states: + type: array + items: + $ref: '#/components/schemas/State' + description: The current state of the clusters. + is_pinned: + type: boolean + description: Whether the clusters are pinned or not. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + ListClustersResponse: + type: object + properties: + clusters: + type: array + items: + $ref: '#/components/schemas/ClusterDetails' + next_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the next page + of results. If the value is "", it means no further results for the + request. + prev_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the previous + page of results. If the value is "", it means no further results for + the request. + ListClustersSortBy: + type: object + properties: + direction: + $ref: '#/components/schemas/ListClustersSortByDirection' + field: + $ref: '#/components/schemas/ListClustersSortByField' + description: >- + The sorting criteria. By default, clusters are sorted by 3 columns + from highest to lowest precedence: cluster state, pinned or + unpinned, then cluster name. + ListGlobalInitScriptsResponse: + type: object + properties: + scripts: + type: array + items: + $ref: '#/components/schemas/GlobalInitScriptDetails' + ListInstancePools: + type: object + properties: + instance_pools: + type: array + items: + $ref: '#/components/schemas/InstancePoolAndStats' + ListInstanceProfilesResponse: + type: object + properties: + instance_profiles: + type: array + items: + $ref: '#/components/schemas/InstanceProfile' + ListNodeTypesResponse: + type: object + properties: + node_types: + type: array + items: + $ref: '#/components/schemas/NodeType' + ListPoliciesResponse: + type: object + properties: + policies: + type: array + items: + $ref: '#/components/schemas/Policy' + ListPolicyFamiliesResponse: + type: object + properties: + next_page_token: + type: string + policy_families: + type: array + items: + $ref: '#/components/schemas/PolicyFamily' + description: List of policy families. + LocalFileInfo: + type: object + properties: + destination: + type: string + required: + - destination + LogAnalyticsInfo: + type: object + properties: + log_analytics_primary_key: + type: string + log_analytics_workspace_id: + type: string + LogSyncStatus: + type: object + properties: + last_attempted: + type: integer + description: >- + The timestamp of last attempt. If the last attempt fails, + `last_exception` will contain the exception in the last attempt. + last_exception: + type: string + description: >- + The exception thrown in the last attempt, it would be null (omitted + in the response) if there is no exception in last attempted. + description: The log delivery status + MavenLibrary: + type: object + properties: + coordinates: + type: string + exclusions: + type: array + items: + type: string + description: >- + List of dependences to exclude. For example: `["slf4j:slf4j", + "*:hadoop-client"]`. Maven dependency exclusions: + https://maven.apache.org/guides/introduction/introduction-to-optional-and-excludes-dependencies.html. + repo: + type: string + description: >- + Maven repo to install the Maven package from. If omitted, both Maven + Central Repository and Spark Packages are searched. + required: + - coordinates + NodeInstanceType: + type: object + properties: + instance_type_id: + type: string + description: Unique identifier across instance types + local_disk_size_gb: + type: integer + description: >- + Size of the individual local disks attached to this instance (i.e. + per local disk). + local_disks: + type: integer + description: Number of local disks that are present on this instance. + local_nvme_disk_size_gb: + type: integer + description: >- + Size of the individual local nvme disks attached to this instance + (i.e. per local disk). + local_nvme_disks: + type: integer + description: Number of local nvme disks that are present on this instance. + required: + - instance_type_id + description: >- + This structure embodies the machine type that hosts spark containers + Note: this should be an + internal data structure for now It is defined in proto in case we want to send it over the wire + in the future (which is likely) + NodeType: + type: object + properties: + node_type_id: + type: string + description: Unique identifier for this node type. + memory_mb: + type: integer + description: Memory (in MB) available for this node type. + num_cores: + type: number + description: >- + Number of CPU cores available for this node type. Note that this can + be fractional, e.g., 2.5 cores, if the the number of cores on a + machine instance is not divisible by the number of Spark nodes on + that machine. + description: + type: string + description: >- + A string description associated with this node type, e.g., + "r3.xlarge". + instance_type_id: + type: string + description: >- + An identifier for the type of hardware that this node runs on, e.g., + "r3.2xlarge" in AWS. + category: + type: string + description: >- + A descriptive category for this node type. Examples include "Memory + Optimized" and "Compute Optimized". + display_order: + type: integer + description: >- + An optional hint at the display order of node types in the UI. + Within a node type category, lowest numbers come first. + is_deprecated: + type: boolean + description: >- + Whether the node type is deprecated. Non-deprecated node types offer + greater performance. + is_encrypted_in_transit: + type: boolean + description: >- + AWS specific, whether this instance supports encryption in transit, + used for hipaa and pci workloads. + is_graviton: + type: boolean + description: Whether this is an Arm-based instance. + is_hidden: + type: boolean + description: Whether this node is hidden from presentation in the UI. + is_io_cache_enabled: + type: boolean + description: Whether this node comes with IO cache enabled by default. + node_info: + $ref: '#/components/schemas/CloudProviderNodeInfo' + description: A collection of node type info reported by the cloud provider + node_instance_type: + $ref: '#/components/schemas/NodeInstanceType' + description: The NodeInstanceType object corresponding to instance_type_id + num_gpus: + type: integer + description: Number of GPUs available for this node type. + photon_driver_capable: + type: boolean + photon_worker_capable: + type: boolean + support_cluster_tags: + type: boolean + description: Whether this node type support cluster tags. + support_ebs_volumes: + type: boolean + description: >- + Whether this node type support EBS volumes. EBS volumes is disabled + for node types that we could place multiple corresponding containers + on the same hosting instance. + support_port_forwarding: + type: boolean + description: Whether this node type supports port forwarding. + required: + - node_type_id + - memory_mb + - num_cores + - description + - instance_type_id + - category + description: >- + A description of a Spark node type including both the dimensions of the + node and the instance + type on which it will be hosted. + NodeTypeFlexibility: + type: object + properties: + alternate_node_type_ids: + type: array + items: + type: string + description: >- + A list of node type IDs to use as fallbacks when the primary node + type is unavailable. + description: >- + Configuration for flexible node types, allowing fallback to alternate + node types during cluster + launch and upscale. + PendingInstanceError: + type: object + properties: + instance_id: + type: string + message: + type: string + description: Error message of a failed pending instances + PermanentDeleteClusterResponse: + type: object + properties: {} + PinClusterResponse: + type: object + properties: {} + Policy: + type: object + properties: + created_at_timestamp: + type: integer + description: >- + Creation time. The timestamp (in millisecond) when this Cluster + Policy was created. + creator_user_name: + type: string + description: >- + Creator user name. The field won't be included in the response if + the user has already been deleted. + definition: + type: string + description: >- + Policy definition document expressed in [Databricks Cluster Policy + Definition Language]. [Databricks Cluster Policy Definition + Language]: + https://docs.databricks.com/administration-guide/clusters/policy-definition.html + description: + type: string + description: Additional human-readable description of the cluster policy. + is_default: + type: boolean + description: >- + If true, policy is a default policy created and managed by + Databricks. Default policies cannot be deleted, and their policy + families cannot be changed. + libraries: + type: array + items: + $ref: '#/components/schemas/Library' + description: >- + A list of libraries to be installed on the next cluster restart that + uses this policy. The maximum number of libraries is 500. + max_clusters_per_user: + type: integer + description: >- + Max number of clusters per user that can be active using this + policy. If not present, there is no max limit. + name: + type: string + description: >- + Cluster Policy name requested by the user. This has to be unique. + Length must be between 1 and 100 characters. + policy_family_definition_overrides: + type: string + description: >- + Policy definition JSON document expressed in [Databricks Policy + Definition Language]. The JSON document must be passed as a string + and cannot be embedded in the requests. You can use this to + customize the policy definition inherited from the policy family. + Policy rules specified here are merged into the inherited policy + definition. [Databricks Policy Definition Language]: + https://docs.databricks.com/administration-guide/clusters/policy-definition.html + policy_family_id: + type: string + description: >- + ID of the policy family. The cluster policy's policy definition + inherits the policy family's policy definition. Cannot be used with + `definition`. Use `policy_family_definition_overrides` instead to + customize the policy definition. + policy_id: + type: string + description: Canonical unique identifier for the Cluster Policy. + description: Describes a Cluster Policy entity. + PolicyFamily: + type: object + properties: + definition: + type: string + description: + type: string + description: Human-readable description of the purpose of the policy family. + name: + type: string + description: Name of the policy family. + policy_family_id: + type: string + description: Unique identifier for the policy family. + PythonPyPiLibrary: + type: object + properties: + package: + type: string + repo: + type: string + description: >- + The repository where the package can be found. If not specified, the + default pip index is used. + required: + - package + RCranLibrary: + type: object + properties: + package: + type: string + repo: + type: string + description: >- + The repository where the package can be found. If not specified, the + default CRAN repo is used. + required: + - package + RemoveResponse: + type: object + properties: {} + ResizeClusterResponse: + type: object + properties: {} + RestartClusterResponse: + type: object + properties: {} + Results: + type: object + properties: + cause: + type: string + data: + type: object + fileName: + type: string + description: >- + The image data in one of the following formats: 1. A Data URL with + base64-encoded image data: `data:image/{type};base64,{base64-data}`. + Example: `data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAUA...` 2. + A FileStore file path for large images: `/plots/{filename}.png`. + Example: `/plots/b6a7ad70-fb2c-4353-8aed-3f1e015174a4.png` + fileNames: + type: array + items: + type: string + description: >- + List of image data for multiple images. Each element follows the + same format as file_name. + isJsonSchema: + type: boolean + description: >- + true if a JSON schema is returned instead of a string representation + of the Hive type. + pos: + type: integer + description: internal field used by SDK + resultType: + $ref: '#/components/schemas/ResultType' + schema: + type: array + items: + type: object + description: The table schema + summary: + type: string + description: The summary of the error + truncated: + type: boolean + description: true if partial results are returned. + S3StorageInfo: + type: object + properties: + destination: + type: string + description: >- + S3 destination, e.g. `s3://my-bucket/some-prefix` Note that logs + will be delivered using cluster iam role, please make sure you set + cluster iam role and the role has write access to the destination. + Please also note that you cannot use AWS keys to deliver logs. + canned_acl: + type: string + description: >- + (Optional) Set canned access control list for the logs, e.g. + `bucket-owner-full-control`. If `canned_cal` is set, please make + sure the cluster iam role has `s3:PutObjectAcl` permission on the + destination bucket and prefix. The full list of possible canned acl + can be found at + http://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl. + Please also note that by default only the object owner gets full + controls. If you are using cross account role for writing data, you + may want to set `bucket-owner-full-control` to make bucket owner + able to read the logs. + enable_encryption: + type: boolean + description: >- + (Optional) Flag to enable server side encryption, `false` by + default. + encryption_type: + type: string + description: >- + (Optional) The encryption type, it could be `sse-s3` or `sse-kms`. + It will be used only when encryption is enabled and the default type + is `sse-s3`. + endpoint: + type: string + description: >- + S3 endpoint, e.g. `https://s3-us-west-2.amazonaws.com`. Either + region or endpoint needs to be set. If both are set, endpoint will + be used. + kms_key: + type: string + description: >- + (Optional) Kms key which will be used if encryption is enabled and + encryption type is set to `sse-kms`. + region: + type: string + description: >- + S3 region, e.g. `us-west-2`. Either region or endpoint needs to be + set. If both are set, endpoint will be used. + required: + - destination + description: A storage location in Amazon S3 + SparkNode: + type: object + properties: + host_private_ip: + type: string + description: The private IP address of the host instance. + instance_id: + type: string + description: >- + Globally unique identifier for the host instance from the cloud + provider. + node_aws_attributes: + $ref: '#/components/schemas/SparkNodeAwsAttributes' + description: Attributes specific to AWS for a Spark node. + node_id: + type: string + description: Globally unique identifier for this node. + private_ip: + type: string + description: >- + Private IP address (typically a 10.x.x.x address) of the Spark node. + Note that this is different from the private IP address of the host + instance. + public_dns: + type: string + description: >- + Public DNS address of this node. This address can be used to access + the Spark JDBC server on the driver node. To communicate with the + JDBC server, traffic must be manually authorized by adding security + group rules to the "worker-unmanaged" security group via the AWS + console. + start_timestamp: + type: integer + description: The timestamp (in millisecond) when the Spark node is launched. + description: Describes a specific Spark driver or executor. + SparkNodeAwsAttributes: + type: object + properties: + is_spot: + type: boolean + description: Whether this node is on an Amazon spot instance. + description: Attributes specific to AWS for a Spark node. + SparkVersion: + type: object + properties: + key: + type: string + name: + type: string + description: A descriptive name for this Spark version, for example "Spark 2.1". + StartClusterResponse: + type: object + properties: {} + TerminationReason: + type: object + properties: + code: + $ref: '#/components/schemas/TerminationReasonCode' + parameters: + type: object + description: >- + list of parameters that provide additional information about why the + cluster was terminated + type: + $ref: '#/components/schemas/TerminationReasonType' + description: type of the termination + UninstallLibrariesResponse: + type: object + properties: {} + UnpinClusterResponse: + type: object + properties: {} + UpdateClusterResource: + type: object + properties: + autoscale: + $ref: '#/components/schemas/AutoScale' + autotermination_minutes: + type: integer + description: >- + Automatically terminates the cluster after it is inactive for this + time in minutes. If not set, this cluster will not be automatically + terminated. If specified, the threshold must be between 10 and 10000 + minutes. Users can also set this value to 0 to explicitly disable + automatic termination. + aws_attributes: + $ref: '#/components/schemas/AwsAttributes' + description: >- + Attributes related to clusters running on Amazon Web Services. If + not specified at cluster creation, a set of default values will be + used. + azure_attributes: + $ref: '#/components/schemas/AzureAttributes' + description: >- + Attributes related to clusters running on Microsoft Azure. If not + specified at cluster creation, a set of default values will be used. + cluster_log_conf: + $ref: '#/components/schemas/ClusterLogConf' + description: >- + The configuration for delivering spark logs to a long-term storage + destination. Three kinds of destinations (DBFS, S3 and Unity Catalog + volumes) are supported. Only one destination can be specified for + one cluster. If the conf is given, the logs will be delivered to the + destination every `5 mins`. The destination of driver logs is + `$destination/$clusterId/driver`, while the destination of executor + logs is `$destination/$clusterId/executor`. + cluster_name: + type: string + description: >- + Cluster name requested by the user. This doesn't have to be unique. + If not specified at creation, the cluster name will be an empty + string. For job clusters, the cluster name is automatically set + based on the job and job run IDs. + custom_tags: + type: object + description: >- + Additional tags for cluster resources. Databricks will tag all + cluster resources (e.g., AWS instances and EBS volumes) with these + tags in addition to `default_tags`. Notes: - Currently, Databricks + allows at most 45 custom tags - Clusters can only reuse cloud + resources if the resources' tags are a subset of the cluster tags + data_security_mode: + $ref: '#/components/schemas/DataSecurityMode' + docker_image: + $ref: '#/components/schemas/DockerImage' + description: Custom docker image BYOC + driver_instance_pool_id: + type: string + description: >- + The optional ID of the instance pool for the driver of the cluster + belongs. The pool cluster uses the instance pool with id + (instance_pool_id) if the driver pool is not assigned. + driver_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for the driver node. + driver_node_type_id: + type: string + description: >- + The node type of the Spark driver. Note that this field is optional; + if unset, the driver node type will be set as the same value as + `node_type_id` defined above. This field, along with node_type_id, + should not be set if virtual_cluster_size is set. If both + driver_node_type_id, node_type_id, and virtual_cluster_size are + specified, driver_node_type_id and node_type_id take precedence. + enable_elastic_disk: + type: boolean + description: >- + Autoscaling Local Storage: when enabled, this cluster will + dynamically acquire additional disk space when its Spark workers are + running low on disk space. + enable_local_disk_encryption: + type: boolean + description: Whether to enable LUKS on cluster VMs' local disks + gcp_attributes: + $ref: '#/components/schemas/GcpAttributes' + description: >- + Attributes related to clusters running on Google Cloud Platform. If + not specified at cluster creation, a set of default values will be + used. + init_scripts: + type: array + items: + $ref: '#/components/schemas/InitScriptInfo' + description: >- + The configuration for storing init scripts. Any number of + destinations can be specified. The scripts are executed sequentially + in the order provided. If `cluster_log_conf` is specified, init + script logs are sent to `//init_scripts`. + instance_pool_id: + type: string + description: The optional ID of the instance pool to which the cluster belongs. + is_single_node: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. When set + to true, Databricks will automatically set single node related + `custom_tags`, `spark_conf`, and `num_workers` + kind: + $ref: '#/components/schemas/Kind' + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources available + to each of the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved + by using the :method:clusters/listNodeTypes API call. + num_workers: + type: integer + description: >- + Number of worker nodes that this cluster should have. A cluster has + one Spark Driver and `num_workers` Executors for a total of + `num_workers` + 1 Spark nodes. Note: When reading the properties of + a cluster, this field reflects the desired number of workers rather + than the actual current number of workers. For instance, if a + cluster is resized from 5 to 10 workers, this field will immediately + be updated to reflect the target size of 10 workers, whereas the + workers listed in `spark_info` will gradually increase from 5 to 10 + as the new nodes are provisioned. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + remote_disk_throughput: + type: integer + description: >- + If set, what the configurable throughput (in Mb/s) for the remote + disk is. Currently only supported for GCP HYPERDISK_BALANCED disks. + runtime_engine: + $ref: '#/components/schemas/RuntimeEngine' + description: >- + Determines the cluster's runtime engine, either standard or Photon. + This field is not compatible with legacy `spark_version` values that + contain `-photon-`. Remove `-photon-` from the `spark_version` and + set `runtime_engine` to `PHOTON`. If left unspecified, the runtime + engine defaults to standard unless the spark_version contains + -photon-, in which case Photon will be used. + single_user_name: + type: string + description: Single user name if data_security_mode is `SINGLE_USER` + spark_conf: + type: object + description: >- + An object containing a set of optional, user-specified Spark + configuration key-value pairs. Users can also pass in a string of + extra JVM options to the driver and the executors via + `spark.driver.extraJavaOptions` and + `spark.executor.extraJavaOptions` respectively. + spark_env_vars: + type: object + description: >- + An object containing a set of optional, user-specified environment + variable key-value pairs. Please note that key-value pair of the + form (X,Y) will be exported as is (i.e., `export X='Y'`) while + launching the driver and workers. In order to specify an additional + set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This + ensures that all default databricks managed environmental variables + are included as well. Example Spark environment variables: + `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": + "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": + "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}` + spark_version: + type: string + description: >- + The Spark version of the cluster, e.g. `3.3.x-scala2.11`. A list of + available Spark versions can be retrieved by using the + :method:clusters/sparkVersions API call. + ssh_public_keys: + type: array + items: + type: string + description: >- + SSH public key contents that will be added to each Spark node in + this cluster. The corresponding private keys can be used to login + with the user name `ubuntu` on port `2200`. Up to 10 keys can be + specified. + total_initial_remote_disk_size: + type: integer + description: >- + If set, what the total initial volume size (in GB) of the remote + disks should be. Currently only supported for GCP HYPERDISK_BALANCED + disks. + use_ml_runtime: + type: boolean + description: >- + This field can only be used when `kind = CLASSIC_PREVIEW`. + `effective_spark_version` is determined by `spark_version` (DBR + release), this field `use_ml_runtime`, and whether `node_type_id` is + gpu node or not. + worker_node_type_flexibility: + $ref: '#/components/schemas/NodeTypeFlexibility' + description: Flexible node type configuration for worker nodes. + workload_type: + $ref: '#/components/schemas/WorkloadType' + UpdateClusterResponse: + type: object + properties: {} + UpdateResponse: + type: object + properties: {} + VolumesStorageInfo: + type: object + properties: + destination: + type: string + description: >- + UC Volumes destination, e.g. + `/Volumes/catalog/schema/vol1/init-scripts/setup-datadog.sh` or + `dbfs:/Volumes/catalog/schema/vol1/init-scripts/setup-datadog.sh` + required: + - destination + description: A storage location back by UC Volumes. + WorkloadType: + type: object + properties: + clients: + $ref: '#/components/schemas/ClientsTypes' + description: >- + defined what type of clients can use the cluster. E.g. Notebooks, + Jobs + required: + - clients + description: Cluster Attributes showing for clusters workload types. + WorkspaceStorageInfo: + type: object + properties: + destination: + type: string + description: >- + wsfs destination, e.g. + `workspace:/cluster-init-scripts/setup-datadog.sh` + required: + - destination + description: A storage location in Workspace Filesystem (WSFS) + AwsAvailability: + type: string + enum: + - ON_DEMAND + - SPOT + - SPOT_WITH_FALLBACK + description: >- + Availability type used for all subsequent nodes past the + `first_on_demand` ones. + + + Note: If `first_on_demand` is zero, this availability type will be used + for the entire cluster. + AzureAvailability: + type: string + enum: + - ON_DEMAND_AZURE + - SPOT_AZURE + - SPOT_WITH_FALLBACK_AZURE + description: >- + Availability type used for all subsequent nodes past the + `first_on_demand` ones. Note: If + + `first_on_demand` is zero, this availability type will be used for the + entire cluster. + CloudProviderNodeStatus: + type: string + enum: + - NotAvailableInRegion + - NotEnabledOnSubscription + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterPermissionLevel: + type: string + enum: + - CAN_ATTACH_TO + - CAN_MANAGE + - CAN_RESTART + description: Permission level + ClusterPolicyPermissionLevel: + type: string + enum: + - CAN_USE + description: Permission level + ClusterSource: + type: string + enum: + - API + - JOB + - MODELS + - PIPELINE + - PIPELINE_MAINTENANCE + - SQL + - UI + description: >- + Determines whether the cluster was created by a user through the UI, + created by the Databricks + + Jobs Scheduler, or through an API request. This is the same as + cluster_creator, but read only. + CommandStatus: + type: string + enum: + - Cancelled + - Cancelling + - Error + - Finished + - Queued + - Running + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ContextStatus: + type: string + enum: + - Error + - Pending + - Running + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DataPlaneEventDetailsEventType: + type: string + enum: + - NODE_BLACKLISTED + - NODE_EXCLUDED_DECOMMISSIONED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DataSecurityMode: + type: string + enum: + - DATA_SECURITY_MODE_AUTO + - DATA_SECURITY_MODE_DEDICATED + - DATA_SECURITY_MODE_STANDARD + - LEGACY_PASSTHROUGH + - LEGACY_SINGLE_USER + - LEGACY_SINGLE_USER_STANDARD + - LEGACY_TABLE_ACL + - NONE + - SINGLE_USER + - USER_ISOLATION + description: >- + Data security mode decides what data governance model to use when + accessing data from a cluster. + + + The following modes can only be used when `kind = CLASSIC_PREVIEW`. * + `DATA_SECURITY_MODE_AUTO`: + + Databricks will choose the most appropriate access mode depending on + your compute configuration. + + * `DATA_SECURITY_MODE_STANDARD`: Alias for `USER_ISOLATION`. * + `DATA_SECURITY_MODE_DEDICATED`: + + Alias for `SINGLE_USER`. + + + The following modes can be used regardless of `kind`. * `NONE`: No + security isolation for + + multiple users sharing the cluster. Data governance features are not + available in this mode. * + + `SINGLE_USER`: A secure cluster that can only be exclusively used by a + single user specified in + + `single_user_name`. Most programming languages, cluster features and + data governance features + + are available in this mode. * `USER_ISOLATION`: A secure cluster that + can be shared by multiple + + users. Cluster users are fully isolated so that they cannot see each + other's data and + + credentials. Most data governance features are supported in this mode. + But programming languages + + and cluster features might be limited. + + + The following modes are deprecated starting with Databricks Runtime 15.0 + and will be removed for + + future Databricks Runtime versions: + + + * `LEGACY_TABLE_ACL`: This mode is for users migrating from legacy Table + ACL clusters. * + + `LEGACY_PASSTHROUGH`: This mode is for users migrating from legacy + Passthrough on high + + concurrency clusters. * `LEGACY_SINGLE_USER`: This mode is for users + migrating from legacy + + Passthrough on standard clusters. * `LEGACY_SINGLE_USER_STANDARD`: This + mode provides a way that + + doesn’t have UC nor passthrough enabled. + DiskTypeAzureDiskVolumeType: + type: string + enum: + - PREMIUM_LRS + - STANDARD_LRS + description: >- + All Azure Disk types that Databricks supports. See + + https://docs.microsoft.com/en-us/azure/storage/storage-about-disks-and-vhds-linux#types-of-disks + DiskTypeEbsVolumeType: + type: string + enum: + - GENERAL_PURPOSE_SSD + - THROUGHPUT_OPTIMIZED_HDD + description: >- + All EBS volume types that Databricks supports. See + https://aws.amazon.com/ebs/details/ for + + details. + EbsVolumeType: + type: string + enum: + - GENERAL_PURPOSE_SSD + - THROUGHPUT_OPTIMIZED_HDD + description: >- + All EBS volume types that Databricks supports. See + https://aws.amazon.com/ebs/details/ for + + details. + EventDetailsCause: + type: string + enum: + - AUTORECOVERY + - AUTOSCALE + - AUTOSCALE_V2 + - REPLACE_BAD_NODES + - USER_REQUEST + description: The cause of a change in target size. + EventType: + type: string + enum: + - ADD_NODES_FAILED + - AUTOMATIC_CLUSTER_UPDATE + - AUTOSCALING_BACKOFF + - AUTOSCALING_FAILED + - AUTOSCALING_STATS_REPORT + - CLUSTER_MIGRATED + - CREATING + - DBFS_DOWN + - DECOMMISSION_ENDED + - DECOMMISSION_STARTED + - DID_NOT_EXPAND_DISK + - DRIVER_HEALTHY + - DRIVER_NOT_RESPONDING + - DRIVER_UNAVAILABLE + - EDITED + - EXPANDED_DISK + - FAILED_TO_EXPAND_DISK + - INIT_SCRIPTS_FINISHED + - INIT_SCRIPTS_STARTED + - METASTORE_DOWN + - NODES_LOST + - NODE_BLACKLISTED + - NODE_EXCLUDED_DECOMMISSIONED + - PINNED + - RESIZING + - RESTARTING + - RUNNING + - SPARK_EXCEPTION + - STARTING + - TERMINATING + - UC_VOLUME_MISCONFIGURED + - UNPINNED + - UPSIZE_COMPLETED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + GcpAvailability: + type: string + enum: + - ON_DEMAND_GCP + - PREEMPTIBLE_GCP + - PREEMPTIBLE_WITH_FALLBACK_GCP + description: >- + This field determines whether the instance pool will contain preemptible + VMs, on-demand VMs, or + + preemptible VMs with a fallback to on-demand VMs if the former is + unavailable. + GetEventsOrder: + type: string + enum: + - ASC + - DESC + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + HardwareAcceleratorType: + type: string + enum: + - GPU_1xA10 + - GPU_8xH100 + description: >- + HardwareAcceleratorType: The type of hardware accelerator to use for + compute workloads. NOTE: + + This enum is referenced and is intended to be used by other Databricks + services that need to + + specify hardware accelerator requirements for AI compute workloads. + InitScriptExecutionDetailsInitScriptExecutionStatus: + type: string + enum: + - FAILED_EXECUTION + - FAILED_FETCH + - FUSE_MOUNT_FAILED + - NOT_EXECUTED + - SKIPPED + - SUCCEEDED + - UNKNOWN + description: Result of attempted script execution + InstancePoolAwsAttributesAvailability: + type: string + enum: + - ON_DEMAND + - SPOT + description: >- + The set of AWS availability types supported when setting up nodes for a + cluster. + InstancePoolAzureAttributesAvailability: + type: string + enum: + - ON_DEMAND_AZURE + - SPOT_AZURE + description: >- + The set of Azure availability types supported when setting up nodes for + a cluster. + InstancePoolPermissionLevel: + type: string + enum: + - CAN_ATTACH_TO + - CAN_MANAGE + description: Permission level + InstancePoolState: + type: string + enum: + - ACTIVE + - DELETED + - STOPPED + description: >- + The state of a Cluster. The current allowable state transitions are as + follows: + + + - ``ACTIVE`` -> ``STOPPED`` - ``ACTIVE`` -> ``DELETED`` - ``STOPPED`` -> + ``ACTIVE`` - + + ``STOPPED`` -> ``DELETED`` + Kind: + type: string + enum: + - CLASSIC_PREVIEW + description: >- + The kind of compute described by this compute specification. + + + Depending on `kind`, different validations and default values will be + applied. + + + Clusters with `kind = CLASSIC_PREVIEW` support the following fields, + whereas clusters with no + + specified `kind` do not. * + [is_single_node](/api/workspace/clusters/create#is_single_node) * + + [use_ml_runtime](/api/workspace/clusters/create#use_ml_runtime) * + + [data_security_mode](/api/workspace/clusters/create#data_security_mode) + set to + + `DATA_SECURITY_MODE_AUTO`, `DATA_SECURITY_MODE_DEDICATED`, or + `DATA_SECURITY_MODE_STANDARD` + + + By using the [simple form], your clusters are automatically using `kind + = CLASSIC_PREVIEW`. + + + [simple form]: https://docs.databricks.com/compute/simple-form.html + Language: + type: string + enum: + - python + - r + - scala + - sql + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + LibraryInstallStatus: + type: string + enum: + - FAILED + - INSTALLED + - INSTALLING + - PENDING + - RESOLVING + - RESTORED + - SKIPPED + - UNINSTALL_ON_RESTART + description: The status of a library on a specific cluster. + ListClustersSortByDirection: + type: string + enum: + - ASC + - DESC + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListClustersSortByField: + type: string + enum: + - CLUSTER_NAME + - DEFAULT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListSortColumn: + type: string + enum: + - POLICY_CREATION_TIME + - POLICY_NAME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListSortOrder: + type: string + enum: + - ASC + - DESC + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ResultType: + type: string + enum: + - error + - image + - images + - table + - text + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + RuntimeEngine: + type: string + enum: + - 'NULL' + - PHOTON + - STANDARD + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + State: + type: string + enum: + - ERROR + - PENDING + - RESIZING + - RESTARTING + - RUNNING + - TERMINATED + - TERMINATING + - UNKNOWN + description: >- + The state of a Cluster. The current allowable state transitions are as + follows: + + + - `PENDING` -> `RUNNING` - `PENDING` -> `TERMINATING` - `RUNNING` -> + `RESIZING` - `RUNNING` -> + + `RESTARTING` - `RUNNING` -> `TERMINATING` - `RESTARTING` -> `RUNNING` - + `RESTARTING` -> + + `TERMINATING` - `RESIZING` -> `RUNNING` - `RESIZING` -> `TERMINATING` - + `TERMINATING` -> + + `TERMINATED` + TerminationReasonCode: + type: string + enum: + - ABUSE_DETECTED + - ACCESS_TOKEN_FAILURE + - ALLOCATION_TIMEOUT + - ALLOCATION_TIMEOUT_NODE_DAEMON_NOT_READY + - ALLOCATION_TIMEOUT_NO_HEALTHY_AND_WARMED_UP_CLUSTERS + - ALLOCATION_TIMEOUT_NO_HEALTHY_CLUSTERS + - ALLOCATION_TIMEOUT_NO_MATCHED_CLUSTERS + - ALLOCATION_TIMEOUT_NO_READY_CLUSTERS + - ALLOCATION_TIMEOUT_NO_UNALLOCATED_CLUSTERS + - ALLOCATION_TIMEOUT_NO_WARMED_UP_CLUSTERS + - ATTACH_PROJECT_FAILURE + - AWS_AUTHORIZATION_FAILURE + - AWS_INACCESSIBLE_KMS_KEY_FAILURE + - AWS_INSTANCE_PROFILE_UPDATE_FAILURE + - AWS_INSUFFICIENT_FREE_ADDRESSES_IN_SUBNET_FAILURE + - AWS_INSUFFICIENT_INSTANCE_CAPACITY_FAILURE + - AWS_INVALID_KEY_PAIR + - AWS_INVALID_KMS_KEY_STATE + - AWS_MAX_SPOT_INSTANCE_COUNT_EXCEEDED_FAILURE + - AWS_REQUEST_LIMIT_EXCEEDED + - AWS_RESOURCE_QUOTA_EXCEEDED + - AWS_UNSUPPORTED_FAILURE + - AZURE_BYOK_KEY_PERMISSION_FAILURE + - AZURE_EPHEMERAL_DISK_FAILURE + - AZURE_INVALID_DEPLOYMENT_TEMPLATE + - AZURE_OPERATION_NOT_ALLOWED_EXCEPTION + - AZURE_PACKED_DEPLOYMENT_PARTIAL_FAILURE + - AZURE_QUOTA_EXCEEDED_EXCEPTION + - AZURE_RESOURCE_MANAGER_THROTTLING + - AZURE_RESOURCE_PROVIDER_THROTTLING + - AZURE_UNEXPECTED_DEPLOYMENT_TEMPLATE_FAILURE + - AZURE_VM_EXTENSION_FAILURE + - AZURE_VNET_CONFIGURATION_FAILURE + - BOOTSTRAP_TIMEOUT + - BOOTSTRAP_TIMEOUT_CLOUD_PROVIDER_EXCEPTION + - BOOTSTRAP_TIMEOUT_DUE_TO_MISCONFIG + - BUDGET_POLICY_LIMIT_ENFORCEMENT_ACTIVATED + - BUDGET_POLICY_RESOLUTION_FAILURE + - CLOUD_ACCOUNT_POD_QUOTA_EXCEEDED + - CLOUD_ACCOUNT_SETUP_FAILURE + - CLOUD_OPERATION_CANCELLED + - CLOUD_PROVIDER_DISK_SETUP_FAILURE + - CLOUD_PROVIDER_INSTANCE_NOT_LAUNCHED + - CLOUD_PROVIDER_LAUNCH_FAILURE + - CLOUD_PROVIDER_LAUNCH_FAILURE_DUE_TO_MISCONFIG + - CLOUD_PROVIDER_RESOURCE_STOCKOUT + - CLOUD_PROVIDER_RESOURCE_STOCKOUT_DUE_TO_MISCONFIG + - CLOUD_PROVIDER_SHUTDOWN + - CLUSTER_OPERATION_THROTTLED + - CLUSTER_OPERATION_TIMEOUT + - COMMUNICATION_LOST + - CONTAINER_LAUNCH_FAILURE + - CONTROL_PLANE_CONNECTION_FAILURE + - CONTROL_PLANE_CONNECTION_FAILURE_DUE_TO_MISCONFIG + - CONTROL_PLANE_REQUEST_FAILURE + - CONTROL_PLANE_REQUEST_FAILURE_DUE_TO_MISCONFIG + - DATABASE_CONNECTION_FAILURE + - DATA_ACCESS_CONFIG_CHANGED + - DBFS_COMPONENT_UNHEALTHY + - DBR_IMAGE_RESOLUTION_FAILURE + - DISASTER_RECOVERY_REPLICATION + - DNS_RESOLUTION_ERROR + - DOCKER_CONTAINER_CREATION_EXCEPTION + - DOCKER_IMAGE_PULL_FAILURE + - DOCKER_IMAGE_TOO_LARGE_FOR_INSTANCE_EXCEPTION + - DOCKER_INVALID_OS_EXCEPTION + - DRIVER_EVICTION + - DRIVER_LAUNCH_TIMEOUT + - DRIVER_NODE_UNREACHABLE + - DRIVER_OUT_OF_DISK + - DRIVER_OUT_OF_MEMORY + - DRIVER_POD_CREATION_FAILURE + - DRIVER_UNEXPECTED_FAILURE + - DRIVER_UNHEALTHY + - DRIVER_UNREACHABLE + - DRIVER_UNRESPONSIVE + - DYNAMIC_SPARK_CONF_SIZE_EXCEEDED + - EOS_SPARK_IMAGE + - EXECUTION_COMPONENT_UNHEALTHY + - EXECUTOR_POD_UNSCHEDULED + - GCP_API_RATE_QUOTA_EXCEEDED + - GCP_DENIED_BY_ORG_POLICY + - GCP_FORBIDDEN + - GCP_IAM_TIMEOUT + - GCP_INACCESSIBLE_KMS_KEY_FAILURE + - GCP_INSUFFICIENT_CAPACITY + - GCP_IP_SPACE_EXHAUSTED + - GCP_KMS_KEY_PERMISSION_DENIED + - GCP_NOT_FOUND + - GCP_QUOTA_EXCEEDED + - GCP_RESOURCE_QUOTA_EXCEEDED + - GCP_SERVICE_ACCOUNT_ACCESS_DENIED + - GCP_SERVICE_ACCOUNT_DELETED + - GCP_SERVICE_ACCOUNT_NOT_FOUND + - GCP_SUBNET_NOT_READY + - GCP_TRUSTED_IMAGE_PROJECTS_VIOLATED + - GKE_BASED_CLUSTER_TERMINATION + - GLOBAL_INIT_SCRIPT_FAILURE + - HIVE_METASTORE_PROVISIONING_FAILURE + - IMAGE_PULL_PERMISSION_DENIED + - INACTIVITY + - INIT_CONTAINER_NOT_FINISHED + - INIT_SCRIPT_FAILURE + - INSTANCE_POOL_CLUSTER_FAILURE + - INSTANCE_POOL_MAX_CAPACITY_REACHED + - INSTANCE_POOL_NOT_FOUND + - INSTANCE_UNREACHABLE + - INSTANCE_UNREACHABLE_DUE_TO_MISCONFIG + - INTERNAL_CAPACITY_FAILURE + - INTERNAL_ERROR + - INVALID_ARGUMENT + - INVALID_AWS_PARAMETER + - INVALID_INSTANCE_PLACEMENT_PROTOCOL + - INVALID_SPARK_IMAGE + - INVALID_WORKER_IMAGE_FAILURE + - IN_PENALTY_BOX + - IP_EXHAUSTION_FAILURE + - JOB_FINISHED + - K8S_ACTIVE_POD_QUOTA_EXCEEDED + - K8S_AUTOSCALING_FAILURE + - K8S_DBR_CLUSTER_LAUNCH_TIMEOUT + - LAZY_ALLOCATION_TIMEOUT + - MAINTENANCE_MODE + - METASTORE_COMPONENT_UNHEALTHY + - MTLS_PORT_CONNECTIVITY_FAILURE + - NEPHOS_RESOURCE_MANAGEMENT + - NETVISOR_SETUP_TIMEOUT + - NETWORK_CHECK_CONTROL_PLANE_FAILURE + - NETWORK_CHECK_CONTROL_PLANE_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_DNS_SERVER_FAILURE + - NETWORK_CHECK_DNS_SERVER_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_METADATA_ENDPOINT_FAILURE + - NETWORK_CHECK_METADATA_ENDPOINT_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_MULTIPLE_COMPONENTS_FAILURE + - NETWORK_CHECK_MULTIPLE_COMPONENTS_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_NIC_FAILURE + - NETWORK_CHECK_NIC_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_STORAGE_FAILURE + - NETWORK_CHECK_STORAGE_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CONFIGURATION_FAILURE + - NFS_MOUNT_FAILURE + - NO_MATCHED_K8S + - NO_MATCHED_K8S_TESTING_TAG + - NPIP_TUNNEL_SETUP_FAILURE + - NPIP_TUNNEL_TOKEN_FAILURE + - POD_ASSIGNMENT_FAILURE + - POD_SCHEDULING_FAILURE + - RATE_LIMITED + - REQUEST_REJECTED + - REQUEST_THROTTLED + - RESOURCE_USAGE_BLOCKED + - SECRET_CREATION_FAILURE + - SECRET_PERMISSION_DENIED + - SECRET_RESOLUTION_ERROR + - SECURITY_DAEMON_REGISTRATION_EXCEPTION + - SELF_BOOTSTRAP_FAILURE + - SERVERLESS_LONG_RUNNING_TERMINATED + - SKIPPED_SLOW_NODES + - SLOW_IMAGE_DOWNLOAD + - SPARK_ERROR + - SPARK_IMAGE_DOWNLOAD_FAILURE + - SPARK_IMAGE_DOWNLOAD_THROTTLED + - SPARK_IMAGE_NOT_FOUND + - SPARK_STARTUP_FAILURE + - SPOT_INSTANCE_TERMINATION + - SSH_BOOTSTRAP_FAILURE + - STORAGE_DOWNLOAD_FAILURE + - STORAGE_DOWNLOAD_FAILURE_DUE_TO_MISCONFIG + - STORAGE_DOWNLOAD_FAILURE_SLOW + - STORAGE_DOWNLOAD_FAILURE_THROTTLED + - STS_CLIENT_SETUP_FAILURE + - SUBNET_EXHAUSTED_FAILURE + - TEMPORARILY_UNAVAILABLE + - TRIAL_EXPIRED + - UNEXPECTED_LAUNCH_FAILURE + - UNEXPECTED_POD_RECREATION + - UNKNOWN + - UNSUPPORTED_INSTANCE_TYPE + - UPDATE_INSTANCE_PROFILE_FAILURE + - USAGE_POLICY_ENTITLEMENT_DENIED + - USER_INITIATED_VM_TERMINATION + - USER_REQUEST + - WORKER_SETUP_FAILURE + - WORKSPACE_CANCELLED_ERROR + - WORKSPACE_CONFIGURATION_ERROR + - WORKSPACE_UPDATE + description: The status code indicating why the cluster was terminated + TerminationReasonType: + type: string + enum: + - CLIENT_ERROR + - CLOUD_FAILURE + - SERVICE_FAULT + - SUCCESS + description: type of the termination + x-stackQL-resources: + cluster_policies: + id: databricks_workspace.compute.cluster_policies + name: cluster_policies + title: Cluster Policies + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1edit/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.policies + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/get' + - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/cluster_policies/methods/replace' + cluster_policy_permission_levels: + id: databricks_workspace.compute.cluster_policy_permission_levels + name: cluster_policy_permission_levels + title: Cluster Policy Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/cluster_policy_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + cluster_policy_permissions: + id: databricks_workspace.compute.cluster_policy_permissions + name: cluster_policy_permissions + title: Cluster Policy Permissions + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1cluster-policies~1{cluster_policy_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/cluster_policy_permissions/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/cluster_policy_permissions/methods/update + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/cluster_policy_permissions/methods/set + clusters: + id: databricks_workspace.compute.clusters + name: clusters + title: Clusters + methods: + change_owner: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1change-owner/post' + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + edit: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1edit/post' + response: + mediaType: application/json + openAPIDocKey: '200' + events: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1events/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.clusters + permanent_delete: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1permanent-delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + pin: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1pin/post' + response: + mediaType: application/json + openAPIDocKey: '200' + resize: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1resize/post' + response: + mediaType: application/json + openAPIDocKey: '200' + restart: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1restart/post' + response: + mediaType: application/json + openAPIDocKey: '200' + start: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1start/post' + response: + mediaType: application/json + openAPIDocKey: '200' + unpin: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1unpin/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1update/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/clusters/methods/get' + - $ref: '#/components/x-stackQL-resources/clusters/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/clusters/methods/create' + update: [] + delete: [] + replace: [] + cluster_permission_levels: + id: databricks_workspace.compute.cluster_permission_levels + name: cluster_permission_levels + title: Cluster Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/cluster_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + cluster_permissions: + id: databricks_workspace.compute.cluster_permissions + name: cluster_permissions + title: Cluster Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1clusters~1{cluster_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/cluster_permissions/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/cluster_permissions/methods/update + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/cluster_permissions/methods/set' + cluster_node_types: + id: databricks_workspace.compute.cluster_node_types + name: cluster_node_types + title: Cluster Node Types + methods: + list: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1list-node-types/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/cluster_node_types/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + cluster_zones: + id: databricks_workspace.compute.cluster_zones + name: cluster_zones + title: Cluster Zones + methods: + list: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1list-zones/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/cluster_zones/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + spark_versions: + id: databricks_workspace.compute.spark_versions + name: spark_versions + title: Spark Versions + methods: + list: + operation: + $ref: '#/paths/~1api~12.1~1clusters~1spark-versions/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/spark_versions/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + command_execution: + id: databricks_workspace.compute.command_execution + name: command_execution + title: Command Execution + methods: + cancel: + operation: + $ref: '#/paths/~1api~11.2~1commands~1cancel/post' + response: + mediaType: application/json + openAPIDocKey: '200' + command_status: + operation: + $ref: '#/paths/~1api~11.2~1commands~1status/get' + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~11.2~1contexts~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + destroy: + operation: + $ref: '#/paths/~1api~11.2~1contexts~1destroy/post' + response: + mediaType: application/json + openAPIDocKey: '200' + execute: + operation: + $ref: '#/paths/~1api~11.2~1commands~1execute/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/command_execution/methods/command_status + insert: + - $ref: '#/components/x-stackQL-resources/command_execution/methods/create' + update: [] + delete: [] + replace: [] + execution_contexts: + id: databricks_workspace.compute.execution_contexts + name: execution_contexts + title: Execution Contexts + methods: + get: + operation: + $ref: '#/paths/~1api~11.2~1contexts~1status/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/execution_contexts/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + global_init_scripts: + id: databricks_workspace.compute.global_init_scripts + name: global_init_scripts + title: Global Init Scripts + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1global-init-scripts/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1global-init-scripts/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.scripts + delete: + operation: + $ref: '#/paths/~1api~12.0~1global-init-scripts~1{script_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1global-init-scripts~1{script_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1global-init-scripts~1{script_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/global_init_scripts/methods/get' + - $ref: '#/components/x-stackQL-resources/global_init_scripts/methods/list' + insert: + - $ref: >- + #/components/x-stackQL-resources/global_init_scripts/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/global_init_scripts/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/global_init_scripts/methods/delete + replace: [] + instance_pools: + id: databricks_workspace.compute.instance_pools + name: instance_pools + title: Instance Pools + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1instance-pools~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1instance-pools~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1instance-pools~1edit/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1instance-pools~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1instance-pools~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.instance_pools + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/instance_pools/methods/get' + - $ref: '#/components/x-stackQL-resources/instance_pools/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/instance_pools/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/instance_pools/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/instance_pools/methods/replace' + instance_pool_permission_levels: + id: databricks_workspace.compute.instance_pool_permission_levels + name: instance_pool_permission_levels + title: Instance Pool Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/instance_pool_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + instance_pool_permissions: + id: databricks_workspace.compute.instance_pool_permissions + name: instance_pool_permissions + title: Instance Pool Permissions + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1instance-pools~1{instance_pool_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/instance_pool_permissions/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/instance_pool_permissions/methods/update + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/instance_pool_permissions/methods/set + instance_profiles: + id: databricks_workspace.compute.instance_profiles + name: instance_profiles + title: Instance Profiles + methods: + add: + operation: + $ref: '#/paths/~1api~12.0~1instance-profiles~1add/post' + response: + mediaType: application/json + openAPIDocKey: '200' + edit: + operation: + $ref: '#/paths/~1api~12.0~1instance-profiles~1edit/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1instance-profiles~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.instance_profiles + remove: + operation: + $ref: '#/paths/~1api~12.0~1instance-profiles~1remove/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/add' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/remove' + replace: + - $ref: '#/components/x-stackQL-resources/instance_profiles/methods/edit' + all_cluster_library_statuses: + id: databricks_workspace.compute.all_cluster_library_statuses + name: all_cluster_library_statuses + title: All Cluster Library Statuses + methods: + list: + operation: + $ref: '#/paths/~1api~12.0~1libraries~1all-cluster-statuses/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.statuses + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/all_cluster_library_statuses/methods/list + insert: [] + update: [] + delete: [] + replace: [] + libraries: + id: databricks_workspace.compute.libraries + name: libraries + title: Libraries + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1libraries~1cluster-status/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.library_statuses + install: + operation: + $ref: '#/paths/~1api~12.0~1libraries~1install/post' + response: + mediaType: application/json + openAPIDocKey: '200' + uninstall: + operation: + $ref: '#/paths/~1api~12.0~1libraries~1uninstall/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/libraries/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/libraries/methods/install' + update: [] + delete: [] + replace: [] + policy_compliance_for_clusters: + id: databricks_workspace.compute.policy_compliance_for_clusters + name: policy_compliance_for_clusters + title: Policy Compliance For Clusters + methods: + enforce: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1enforce-compliance/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1get-compliance/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1policies~1clusters~1list-compliance/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.clusters + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/policy_compliance_for_clusters/methods/get + - $ref: >- + #/components/x-stackQL-resources/policy_compliance_for_clusters/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/policy_compliance_for_clusters/methods/enforce + update: [] + delete: [] + replace: [] + policy_families: + id: databricks_workspace.compute.policy_families + name: policy_families + title: Policy Families + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1policy-families~1{policy_family_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1policy-families/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.policy_families + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/policy_families/methods/get' + - $ref: '#/components/x-stackQL-resources/policy_families/methods/list' + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/dashboards.yaml b/providers/src/databricks_workspace/v00.00.00000/services/dashboards.yaml new file mode 100644 index 00000000..c554f198 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/dashboards.yaml @@ -0,0 +1,3668 @@ +openapi: 3.0.0 +info: + title: Databricks Dashboards API (workspace) + description: >- + OpenAPI specification for the Databricks dashboards service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.dashboards +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages: + post: + operationId: genie_create_message + summary: >- + Create new message in a [conversation](:method:genie/startconversation). + The AI response uses all + tags: + - dashboards + - genie + description: >- + Create new message in a [conversation](:method:genie/startconversation). + The AI response uses all + + previously created messages in the conversation to respond. + + + :param space_id: str + The ID associated with the Genie space where the conversation is started. + :param conversation_id: str + The ID associated with the conversation. + :param content: str + User message content. + + :returns: + Long-running operation waiter for :class:`GenieMessage`. + See :method:wait_get_message_genie_completed for more details. + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: >- + The ID associated with the Genie space where the conversation is + started. + - name: conversation_id + in: path + required: true + schema: + type: string + description: The ID associated with the conversation. + requestBody: + content: + application/json: + schema: + type: object + properties: + content: + type: string + description: User message content. + required: + - content + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieMessage' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + get: + operationId: genie_list_conversation_messages + summary: List messages in a conversation + tags: + - dashboards + - genie + description: |- + List messages in a conversation + + :param space_id: str + The ID associated with the Genie space where the conversation is located + :param conversation_id: str + The ID of the conversation to list messages from + :param page_size: int (optional) + Maximum number of messages to return per page + :param page_token: str (optional) + Token to get the next page of results + + :returns: :class:`GenieListConversationMessagesResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: >- + The ID associated with the Genie space where the conversation is + located + - name: conversation_id + in: path + required: true + schema: + type: string + description: The ID of the conversation to list messages from + - name: page_size + in: query + required: false + schema: + type: string + description: Maximum number of messages to return per page + - name: page_token + in: query + required: false + schema: + type: string + description: Token to get the next page of results + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieListConversationMessagesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces: + post: + operationId: genie_create_space + summary: Creates a Genie space from a serialized payload. + tags: + - dashboards + - genie + description: |- + Creates a Genie space from a serialized payload. + + :param warehouse_id: str + Warehouse to associate with the new space + :param serialized_space: str + The contents of the Genie Space in serialized string form. Use the [Get Genie + Space](:method:genie/getspace) API to retrieve an example response, which includes the + `serialized_space` field. This field provides the structure of the JSON string that represents the + space's layout and components. + :param description: str (optional) + Optional description + :param parent_path: str (optional) + Parent folder path where the space will be registered + :param title: str (optional) + Optional title override + + :returns: :class:`GenieSpace` + requestBody: + content: + application/json: + schema: + type: object + properties: + warehouse_id: + type: string + description: Warehouse to associate with the new space + serialized_space: + type: string + description: >- + The contents of the Genie Space in serialized string form. + Use the [Get Genie Space](:method:genie/getspace) API to + retrieve an example response, which includes the + `serialized_space` field. This field provides the structure + of the JSON string that represents the space's layout and + components. + description: + type: string + description: Optional description + parent_path: + type: string + description: Parent folder path where the space will be registered + title: + type: string + description: Optional title override + required: + - warehouse_id + - serialized_space + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieSpace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + get: + operationId: genie_list_spaces + summary: Get list of Genie Spaces. + tags: + - dashboards + - genie + description: |- + Get list of Genie Spaces. + + :param page_size: int (optional) + Maximum number of spaces to return per page + :param page_token: str (optional) + Pagination token for getting the next page of results + + :returns: :class:`GenieListSpacesResponse` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: Maximum number of spaces to return per page + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token for getting the next page of results + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieListSpacesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}: + delete: + operationId: genie_delete_conversation + summary: Delete a conversation. + tags: + - dashboards + - genie + description: |- + Delete a conversation. + + :param space_id: str + The ID associated with the Genie space where the conversation is located. + :param conversation_id: str + The ID of the conversation to delete. + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: >- + The ID associated with the Genie space where the conversation is + located. + - name: conversation_id + in: path + required: true + schema: + type: string + description: The ID of the conversation to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}: + delete: + operationId: genie_delete_conversation_message + summary: Delete a conversation message. + tags: + - dashboards + - genie + description: |- + Delete a conversation message. + + :param space_id: str + The ID associated with the Genie space where the message is located. + :param conversation_id: str + The ID associated with the conversation. + :param message_id: str + The ID associated with the message to delete. + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: The ID associated with the Genie space where the message is located. + - name: conversation_id + in: path + required: true + schema: + type: string + description: The ID associated with the conversation. + - name: message_id + in: path + required: true + schema: + type: string + description: The ID associated with the message to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + get: + operationId: genie_get_message + summary: Get message from conversation. + tags: + - dashboards + - genie + description: |- + Get message from conversation. + + :param space_id: str + The ID associated with the Genie space where the target conversation is located. + :param conversation_id: str + The ID associated with the target conversation. + :param message_id: str + The ID associated with the target message from the identified conversation. + + :returns: :class:`GenieMessage` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: >- + The ID associated with the Genie space where the target conversation + is located. + - name: conversation_id + in: path + required: true + schema: + type: string + description: The ID associated with the target conversation. + - name: message_id + in: path + required: true + schema: + type: string + description: >- + The ID associated with the target message from the identified + conversation. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieMessage' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/attachments/{attachment_id}/execute-query: + post: + operationId: genie_execute_message_attachment_query + summary: >- + Execute the SQL for a message query attachment. Use this API when the + query attachment has expired and + tags: + - dashboards + - genie + description: >- + Execute the SQL for a message query attachment. Use this API when the + query attachment has expired and + + needs to be re-executed. + + + :param space_id: str + Genie space ID + :param conversation_id: str + Conversation ID + :param message_id: str + Message ID + :param attachment_id: str + Attachment ID + + :returns: :class:`GenieGetMessageQueryResultResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + - name: conversation_id + in: path + required: true + schema: + type: string + description: Conversation ID + - name: message_id + in: path + required: true + schema: + type: string + description: Message ID + - name: attachment_id + in: path + required: true + schema: + type: string + description: Attachment ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieGetMessageQueryResultResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/execute-query: + post: + operationId: genie_execute_message_query + summary: >- + DEPRECATED: Use [Execute Message Attachment + Query](:method:genie/executemessageattachmentquery) + tags: + - dashboards + - genie + description: >- + DEPRECATED: Use [Execute Message Attachment + Query](:method:genie/executemessageattachmentquery) + + instead. + + + :param space_id: str + Genie space ID + :param conversation_id: str + Conversation ID + :param message_id: str + Message ID + + :returns: :class:`GenieGetMessageQueryResultResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + - name: conversation_id + in: path + required: true + schema: + type: string + description: Conversation ID + - name: message_id + in: path + required: true + schema: + type: string + description: Message ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieGetMessageQueryResultResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/attachments/{attachment_id}/downloads: + post: + operationId: genie_generate_download_full_query_result + summary: >- + Initiates a new SQL execution and returns a `download_id` and + `download_id_signature` that you can use + tags: + - dashboards + - genie + description: >- + Initiates a new SQL execution and returns a `download_id` and + `download_id_signature` that you can use + + to track the progress of the download. The query result is stored in an + external link and can be + + retrieved using the [Get Download Full Query + Result](:method:genie/getdownloadfullqueryresult) API. + + Both `download_id` and `download_id_signature` must be provided when + calling the Get endpoint. + + + ---- + + + ### **Warning: Databricks strongly recommends that you protect the URLs + that are returned by the + + `EXTERNAL_LINKS` disposition.** + + + When you use the `EXTERNAL_LINKS` disposition, a short-lived, URL is + generated, which can be used to + + download the results directly from . As a short-lived is embedded in + this URL, you should protect the + + URL. + + + Because URLs are already generated with embedded temporary s, you must + not set an `Authorization` + + header in the download requests. + + + See [Execute Statement](:method:statementexecution/executestatement) for + more details. + + + ---- + + + :param space_id: str + Genie space ID + :param conversation_id: str + Conversation ID + :param message_id: str + Message ID + :param attachment_id: str + Attachment ID + + :returns: :class:`GenieGenerateDownloadFullQueryResultResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + - name: conversation_id + in: path + required: true + schema: + type: string + description: Conversation ID + - name: message_id + in: path + required: true + schema: + type: string + description: Message ID + - name: attachment_id + in: path + required: true + schema: + type: string + description: Attachment ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/GenieGenerateDownloadFullQueryResultResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/attachments/{attachment_id}/downloads/{download_id}: + get: + operationId: genie_get_download_full_query_result + summary: >- + After [Generating a Full Query Result + Download](:method:genie/generatedownloadfullqueryresult) and + tags: + - dashboards + - genie + description: >- + After [Generating a Full Query Result + Download](:method:genie/generatedownloadfullqueryresult) and + + successfully receiving a `download_id` and `download_id_signature`, use + this API to poll the download + + progress. Both `download_id` and `download_id_signature` are required to + call this endpoint. When the + + download is complete, the API returns the result in the `EXTERNAL_LINKS` + disposition, containing one + + or more external links to the query result files. + + + ---- + + + ### **Warning: Databricks strongly recommends that you protect the URLs + that are returned by the + + `EXTERNAL_LINKS` disposition.** + + + When you use the `EXTERNAL_LINKS` disposition, a short-lived, URL is + generated, which can be used to + + download the results directly from . As a short-lived is embedded in + this URL, you should protect the + + URL. + + + Because URLs are already generated with embedded temporary s, you must + not set an `Authorization` + + header in the download requests. + + + See [Execute Statement](:method:statementexecution/executestatement) for + more details. + + + ---- + + + :param space_id: str + Genie space ID + :param conversation_id: str + Conversation ID + :param message_id: str + Message ID + :param attachment_id: str + Attachment ID + :param download_id: str + Download ID. This ID is provided by the [Generate Download + endpoint](:method:genie/generateDownloadFullQueryResult) + :param download_id_signature: str (optional) + JWT signature for the download_id to ensure secure access to query results + + :returns: :class:`GenieGetDownloadFullQueryResultResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + - name: conversation_id + in: path + required: true + schema: + type: string + description: Conversation ID + - name: message_id + in: path + required: true + schema: + type: string + description: Message ID + - name: attachment_id + in: path + required: true + schema: + type: string + description: Attachment ID + - name: download_id + in: path + required: true + schema: + type: string + description: >- + Download ID. This ID is provided by the [Generate Download + endpoint](:method:genie/generateDownloadFullQueryResult) + - name: download_id_signature + in: query + required: false + schema: + type: string + description: >- + JWT signature for the download_id to ensure secure access to query + results + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieGetDownloadFullQueryResultResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/attachments/{attachment_id}/query-result: + get: + operationId: genie_get_message_attachment_query_result + summary: >- + Get the result of SQL query if the message has a query attachment. This + is only available if a message + tags: + - dashboards + - genie + description: >- + Get the result of SQL query if the message has a query attachment. This + is only available if a message + + has a query attachment and the message status is `EXECUTING_QUERY` OR + `COMPLETED`. + + + :param space_id: str + Genie space ID + :param conversation_id: str + Conversation ID + :param message_id: str + Message ID + :param attachment_id: str + Attachment ID + + :returns: :class:`GenieGetMessageQueryResultResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + - name: conversation_id + in: path + required: true + schema: + type: string + description: Conversation ID + - name: message_id + in: path + required: true + schema: + type: string + description: Message ID + - name: attachment_id + in: path + required: true + schema: + type: string + description: Attachment ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieGetMessageQueryResultResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/query-result: + get: + operationId: genie_get_message_query_result + summary: >- + DEPRECATED: Use [Get Message Attachment Query + Result](:method:genie/getmessageattachmentqueryresult) + tags: + - dashboards + - genie + description: >- + DEPRECATED: Use [Get Message Attachment Query + Result](:method:genie/getmessageattachmentqueryresult) + + instead. + + + :param space_id: str + Genie space ID + :param conversation_id: str + Conversation ID + :param message_id: str + Message ID + + :returns: :class:`GenieGetMessageQueryResultResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + - name: conversation_id + in: path + required: true + schema: + type: string + description: Conversation ID + - name: message_id + in: path + required: true + schema: + type: string + description: Message ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieGetMessageQueryResultResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/query-result/{attachment_id}: + get: + operationId: genie_get_message_query_result_by_attachment + summary: >- + DEPRECATED: Use [Get Message Attachment Query + Result](:method:genie/getmessageattachmentqueryresult) + tags: + - dashboards + - genie + description: >- + DEPRECATED: Use [Get Message Attachment Query + Result](:method:genie/getmessageattachmentqueryresult) + + instead. + + + :param space_id: str + Genie space ID + :param conversation_id: str + Conversation ID + :param message_id: str + Message ID + :param attachment_id: str + Attachment ID + + :returns: :class:`GenieGetMessageQueryResultResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + - name: conversation_id + in: path + required: true + schema: + type: string + description: Conversation ID + - name: message_id + in: path + required: true + schema: + type: string + description: Message ID + - name: attachment_id + in: path + required: true + schema: + type: string + description: Attachment ID + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieGetMessageQueryResultResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}: + get: + operationId: genie_get_space + summary: Get details of a Genie Space. + tags: + - dashboards + - genie + description: |- + Get details of a Genie Space. + + :param space_id: str + The ID associated with the Genie space + :param include_serialized_space: bool (optional) + Whether to include the serialized space export in the response. Requires at least CAN EDIT + permission on the space. + + :returns: :class:`GenieSpace` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: The ID associated with the Genie space + - name: include_serialized_space + in: query + required: false + schema: + type: string + description: >- + Whether to include the serialized space export in the response. + Requires at least CAN EDIT permission on the space. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieSpace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + delete: + operationId: genie_trash_space + summary: Move a Genie Space to the trash. + tags: + - dashboards + - genie + description: |- + Move a Genie Space to the trash. + + :param space_id: str + The ID associated with the Genie space to be sent to the trash. + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: The ID associated with the Genie space to be sent to the trash. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + patch: + operationId: genie_update_space + summary: Updates a Genie space with a serialized payload. + tags: + - dashboards + - genie + description: |- + Updates a Genie space with a serialized payload. + + :param space_id: str + Genie space ID + :param description: str (optional) + Optional description + :param serialized_space: str (optional) + The contents of the Genie Space in serialized string form (full replacement). Use the [Get Genie + Space](:method:genie/getspace) API to retrieve an example response, which includes the + `serialized_space` field. This field provides the structure of the JSON string that represents the + space's layout and components. + :param title: str (optional) + Optional title override + :param warehouse_id: str (optional) + Optional warehouse override + + :returns: :class:`GenieSpace` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: Genie space ID + requestBody: + content: + application/json: + schema: + type: object + properties: + description: + type: string + description: Optional description + serialized_space: + type: string + description: >- + The contents of the Genie Space in serialized string form + (full replacement). Use the [Get Genie + Space](:method:genie/getspace) API to retrieve an example + response, which includes the `serialized_space` field. This + field provides the structure of the JSON string that + represents the space's layout and components. + title: + type: string + description: Optional title override + warehouse_id: + type: string + description: Optional warehouse override + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieSpace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations: + get: + operationId: genie_list_conversations + summary: Get a list of conversations in a Genie Space. + tags: + - dashboards + - genie + description: |- + Get a list of conversations in a Genie Space. + + :param space_id: str + The ID of the Genie space to retrieve conversations from. + :param include_all: bool (optional) + Include all conversations in the space across all users. Requires at least CAN MANAGE permission on + the space. + :param page_size: int (optional) + Maximum number of conversations to return per page + :param page_token: str (optional) + Token to get the next page of results + + :returns: :class:`GenieListConversationsResponse` + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: The ID of the Genie space to retrieve conversations from. + - name: include_all + in: query + required: false + schema: + type: string + description: >- + Include all conversations in the space across all users. Requires at + least CAN MANAGE permission on the space. + - name: page_size + in: query + required: false + schema: + type: string + description: Maximum number of conversations to return per page + - name: page_token + in: query + required: false + schema: + type: string + description: Token to get the next page of results + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieListConversationsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/conversations/{conversation_id}/messages/{message_id}/feedback: + post: + operationId: genie_send_message_feedback + summary: Send feedback for a message. + tags: + - dashboards + - genie + description: |- + Send feedback for a message. + + :param space_id: str + The ID associated with the Genie space where the message is located. + :param conversation_id: str + The ID associated with the conversation. + :param message_id: str + The ID associated with the message to provide feedback for. + :param rating: :class:`GenieFeedbackRating` + The rating (POSITIVE, NEGATIVE, or NONE). + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: The ID associated with the Genie space where the message is located. + - name: conversation_id + in: path + required: true + schema: + type: string + description: The ID associated with the conversation. + - name: message_id + in: path + required: true + schema: + type: string + description: The ID associated with the message to provide feedback for. + requestBody: + content: + application/json: + schema: + type: object + properties: + rating: + type: string + description: The rating (POSITIVE, NEGATIVE, or NONE). + required: + - rating + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/genie/spaces/{space_id}/start-conversation: + post: + operationId: genie_start_conversation + summary: Start a new conversation. + tags: + - dashboards + - genie + description: |- + Start a new conversation. + + :param space_id: str + The ID associated with the Genie space where you want to start a conversation. + :param content: str + The text of the message that starts the conversation. + + :returns: + Long-running operation waiter for :class:`GenieMessage`. + See :method:wait_get_message_genie_completed for more details. + parameters: + - name: space_id + in: path + required: true + schema: + type: string + description: >- + The ID associated with the Genie space where you want to start a + conversation. + requestBody: + content: + application/json: + schema: + type: object + properties: + content: + type: string + description: The text of the message that starts the conversation. + required: + - content + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GenieMessage' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GenieAPI + /api/2.0/lakeview/dashboards: + post: + operationId: lakeview_create + summary: Create a draft dashboard. + tags: + - dashboards + - lakeview + description: |- + Create a draft dashboard. + + :param dashboard: :class:`Dashboard` + :param dataset_catalog: str (optional) + Sets the default catalog for all datasets in this dashboard. Does not impact table references that + use fully qualified catalog names (ex: samples.nyctaxi.trips). Leave blank to keep each dataset’s + existing configuration. + :param dataset_schema: str (optional) + Sets the default schema for all datasets in this dashboard. Does not impact table references that + use fully qualified schema names (ex: nyctaxi.trips). Leave blank to keep each dataset’s existing + configuration. + + :returns: :class:`Dashboard` + parameters: + - name: dataset_catalog + in: query + required: false + schema: + type: string + - name: dataset_schema + in: query + required: false + schema: + type: string + description: >- + Sets the default schema for all datasets in this dashboard. Does not + impact table references that use fully qualified schema names (ex: + nyctaxi.trips). Leave blank to keep each dataset’s existing + configuration. + requestBody: + content: + application/json: + schema: + type: object + properties: + dashboard: + type: string + description: >- + :param dataset_catalog: str (optional) Sets the default + catalog for all datasets in this dashboard. Does not impact + table references that use fully qualified catalog names (ex: + samples.nyctaxi.trips). Leave blank to keep each dataset’s + existing configuration. + required: + - dashboard + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Dashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + get: + operationId: lakeview_list + summary: List dashboards. + tags: + - dashboards + - lakeview + description: |- + List dashboards. + + :param page_size: int (optional) + The number of dashboards to return per page. + :param page_token: str (optional) + A page token, received from a previous `ListDashboards` call. This token can be used to retrieve the + subsequent page. + :param show_trashed: bool (optional) + The flag to include dashboards located in the trash. If unspecified, only active dashboards will be + returned. + :param view: :class:`DashboardView` (optional) + `DASHBOARD_VIEW_BASIC`only includes summary metadata from the dashboard. + + :returns: Iterator over :class:`Dashboard` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: The number of dashboards to return per page. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous `ListDashboards` call. This + token can be used to retrieve the subsequent page. + - name: show_trashed + in: query + required: false + schema: + type: string + description: >- + The flag to include dashboards located in the trash. If unspecified, + only active dashboards will be returned. + - name: view + in: query + required: false + schema: + type: string + description: >- + `DASHBOARD_VIEW_BASIC`only includes summary metadata from the + dashboard. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListDashboardsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/{dashboard_id}/schedules: + post: + operationId: lakeview_create_schedule + summary: Create dashboard schedule. + tags: + - dashboards + - lakeview + description: |- + Create dashboard schedule. + + :param dashboard_id: str + UUID identifying the dashboard to which the schedule belongs. + :param schedule: :class:`Schedule` + The schedule to create. A dashboard is limited to 10 schedules. + + :returns: :class:`Schedule` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard to which the schedule belongs. + requestBody: + content: + application/json: + schema: + type: object + properties: + schedule: + type: string + description: >- + The schedule to create. A dashboard is limited to 10 + schedules. + required: + - schedule + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Schedule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + get: + operationId: lakeview_list_schedules + summary: List dashboard schedules. + tags: + - dashboards + - lakeview + description: |- + List dashboard schedules. + + :param dashboard_id: str + UUID identifying the dashboard to which the schedules belongs. + :param page_size: int (optional) + The number of schedules to return per page. + :param page_token: str (optional) + A page token, received from a previous `ListSchedules` call. Use this to retrieve the subsequent + page. + + :returns: Iterator over :class:`Schedule` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard to which the schedules belongs. + - name: page_size + in: query + required: false + schema: + type: string + description: The number of schedules to return per page. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous `ListSchedules` call. Use + this to retrieve the subsequent page. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListSchedulesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/{dashboard_id}/schedules/{schedule_id}/subscriptions: + post: + operationId: lakeview_create_subscription + summary: Create schedule subscription. + tags: + - dashboards + - lakeview + description: |- + Create schedule subscription. + + :param dashboard_id: str + UUID identifying the dashboard to which the subscription belongs. + :param schedule_id: str + UUID identifying the schedule to which the subscription belongs. + :param subscription: :class:`Subscription` + The subscription to create. A schedule is limited to 100 subscriptions. + + :returns: :class:`Subscription` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard to which the subscription belongs. + - name: schedule_id + in: path + required: true + schema: + type: string + description: UUID identifying the schedule to which the subscription belongs. + requestBody: + content: + application/json: + schema: + type: object + properties: + subscription: + type: string + description: >- + The subscription to create. A schedule is limited to 100 + subscriptions. + required: + - subscription + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Subscription' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + get: + operationId: lakeview_list_subscriptions + summary: List schedule subscriptions. + tags: + - dashboards + - lakeview + description: |- + List schedule subscriptions. + + :param dashboard_id: str + UUID identifying the dashboard which the subscriptions belongs. + :param schedule_id: str + UUID identifying the schedule which the subscriptions belongs. + :param page_size: int (optional) + The number of subscriptions to return per page. + :param page_token: str (optional) + A page token, received from a previous `ListSubscriptions` call. Use this to retrieve the subsequent + page. + + :returns: Iterator over :class:`Subscription` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard which the subscriptions belongs. + - name: schedule_id + in: path + required: true + schema: + type: string + description: UUID identifying the schedule which the subscriptions belongs. + - name: page_size + in: query + required: false + schema: + type: string + description: The number of subscriptions to return per page. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous `ListSubscriptions` call. Use + this to retrieve the subsequent page. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListSubscriptionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/{dashboard_id}/schedules/{schedule_id}: + delete: + operationId: lakeview_delete_schedule + summary: Delete dashboard schedule. + tags: + - dashboards + - lakeview + description: |- + Delete dashboard schedule. + + :param dashboard_id: str + UUID identifying the dashboard to which the schedule belongs. + :param schedule_id: str + UUID identifying the schedule. + :param etag: str (optional) + The etag for the schedule. Optionally, it can be provided to verify that the schedule has not been + modified from its last retrieval. + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard to which the schedule belongs. + - name: schedule_id + in: path + required: true + schema: + type: string + description: UUID identifying the schedule. + - name: etag + in: query + required: false + schema: + type: string + description: >- + The etag for the schedule. Optionally, it can be provided to verify + that the schedule has not been modified from its last retrieval. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + get: + operationId: lakeview_get_schedule + summary: Get dashboard schedule. + tags: + - dashboards + - lakeview + description: |- + Get dashboard schedule. + + :param dashboard_id: str + UUID identifying the dashboard to which the schedule belongs. + :param schedule_id: str + UUID identifying the schedule. + + :returns: :class:`Schedule` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard to which the schedule belongs. + - name: schedule_id + in: path + required: true + schema: + type: string + description: UUID identifying the schedule. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Schedule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + put: + operationId: lakeview_update_schedule + summary: Update dashboard schedule. + tags: + - dashboards + - lakeview + description: |- + Update dashboard schedule. + + :param dashboard_id: str + UUID identifying the dashboard to which the schedule belongs. + :param schedule_id: str + UUID identifying the schedule. + :param schedule: :class:`Schedule` + The schedule to update. + + :returns: :class:`Schedule` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard to which the schedule belongs. + - name: schedule_id + in: path + required: true + schema: + type: string + description: UUID identifying the schedule. + requestBody: + content: + application/json: + schema: + type: object + properties: + schedule: + type: string + description: The schedule to update. + required: + - schedule + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Schedule' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/{dashboard_id}/schedules/{schedule_id}/subscriptions/{subscription_id}: + delete: + operationId: lakeview_delete_subscription + summary: Delete schedule subscription. + tags: + - dashboards + - lakeview + description: |- + Delete schedule subscription. + + :param dashboard_id: str + UUID identifying the dashboard which the subscription belongs. + :param schedule_id: str + UUID identifying the schedule which the subscription belongs. + :param subscription_id: str + UUID identifying the subscription. + :param etag: str (optional) + The etag for the subscription. Can be optionally provided to ensure that the subscription has not + been modified since the last read. + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard which the subscription belongs. + - name: schedule_id + in: path + required: true + schema: + type: string + description: UUID identifying the schedule which the subscription belongs. + - name: subscription_id + in: path + required: true + schema: + type: string + description: UUID identifying the subscription. + - name: etag + in: query + required: false + schema: + type: string + description: >- + The etag for the subscription. Can be optionally provided to ensure + that the subscription has not been modified since the last read. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + get: + operationId: lakeview_get_subscription + summary: Get schedule subscription. + tags: + - dashboards + - lakeview + description: |- + Get schedule subscription. + + :param dashboard_id: str + UUID identifying the dashboard which the subscription belongs. + :param schedule_id: str + UUID identifying the schedule which the subscription belongs. + :param subscription_id: str + UUID identifying the subscription. + + :returns: :class:`Subscription` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard which the subscription belongs. + - name: schedule_id + in: path + required: true + schema: + type: string + description: UUID identifying the schedule which the subscription belongs. + - name: subscription_id + in: path + required: true + schema: + type: string + description: UUID identifying the subscription. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Subscription' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/{dashboard_id}: + get: + operationId: lakeview_get + summary: Get a draft dashboard. + tags: + - dashboards + - lakeview + description: |- + Get a draft dashboard. + + :param dashboard_id: str + UUID identifying the dashboard. + + :returns: :class:`Dashboard` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Dashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + delete: + operationId: lakeview_trash + summary: Trash a dashboard. + tags: + - dashboards + - lakeview + description: |- + Trash a dashboard. + + :param dashboard_id: str + UUID identifying the dashboard. + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + patch: + operationId: lakeview_update + summary: Update a draft dashboard. + tags: + - dashboards + - lakeview + description: |- + Update a draft dashboard. + + :param dashboard_id: str + UUID identifying the dashboard. + :param dashboard: :class:`Dashboard` + :param dataset_catalog: str (optional) + Sets the default catalog for all datasets in this dashboard. Does not impact table references that + use fully qualified catalog names (ex: samples.nyctaxi.trips). Leave blank to keep each dataset’s + existing configuration. + :param dataset_schema: str (optional) + Sets the default schema for all datasets in this dashboard. Does not impact table references that + use fully qualified schema names (ex: nyctaxi.trips). Leave blank to keep each dataset’s existing + configuration. + + :returns: :class:`Dashboard` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard. + - name: dataset_catalog + in: query + required: false + schema: + type: string + - name: dataset_schema + in: query + required: false + schema: + type: string + description: >- + Sets the default schema for all datasets in this dashboard. Does not + impact table references that use fully qualified schema names (ex: + nyctaxi.trips). Leave blank to keep each dataset’s existing + configuration. + requestBody: + content: + application/json: + schema: + type: object + properties: + dashboard: + type: string + description: >- + :param dataset_catalog: str (optional) Sets the default + catalog for all datasets in this dashboard. Does not impact + table references that use fully qualified catalog names (ex: + samples.nyctaxi.trips). Leave blank to keep each dataset’s + existing configuration. + required: + - dashboard + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Dashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/{dashboard_id}/published: + get: + operationId: lakeview_get_published + summary: Get the current published dashboard. + tags: + - dashboards + - lakeview + description: |- + Get the current published dashboard. + + :param dashboard_id: str + UUID identifying the published dashboard. + + :returns: :class:`PublishedDashboard` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the published dashboard. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PublishedDashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + post: + operationId: lakeview_publish + summary: Publish the current draft dashboard. + tags: + - dashboards + - lakeview + description: |- + Publish the current draft dashboard. + + :param dashboard_id: str + UUID identifying the dashboard to be published. + :param embed_credentials: bool (optional) + Flag to indicate if the publisher's credentials should be embedded in the published dashboard. These + embedded credentials will be used to execute the published dashboard's queries. + :param warehouse_id: str (optional) + The ID of the warehouse that can be used to override the warehouse which was set in the draft. + + :returns: :class:`PublishedDashboard` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the dashboard to be published. + requestBody: + content: + application/json: + schema: + type: object + properties: + embed_credentials: + type: string + description: >- + Flag to indicate if the publisher's credentials should be + embedded in the published dashboard. These embedded + credentials will be used to execute the published + dashboard's queries. + warehouse_id: + type: string + description: >- + The ID of the warehouse that can be used to override the + warehouse which was set in the draft. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PublishedDashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + delete: + operationId: lakeview_unpublish + summary: Unpublish the dashboard. + tags: + - dashboards + - lakeview + description: |- + Unpublish the dashboard. + + :param dashboard_id: str + UUID identifying the published dashboard. + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the published dashboard. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/migrate: + post: + operationId: lakeview_migrate + summary: Migrates a classic SQL dashboard to Lakeview. + tags: + - dashboards + - lakeview + description: |- + Migrates a classic SQL dashboard to Lakeview. + + :param source_dashboard_id: str + UUID of the dashboard to be migrated. + :param display_name: str (optional) + Display name for the new Lakeview dashboard. + :param parent_path: str (optional) + The workspace path of the folder to contain the migrated Lakeview dashboard. + :param update_parameter_syntax: bool (optional) + Flag to indicate if mustache parameter syntax ({{ param }}) should be auto-updated to named syntax + (:param) when converting datasets in the dashboard. + + :returns: :class:`Dashboard` + requestBody: + content: + application/json: + schema: + type: object + properties: + source_dashboard_id: + type: string + description: UUID of the dashboard to be migrated. + display_name: + type: string + description: Display name for the new Lakeview dashboard. + parent_path: + type: string + description: >- + The workspace path of the folder to contain the migrated + Lakeview dashboard. + update_parameter_syntax: + type: string + description: >- + Flag to indicate if mustache parameter syntax ({{ param }}) + should be auto-updated to named syntax (:param) when + converting datasets in the dashboard. + required: + - source_dashboard_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Dashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewAPI + /api/2.0/lakeview/dashboards/{dashboard_id}/published/tokeninfo: + get: + operationId: lakeview_embedded_get_published_dashboard_token_info + summary: >- + Get a required authorization details and scopes of a published dashboard + to mint an OAuth token. + tags: + - dashboards + - lakeview_embedded + description: >- + Get a required authorization details and scopes of a published dashboard + to mint an OAuth token. + + + :param dashboard_id: str + UUID identifying the published dashboard. + :param external_value: str (optional) + Provided external value to be included in the custom claim. + :param external_viewer_id: str (optional) + Provided external viewer id to be included in the custom claim. + + :returns: :class:`GetPublishedDashboardTokenInfoResponse` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: UUID identifying the published dashboard. + - name: external_value + in: query + required: false + schema: + type: string + description: Provided external value to be included in the custom claim. + - name: external_viewer_id + in: query + required: false + schema: + type: string + description: Provided external viewer id to be included in the custom claim. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPublishedDashboardTokenInfoResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LakeviewEmbeddedAPI +components: + schemas: + AuthorizationDetails: + type: object + properties: + grant_rules: + type: array + items: + $ref: '#/components/schemas/AuthorizationDetailsGrantRule' + resource_legacy_acl_path: + type: string + description: The acl path of the tree store resource resource. + resource_name: + type: string + description: >- + The resource name to which the authorization rule applies. This + field is specific to `workspace_rule_set` constraint. Format: + `workspaces/{workspace_id}/dashboards/{dashboard_id}` + type: + type: string + description: >- + The type of authorization downscoping policy. Ex: + `workspace_rule_set` defines access rules for a specific workspace + resource + AuthorizationDetailsGrantRule: + type: object + properties: + permission_set: + type: string + CronSchedule: + type: object + properties: + quartz_cron_expression: + type: string + timezone_id: + type: string + description: >- + A Java timezone id. The schedule will be resolved with respect to + this timezone. See [Java TimeZone] for details. [Java TimeZone]: + https://docs.oracle.com/javase/7/docs/api/java/util/TimeZone.html + required: + - quartz_cron_expression + - timezone_id + Dashboard: + type: object + properties: + create_time: + type: string + dashboard_id: + type: string + description: UUID identifying the dashboard. + display_name: + type: string + description: The display name of the dashboard. + etag: + type: string + description: >- + The etag for the dashboard. Can be optionally provided on updates to + ensure that the dashboard has not been modified since the last read. + This field is excluded in List Dashboards responses. + lifecycle_state: + $ref: '#/components/schemas/LifecycleState' + description: >- + The state of the dashboard resource. Used for tracking trashed + status. + parent_path: + type: string + description: >- + The workspace path of the folder containing the dashboard. Includes + leading slash and no trailing slash. This field is excluded in List + Dashboards responses. + path: + type: string + description: >- + The workspace path of the dashboard asset, including the file name. + Exported dashboards always have the file extension `.lvdash.json`. + This field is excluded in List Dashboards responses. + serialized_dashboard: + type: string + description: >- + The contents of the dashboard in serialized string form. This field + is excluded in List Dashboards responses. Use the [get dashboard + API] to retrieve an example response, which includes the + `serialized_dashboard` field. This field provides the structure of + the JSON string that represents the dashboard's layout and + components. [get dashboard API]: + https://docs.databricks.com/api/workspace/lakeview/get + update_time: + type: string + description: >- + The timestamp of when the dashboard was last updated by the user. + This field is excluded in List Dashboards responses. + warehouse_id: + type: string + description: The warehouse ID used to run the dashboard. + GenieAttachment: + type: object + properties: + attachment_id: + type: string + description: Attachment ID + query: + $ref: '#/components/schemas/GenieQueryAttachment' + description: Query Attachment if Genie responds with a SQL query + suggested_questions: + $ref: '#/components/schemas/GenieSuggestedQuestionsAttachment' + description: Follow-up questions suggested by Genie + text: + $ref: '#/components/schemas/TextAttachment' + description: >- + Text Attachment if Genie responds with text. This also contains the + final summary when available. + description: Genie AI Response + GenieConversation: + type: object + properties: + id: + type: string + space_id: + type: string + description: Genie space ID + user_id: + type: integer + description: ID of the user who created the conversation + title: + type: string + description: Conversation title + conversation_id: + type: string + description: Conversation ID + created_timestamp: + type: integer + description: Timestamp when the message was created + last_updated_timestamp: + type: integer + description: Timestamp when the message was last updated + required: + - id + - space_id + - user_id + - title + - conversation_id + GenieConversationSummary: + type: object + properties: + conversation_id: + type: string + title: + type: string + created_timestamp: + type: integer + required: + - conversation_id + - title + - created_timestamp + GenieFeedback: + type: object + properties: + rating: + $ref: '#/components/schemas/GenieFeedbackRating' + description: The feedback rating + description: Feedback containing rating and optional comment + GenieGenerateDownloadFullQueryResultResponse: + type: object + properties: + download_id: + type: string + download_id_signature: + type: string + description: >- + JWT signature for the download_id to ensure secure access to query + results + GenieGetDownloadFullQueryResultResponse: + type: object + properties: + statement_response: + type: string + GenieGetMessageQueryResultResponse: + type: object + properties: + statement_response: + type: string + GenieListConversationMessagesResponse: + type: object + properties: + messages: + type: array + items: + $ref: '#/components/schemas/GenieMessage' + next_page_token: + type: string + description: The token to use for retrieving the next page of results. + GenieListConversationsResponse: + type: object + properties: + conversations: + type: array + items: + $ref: '#/components/schemas/GenieConversationSummary' + next_page_token: + type: string + description: Token to get the next page of results + GenieListSpacesResponse: + type: object + properties: + next_page_token: + type: string + spaces: + type: array + items: + $ref: '#/components/schemas/GenieSpace' + description: List of Genie spaces + GenieMessage: + type: object + properties: + id: + type: string + space_id: + type: string + description: Genie space ID + conversation_id: + type: string + description: Conversation ID + content: + type: string + description: User message content + message_id: + type: string + description: Message ID + attachments: + type: array + items: + $ref: '#/components/schemas/GenieAttachment' + description: AI-generated response to the message + created_timestamp: + type: integer + description: Timestamp when the message was created + error: + $ref: '#/components/schemas/MessageError' + description: Error message if Genie failed to respond to the message + feedback: + $ref: '#/components/schemas/GenieFeedback' + description: User feedback for the message if provided + last_updated_timestamp: + type: integer + description: Timestamp when the message was last updated + query_result: + $ref: '#/components/schemas/Result' + description: >- + The result of SQL query if the message includes a query attachment. + Deprecated. Use `query_result_metadata` in `GenieQueryAttachment` + instead. + status: + $ref: '#/components/schemas/MessageStatus' + user_id: + type: integer + description: ID of the user who created the message + required: + - id + - space_id + - conversation_id + - content + - message_id + GenieQueryAttachment: + type: object + properties: + description: + type: string + id: + type: string + last_updated_timestamp: + type: integer + description: Time when the user updated the query last + parameters: + type: array + items: + $ref: '#/components/schemas/QueryAttachmentParameter' + query: + type: string + description: AI generated SQL query + query_result_metadata: + $ref: '#/components/schemas/GenieResultMetadata' + description: Metadata associated with the query result. + statement_id: + type: string + description: >- + Statement Execution API statement id. Use [Get status, manifest, and + result first chunk](:method:statementexecution/getstatement) to get + the full result data. + title: + type: string + description: Name of the query + GenieResultMetadata: + type: object + properties: + is_truncated: + type: boolean + row_count: + type: integer + description: The number of rows in the result set. + GenieSpace: + type: object + properties: + space_id: + type: string + title: + type: string + description: Title of the Genie Space + description: + type: string + description: Description of the Genie Space + serialized_space: + type: string + description: >- + The contents of the Genie Space in serialized string form. This + field is excluded in List Genie spaces responses. Use the [Get Genie + Space](:method:genie/getspace) API to retrieve an example response, + which includes the `serialized_space` field. This field provides the + structure of the JSON string that represents the space's layout and + components. + warehouse_id: + type: string + description: Warehouse associated with the Genie Space + required: + - space_id + - title + GenieStartConversationResponse: + type: object + properties: + message_id: + type: string + conversation_id: + type: string + description: Conversation ID + conversation: + $ref: '#/components/schemas/GenieConversation' + message: + $ref: '#/components/schemas/GenieMessage' + required: + - message_id + - conversation_id + GenieSuggestedQuestionsAttachment: + type: object + properties: + questions: + type: array + items: + type: string + description: The suggested follow-up questions + description: Follow-up questions suggested by Genie + GetPublishedDashboardTokenInfoResponse: + type: object + properties: + authorization_details: + type: array + items: + $ref: '#/components/schemas/AuthorizationDetails' + custom_claim: + type: string + description: >- + Custom claim generated from external_value and external_viewer_id. + Format: + `urn:aibi:external_data:::` + scope: + type: string + description: Scope defining access permissions. + ListDashboardsResponse: + type: object + properties: + dashboards: + type: array + items: + $ref: '#/components/schemas/Dashboard' + next_page_token: + type: string + description: >- + A token, which can be sent as `page_token` to retrieve the next + page. If this field is omitted, there are no subsequent dashboards. + ListSchedulesResponse: + type: object + properties: + next_page_token: + type: string + schedules: + type: array + items: + $ref: '#/components/schemas/Schedule' + ListSubscriptionsResponse: + type: object + properties: + next_page_token: + type: string + subscriptions: + type: array + items: + $ref: '#/components/schemas/Subscription' + MessageError: + type: object + properties: + error: + type: string + type: + $ref: '#/components/schemas/MessageErrorType' + PublishedDashboard: + type: object + properties: + display_name: + type: string + embed_credentials: + type: boolean + description: >- + Indicates whether credentials are embedded in the published + dashboard. + revision_create_time: + type: string + description: The timestamp of when the published dashboard was last revised. + warehouse_id: + type: string + description: The warehouse ID used to run the published dashboard. + QueryAttachmentParameter: + type: object + properties: + keyword: + type: string + sql_type: + type: string + value: + type: string + Result: + type: object + properties: + is_truncated: + type: boolean + row_count: + type: integer + description: Row count of the result + statement_id: + type: string + description: >- + Statement Execution API statement id. Use [Get status, manifest, and + result first chunk](:method:statementexecution/getstatement) to get + the full result data. + statement_id_signature: + type: string + description: JWT corresponding to the statement contained in this result + Schedule: + type: object + properties: + cron_schedule: + $ref: '#/components/schemas/CronSchedule' + create_time: + type: string + description: A timestamp indicating when the schedule was created. + dashboard_id: + type: string + description: UUID identifying the dashboard to which the schedule belongs. + display_name: + type: string + description: The display name for schedule. + etag: + type: string + description: >- + The etag for the schedule. Must be left empty on create, must be + provided on updates to ensure that the schedule has not been + modified since the last read, and can be optionally provided on + delete. + pause_status: + $ref: '#/components/schemas/SchedulePauseStatus' + description: The status indicates whether this schedule is paused or not. + schedule_id: + type: string + description: UUID identifying the schedule. + update_time: + type: string + description: A timestamp indicating when the schedule was last updated. + warehouse_id: + type: string + description: The warehouse id to run the dashboard with for the schedule. + required: + - cron_schedule + Subscriber: + type: object + properties: + destination_subscriber: + $ref: '#/components/schemas/SubscriptionSubscriberDestination' + user_subscriber: + $ref: '#/components/schemas/SubscriptionSubscriberUser' + description: >- + The user to receive the subscription email. This parameter is + mutually exclusive with `destination_subscriber`. + Subscription: + type: object + properties: + subscriber: + $ref: '#/components/schemas/Subscriber' + create_time: + type: string + description: A timestamp indicating when the subscription was created. + created_by_user_id: + type: integer + description: >- + UserId of the user who adds subscribers (users or notification + destinations) to the dashboard's schedule. + dashboard_id: + type: string + description: UUID identifying the dashboard to which the subscription belongs. + etag: + type: string + description: >- + The etag for the subscription. Must be left empty on create, can be + optionally provided on delete to ensure that the subscription has + not been deleted since the last read. + schedule_id: + type: string + description: UUID identifying the schedule to which the subscription belongs. + subscription_id: + type: string + description: UUID identifying the subscription. + update_time: + type: string + description: A timestamp indicating when the subscription was last updated. + required: + - subscriber + SubscriptionSubscriberDestination: + type: object + properties: + destination_id: + type: string + required: + - destination_id + SubscriptionSubscriberUser: + type: object + properties: + user_id: + type: integer + required: + - user_id + TextAttachment: + type: object + properties: + content: + type: string + id: + type: string + purpose: + $ref: '#/components/schemas/TextAttachmentPurpose' + description: Purpose/intent of this text attachment + TrashDashboardResponse: + type: object + properties: {} + UnpublishDashboardResponse: + type: object + properties: {} + DashboardView: + type: string + enum: + - DASHBOARD_VIEW_BASIC + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + GenieFeedbackRating: + type: string + enum: + - NEGATIVE + - NONE + - POSITIVE + description: Feedback rating for Genie messages + LifecycleState: + type: string + enum: + - ACTIVE + - TRASHED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MessageErrorType: + type: string + enum: + - BLOCK_MULTIPLE_EXECUTIONS_EXCEPTION + - CHAT_COMPLETION_CLIENT_EXCEPTION + - CHAT_COMPLETION_CLIENT_TIMEOUT_EXCEPTION + - CHAT_COMPLETION_NETWORK_EXCEPTION + - CONTENT_FILTER_EXCEPTION + - CONTEXT_EXCEEDED_EXCEPTION + - COULD_NOT_GET_MODEL_DEPLOYMENTS_EXCEPTION + - COULD_NOT_GET_UC_SCHEMA_EXCEPTION + - DEPLOYMENT_NOT_FOUND_EXCEPTION + - DESCRIBE_QUERY_INVALID_SQL_ERROR + - DESCRIBE_QUERY_TIMEOUT + - DESCRIBE_QUERY_UNEXPECTED_FAILURE + - EXCEEDED_MAX_TOKEN_LENGTH_EXCEPTION + - FUNCTIONS_NOT_AVAILABLE_EXCEPTION + - FUNCTION_ARGUMENTS_INVALID_EXCEPTION + - FUNCTION_ARGUMENTS_INVALID_JSON_EXCEPTION + - FUNCTION_ARGUMENTS_INVALID_TYPE_EXCEPTION + - FUNCTION_CALL_MISSING_PARAMETER_EXCEPTION + - GENERATED_SQL_QUERY_TOO_LONG_EXCEPTION + - GENERIC_CHAT_COMPLETION_EXCEPTION + - GENERIC_CHAT_COMPLETION_SERVICE_EXCEPTION + - GENERIC_SQL_EXEC_API_CALL_EXCEPTION + - ILLEGAL_PARAMETER_DEFINITION_EXCEPTION + - INTERNAL_CATALOG_ASSET_CREATION_FAILED_EXCEPTION + - INTERNAL_CATALOG_ASSET_CREATION_ONGOING_EXCEPTION + - INTERNAL_CATALOG_ASSET_CREATION_UNSUPPORTED_EXCEPTION + - INTERNAL_CATALOG_MISSING_UC_PATH_EXCEPTION + - INTERNAL_CATALOG_PATH_OVERLAP_EXCEPTION + - INVALID_CERTIFIED_ANSWER_FUNCTION_EXCEPTION + - INVALID_CERTIFIED_ANSWER_IDENTIFIER_EXCEPTION + - INVALID_CHAT_COMPLETION_JSON_EXCEPTION + - INVALID_COMPLETION_REQUEST_EXCEPTION + - INVALID_FUNCTION_CALL_EXCEPTION + - INVALID_SQL_MULTIPLE_DATASET_REFERENCES_EXCEPTION + - INVALID_SQL_MULTIPLE_STATEMENTS_EXCEPTION + - INVALID_SQL_UNKNOWN_TABLE_EXCEPTION + - INVALID_TABLE_IDENTIFIER_EXCEPTION + - LOCAL_CONTEXT_EXCEEDED_EXCEPTION + - MESSAGE_ATTACHMENT_TOO_LONG_ERROR + - MESSAGE_CANCELLED_WHILE_EXECUTING_EXCEPTION + - MESSAGE_DELETED_WHILE_EXECUTING_EXCEPTION + - MESSAGE_UPDATED_WHILE_EXECUTING_EXCEPTION + - MISSING_SQL_QUERY_EXCEPTION + - NO_DEPLOYMENTS_AVAILABLE_TO_WORKSPACE + - NO_QUERY_TO_VISUALIZE_EXCEPTION + - NO_TABLES_TO_QUERY_EXCEPTION + - RATE_LIMIT_EXCEEDED_GENERIC_EXCEPTION + - RATE_LIMIT_EXCEEDED_SPECIFIED_WAIT_EXCEPTION + - REPLY_PROCESS_TIMEOUT_EXCEPTION + - RETRYABLE_PROCESSING_EXCEPTION + - SQL_EXECUTION_EXCEPTION + - STOP_PROCESS_DUE_TO_AUTO_REGENERATE + - TABLES_MISSING_EXCEPTION + - TOO_MANY_CERTIFIED_ANSWERS_EXCEPTION + - TOO_MANY_TABLES_EXCEPTION + - UNEXPECTED_REPLY_PROCESS_EXCEPTION + - UNKNOWN_AI_MODEL + - UNSUPPORTED_CONVERSATION_TYPE_EXCEPTION + - WAREHOUSE_ACCESS_MISSING_EXCEPTION + - WAREHOUSE_NOT_FOUND_EXCEPTION + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MessageStatus: + type: string + enum: + - ASKING_AI + - CANCELLED + - COMPLETED + - EXECUTING_QUERY + - FAILED + - FETCHING_METADATA + - FILTERING_CONTEXT + - PENDING_WAREHOUSE + - QUERY_RESULT_EXPIRED + - SUBMITTED + description: >- + MessageStatus. The possible values are: * `FETCHING_METADATA`: Fetching + metadata from the data + + sources. * `FILTERING_CONTEXT`: Running smart context step to determine + relevant context. * + + `ASKING_AI`: Waiting for the LLM to respond to the user's question. * + `PENDING_WAREHOUSE`: + + Waiting for warehouse before the SQL query can start executing. * + `EXECUTING_QUERY`: Executing a + + generated SQL query. Get the SQL query result by calling + + [getMessageAttachmentQueryResult](:method:genie/getMessageAttachmentQueryResult) + API. * + + `FAILED`: The response generation or query execution failed. See `error` + field. * `COMPLETED`: + + Message processing is completed. Results are in the `attachments` field. + Get the SQL query + + result by calling + + [getMessageAttachmentQueryResult](:method:genie/getMessageAttachmentQueryResult) + API. * + + `SUBMITTED`: Message has been submitted. * `QUERY_RESULT_EXPIRED`: SQL + result is not available + + anymore. The user needs to rerun the query. Rerun the SQL query result + by calling + + [executeMessageAttachmentQuery](:method:genie/executeMessageAttachmentQuery) + API. * `CANCELLED`: + + Message has been cancelled. + SchedulePauseStatus: + type: string + enum: + - PAUSED + - UNPAUSED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TextAttachmentPurpose: + type: string + enum: + - FOLLOW_UP_QUESTION + description: Purpose/intent of a text attachment + x-stackQL-resources: + genie_messages: + id: databricks_workspace.dashboards.genie_messages + name: genie_messages + title: Genie Messages + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages/get + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + execute_query: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1execute-query/post + response: + mediaType: application/json + openAPIDocKey: '200' + get_query_result_deprecated: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1query-result/get + response: + mediaType: application/json + openAPIDocKey: '200' + get_query_result_by_attachment: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1query-result~1{attachment_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + send_feedback: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1feedback/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/genie_messages/methods/get_query_result_by_attachment + - $ref: '#/components/x-stackQL-resources/genie_messages/methods/get' + - $ref: '#/components/x-stackQL-resources/genie_messages/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/genie_messages/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/genie_messages/methods/delete' + replace: [] + genie: + id: databricks_workspace.dashboards.genie + name: genie + title: Genie + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1genie~1spaces/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1genie~1spaces/get' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1genie~1spaces~1{space_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1genie~1spaces~1{space_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1genie~1spaces~1{space_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/genie/methods/get' + - $ref: '#/components/x-stackQL-resources/genie/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/genie/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/genie/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/genie/methods/delete' + replace: [] + genie_conversations: + id: databricks_workspace.dashboards.genie_conversations + name: genie_conversations + title: Genie Conversations + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations/get' + response: + mediaType: application/json + openAPIDocKey: '200' + start: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1start-conversation/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/genie_conversations/methods/list' + insert: [] + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/genie_conversations/methods/delete + replace: [] + genie_message_attachments: + id: databricks_workspace.dashboards.genie_message_attachments + name: genie_message_attachments + title: Genie Message Attachments + methods: + execute_attachment_query: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1attachments~1{attachment_id}~1execute-query/post + response: + mediaType: application/json + openAPIDocKey: '200' + get_query_result: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1attachments~1{attachment_id}~1query-result/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/genie_message_attachments/methods/get_query_result + insert: [] + update: [] + delete: [] + replace: [] + genie_message_attachment_downloads: + id: databricks_workspace.dashboards.genie_message_attachment_downloads + name: genie_message_attachment_downloads + title: Genie Message Attachment Downloads + methods: + generate_download: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1attachments~1{attachment_id}~1downloads/post + response: + mediaType: application/json + openAPIDocKey: '200' + get_download: + operation: + $ref: >- + #/paths/~1api~12.0~1genie~1spaces~1{space_id}~1conversations~1{conversation_id}~1messages~1{message_id}~1attachments~1{attachment_id}~1downloads~1{download_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/genie_message_attachment_downloads/methods/get_download + insert: [] + update: [] + delete: [] + replace: [] + lakeview: + id: databricks_workspace.dashboards.lakeview + name: lakeview + title: Lakeview + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1lakeview~1dashboards/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1lakeview~1dashboards/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.dashboards + update: + operation: + $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + migrate: + operation: + $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1migrate/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/lakeview/methods/get' + - $ref: '#/components/x-stackQL-resources/lakeview/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/lakeview/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/lakeview/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/lakeview/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/lakeview/methods/update' + lakeview_schedules: + id: databricks_workspace.dashboards.lakeview_schedules + name: lakeview_schedules + title: Lakeview Schedules + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.schedules + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/lakeview_schedules/methods/get' + - $ref: '#/components/x-stackQL-resources/lakeview_schedules/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/lakeview_schedules/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/lakeview_schedules/methods/delete' + replace: [] + lakeview_subscriptions: + id: databricks_workspace.dashboards.lakeview_subscriptions + name: lakeview_subscriptions + title: Lakeview Subscriptions + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.subscriptions + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions~1{subscription_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions~1{subscription_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/lakeview_subscriptions/methods/get + - $ref: >- + #/components/x-stackQL-resources/lakeview_subscriptions/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/lakeview_subscriptions/methods/create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/lakeview_subscriptions/methods/delete + replace: [] + lakeview_published: + id: databricks_workspace.dashboards.lakeview_published + name: lakeview_published + title: Lakeview Published + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1published/get + response: + mediaType: application/json + openAPIDocKey: '200' + publish: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1published/post + response: + mediaType: application/json + openAPIDocKey: '200' + unpublish: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1published/delete + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/lakeview_published/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + lakeview_embedded: + id: databricks_workspace.dashboards.lakeview_embedded + name: lakeview_embedded + title: Lakeview Embedded + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1published~1tokeninfo/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/lakeview_embedded/methods/get' + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/database.yaml b/providers/src/databricks_workspace/v00.00.00000/services/database.yaml new file mode 100644 index 00000000..b4b80138 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/database.yaml @@ -0,0 +1,2329 @@ +openapi: 3.0.0 +info: + title: Databricks Database API (workspace) + description: >- + OpenAPI specification for the Databricks database service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.database +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/database/catalogs: + post: + operationId: database_create_database_catalog + summary: Create a Database Catalog. + tags: + - database + description: |- + Create a Database Catalog. + + :param catalog: :class:`DatabaseCatalog` + + :returns: :class:`DatabaseCatalog` + requestBody: + content: + application/json: + schema: + type: object + properties: + catalog: + type: string + description: ':returns: :class:`DatabaseCatalog`' + required: + - catalog + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseCatalog' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/instances: + post: + operationId: database_create_database_instance + summary: Create a Database Instance. + tags: + - database + description: |- + Create a Database Instance. + + :param database_instance: :class:`DatabaseInstance` + Instance to create. + + :returns: + Long-running operation waiter for :class:`DatabaseInstance`. + See :method:wait_get_database_instance_database_available for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + database_instance: + type: string + description: Instance to create. + required: + - database_instance + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseInstance' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + get: + operationId: database_list_database_instances + summary: List Database Instances. + tags: + - database + description: |- + List Database Instances. + + :param page_size: int (optional) + Upper bound for items returned. + :param page_token: str (optional) + Pagination token to go to the next page of Database Instances. Requests first page if absent. + + :returns: Iterator over :class:`DatabaseInstance` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of Database Instances. + Requests first page if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListDatabaseInstancesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/instances/{instance_name}/roles: + post: + operationId: database_create_database_instance_role + summary: Create a role for a Database Instance. + tags: + - database + description: |- + Create a role for a Database Instance. + + :param instance_name: str + :param database_instance_role: :class:`DatabaseInstanceRole` + :param database_instance_name: str (optional) + + :returns: :class:`DatabaseInstanceRole` + parameters: + - name: instance_name + in: path + required: true + schema: + type: string + description: ':param database_instance_role: :class:`DatabaseInstanceRole`' + - name: database_instance_name + in: query + required: false + schema: + type: string + description: ':returns: :class:`DatabaseInstanceRole`' + requestBody: + content: + application/json: + schema: + type: object + properties: + database_instance_role: + type: string + required: + - database_instance_role + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseInstanceRole' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + get: + operationId: database_list_database_instance_roles + summary: >- + START OF PG ROLE APIs Section These APIs are marked a PUBLIC with stage + < PUBLIC_PREVIEW. With more + tags: + - database + description: >- + START OF PG ROLE APIs Section These APIs are marked a PUBLIC with stage + < PUBLIC_PREVIEW. With more + + recent Lakebase V2 plans, we don't plan to ever advance these to + PUBLIC_PREVIEW. These APIs will + + remain effectively undocumented/UI-only and we'll aim for a new public + roles API as part of V2 PuPr. + + + :param instance_name: str + + :param page_size: int (optional) + Upper bound for items returned. + :param page_token: str (optional) + Pagination token to go to the next page of Database Instances. Requests first page if absent. + + :returns: Iterator over :class:`DatabaseInstanceRole` + parameters: + - name: instance_name + in: path + required: true + schema: + type: string + description: ':param page_size: int (optional) Upper bound for items returned.' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of Database Instances. + Requests first page if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListDatabaseInstanceRolesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/tables: + post: + operationId: database_create_database_table + summary: >- + Create a Database Table. Useful for registering pre-existing PG tables + in UC. See + tags: + - database + description: >- + Create a Database Table. Useful for registering pre-existing PG tables + in UC. See + + CreateSyncedDatabaseTable for creating synced tables in PG from a source + table in UC. + + + :param table: :class:`DatabaseTable` + + + :returns: :class:`DatabaseTable` + requestBody: + content: + application/json: + schema: + type: object + properties: + table: + type: string + description: ':returns: :class:`DatabaseTable`' + required: + - table + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseTable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/synced_tables: + post: + operationId: database_create_synced_database_table + summary: Create a Synced Database Table. + tags: + - database + description: |- + Create a Synced Database Table. + + :param synced_table: :class:`SyncedDatabaseTable` + + :returns: :class:`SyncedDatabaseTable` + requestBody: + content: + application/json: + schema: + type: object + properties: + synced_table: + type: string + description: ':returns: :class:`SyncedDatabaseTable`' + required: + - synced_table + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SyncedDatabaseTable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/catalogs/{name}: + delete: + operationId: database_delete_database_catalog + summary: Delete a Database Catalog. + tags: + - database + description: |- + Delete a Database Catalog. + + :param name: str + parameters: + - name: name + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + get: + operationId: database_get_database_catalog + summary: Get a Database Catalog. + tags: + - database + description: |- + Get a Database Catalog. + + :param name: str + + :returns: :class:`DatabaseCatalog` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: ':returns: :class:`DatabaseCatalog`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseCatalog' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + patch: + operationId: database_update_database_catalog + summary: This API is currently unimplemented, but exposed for Terraform support. + tags: + - database + description: |- + This API is currently unimplemented, but exposed for Terraform support. + + :param name: str + The name of the catalog in UC. + :param database_catalog: :class:`DatabaseCatalog` + Note that updating a database catalog is not yet supported. + :param update_mask: str + The list of fields to update. Setting this field is not yet supported. + + :returns: :class:`DatabaseCatalog` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the catalog in UC. + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + The list of fields to update. Setting this field is not yet + supported. + requestBody: + content: + application/json: + schema: + type: object + properties: + database_catalog: + type: string + description: Note that updating a database catalog is not yet supported. + required: + - database_catalog + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseCatalog' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/instances/{name}: + delete: + operationId: database_delete_database_instance + summary: Delete a Database Instance. + tags: + - database + description: |- + Delete a Database Instance. + + :param name: str + Name of the instance to delete. + :param force: bool (optional) + By default, a instance cannot be deleted if it has descendant instances created via PITR. If this + flag is specified as true, all descendent instances will be deleted as well. + :param purge: bool (optional) + Deprecated. Omitting the field or setting it to true will result in the field being hard deleted. + Setting a value of false will throw a bad request. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the instance to delete. + - name: force + in: query + required: false + schema: + type: string + description: >- + By default, a instance cannot be deleted if it has descendant + instances created via PITR. If this flag is specified as true, all + descendent instances will be deleted as well. + - name: purge + in: query + required: false + schema: + type: string + description: >- + Deprecated. Omitting the field or setting it to true will result in + the field being hard deleted. Setting a value of false will throw a + bad request. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + get: + operationId: database_get_database_instance + summary: Get a Database Instance. + tags: + - database + description: |- + Get a Database Instance. + + :param name: str + Name of the cluster to get. + + :returns: :class:`DatabaseInstance` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the cluster to get. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseInstance' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + patch: + operationId: database_update_database_instance + summary: Update a Database Instance. + tags: + - database + description: |- + Update a Database Instance. + + :param name: str + The name of the instance. This is the unique identifier for the instance. + :param database_instance: :class:`DatabaseInstance` + :param update_mask: str + The list of fields to update. If unspecified, all fields will be updated when possible. To wipe out + custom_tags, specify custom_tags in the update_mask with an empty custom_tags map. + + :returns: :class:`DatabaseInstance` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The name of the instance. This is the unique identifier for the + instance. + - name: update_mask + in: query + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + database_instance: + type: string + description: >- + :param update_mask: str The list of fields to update. If + unspecified, all fields will be updated when possible. To + wipe out custom_tags, specify custom_tags in the update_mask + with an empty custom_tags map. + required: + - database_instance + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseInstance' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/instances/{instance_name}/roles/{name}: + delete: + operationId: database_delete_database_instance_role + summary: Deletes a role for a Database Instance. + tags: + - database + description: |- + Deletes a role for a Database Instance. + + :param instance_name: str + :param name: str + :param allow_missing: bool (optional) + This is the AIP standard name for the equivalent of Postgres' `IF EXISTS` option + :param reassign_owned_to: str (optional) + parameters: + - name: instance_name + in: path + required: true + schema: + type: string + description: ':param name: str' + - name: name + in: path + required: true + schema: + type: string + - name: allow_missing + in: query + required: false + schema: + type: string + description: >- + This is the AIP standard name for the equivalent of Postgres' `IF + EXISTS` option + - name: reassign_owned_to + in: query + required: false + schema: + type: string + description: str (optional) + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + get: + operationId: database_get_database_instance_role + summary: Gets a role for a Database Instance. + tags: + - database + description: |- + Gets a role for a Database Instance. + + :param instance_name: str + :param name: str + + :returns: :class:`DatabaseInstanceRole` + parameters: + - name: instance_name + in: path + required: true + schema: + type: string + description: ':param name: str' + - name: name + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseInstanceRole' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/tables/{name}: + delete: + operationId: database_delete_database_table + summary: Delete a Database Table. + tags: + - database + description: |- + Delete a Database Table. + + :param name: str + parameters: + - name: name + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + get: + operationId: database_get_database_table + summary: Get a Database Table. + tags: + - database + description: |- + Get a Database Table. + + :param name: str + + :returns: :class:`DatabaseTable` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: ':returns: :class:`DatabaseTable`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseTable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/synced_tables/{name}: + delete: + operationId: database_delete_synced_database_table + summary: Delete a Synced Database Table. + tags: + - database + description: |- + Delete a Synced Database Table. + + :param name: str + :param purge_data: bool (optional) + Optional. When set to true, the actual PostgreSQL table will be dropped from the database. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + :param purge_data: bool (optional) Optional. When set to true, the + actual PostgreSQL table will be dropped from the database. + - name: purge_data + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + get: + operationId: database_get_synced_database_table + summary: Get a Synced Database Table. + tags: + - database + description: |- + Get a Synced Database Table. + + :param name: str + + :returns: :class:`SyncedDatabaseTable` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: ':returns: :class:`SyncedDatabaseTable`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SyncedDatabaseTable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + patch: + operationId: database_update_synced_database_table + summary: This API is currently unimplemented, but exposed for Terraform support. + tags: + - database + description: |- + This API is currently unimplemented, but exposed for Terraform support. + + :param name: str + Full three-part (catalog, schema, table) name of the table. + :param synced_table: :class:`SyncedDatabaseTable` + Note that updating a synced database table is not yet supported. + :param update_mask: str + The list of fields to update. Setting this field is not yet supported. + + :returns: :class:`SyncedDatabaseTable` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Full three-part (catalog, schema, table) name of the table. + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + The list of fields to update. Setting this field is not yet + supported. + requestBody: + content: + application/json: + schema: + type: object + properties: + synced_table: + type: string + description: >- + Note that updating a synced database table is not yet + supported. + required: + - synced_table + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SyncedDatabaseTable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/instances:findByUid: + get: + operationId: database_find_database_instance_by_uid + summary: Find a Database Instance by uid. + tags: + - database + description: |- + Find a Database Instance by uid. + + :param uid: str (optional) + UID of the cluster to get. + + :returns: :class:`DatabaseInstance` + parameters: + - name: uid + in: query + required: true + schema: + type: string + description: UID of the cluster to get. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseInstance' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/credentials: + post: + operationId: database_generate_database_credential + summary: Generates a credential that can be used to access database instances. + tags: + - database + description: |- + Generates a credential that can be used to access database instances. + + :param claims: List[:class:`RequestedClaims`] (optional) + The returned token will be scoped to the union of instance_names and instances containing the + specified UC tables, so instance_names is allowed to be empty. + :param instance_names: List[str] (optional) + Instances to which the token will be scoped. + :param request_id: str (optional) + + :returns: :class:`DatabaseCredential` + requestBody: + content: + application/json: + schema: + type: object + properties: + claims: + type: string + description: >- + The returned token will be scoped to the union of + instance_names and instances containing the specified UC + tables, so instance_names is allowed to be empty. + instance_names: + type: string + description: Instances to which the token will be scoped. + request_id: + type: string + description: ':returns: :class:`DatabaseCredential`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseCredential' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/instances/{instance_name}/catalogs: + get: + operationId: database_list_database_catalogs + summary: This API is currently unimplemented, but exposed for Terraform support. + tags: + - database + description: |- + This API is currently unimplemented, but exposed for Terraform support. + + :param instance_name: str + Name of the instance to get database catalogs for. + :param page_size: int (optional) + Upper bound for items returned. + :param page_token: str (optional) + Pagination token to go to the next page of synced database tables. Requests first page if absent. + + :returns: Iterator over :class:`DatabaseCatalog` + parameters: + - name: instance_name + in: path + required: true + schema: + type: string + description: Name of the instance to get database catalogs for. + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of synced database tables. + Requests first page if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListDatabaseCatalogsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI + /api/2.0/database/instances/{instance_name}/synced_tables: + get: + operationId: database_list_synced_database_tables + summary: This API is currently unimplemented, but exposed for Terraform support. + tags: + - database + description: |- + This API is currently unimplemented, but exposed for Terraform support. + + :param instance_name: str + Name of the instance to get synced tables for. + :param page_size: int (optional) + Upper bound for items returned. + :param page_token: str (optional) + Pagination token to go to the next page of synced database tables. Requests first page if absent. + + :returns: Iterator over :class:`SyncedDatabaseTable` + parameters: + - name: instance_name + in: path + required: true + schema: + type: string + description: Name of the instance to get synced tables for. + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of synced database tables. + Requests first page if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListSyncedDatabaseTablesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DatabaseAPI +components: + schemas: + CustomTag: + type: object + properties: + key: + type: string + value: + type: string + description: The value of the custom tag. + DatabaseCatalog: + type: object + properties: + name: + type: string + database_instance_name: + type: string + description: The name of the DatabaseInstance housing the database. + database_name: + type: string + description: >- + The name of the database (in a instance) associated with the + catalog. + create_database_if_not_exists: + type: boolean + uid: + type: string + required: + - name + - database_instance_name + - database_name + DatabaseCredential: + type: object + properties: + expiration_time: + type: string + token: + type: string + DatabaseInstance: + type: object + properties: + name: + type: string + description: >- + The name of the instance. This is the unique identifier for the + instance. + capacity: + type: string + description: >- + The sku of the instance. Valid values are "CU_1", "CU_2", "CU_4", + "CU_8". + child_instance_refs: + type: array + items: + $ref: '#/components/schemas/DatabaseInstanceRef' + description: >- + The refs of the child instances. This is only available if the + instance is parent instance. + creation_time: + type: string + description: The timestamp when the instance was created. + creator: + type: string + description: The email of the creator of the instance. + custom_tags: + type: array + items: + $ref: '#/components/schemas/CustomTag' + description: >- + Custom tags associated with the instance. This field is only + included on create and update responses. + effective_capacity: + type: string + description: >- + Deprecated. The sku of the instance; this field will always match + the value of capacity. This is an output only field that contains + the value computed from the input field combined with server side + defaults. Use the field without the effective_ prefix to set the + value. + effective_custom_tags: + type: array + items: + $ref: '#/components/schemas/CustomTag' + description: >- + The recorded custom tags associated with the instance. This is an + output only field that contains the value computed from the input + field combined with server side defaults. Use the field without the + effective_ prefix to set the value. + effective_enable_pg_native_login: + type: boolean + description: >- + Whether the instance has PG native password login enabled. This is + an output only field that contains the value computed from the input + field combined with server side defaults. Use the field without the + effective_ prefix to set the value. + effective_enable_readable_secondaries: + type: boolean + description: >- + Whether secondaries serving read-only traffic are enabled. Defaults + to false. This is an output only field that contains the value + computed from the input field combined with server side defaults. + Use the field without the effective_ prefix to set the value. + effective_node_count: + type: integer + description: >- + The number of nodes in the instance, composed of 1 primary and 0 or + more secondaries. Defaults to 1 primary and 0 secondaries. This is + an output only field that contains the value computed from the input + field combined with server side defaults. Use the field without the + effective_ prefix to set the value. + effective_retention_window_in_days: + type: integer + description: >- + The retention window for the instance. This is the time window in + days for which the historical data is retained. This is an output + only field that contains the value computed from the input field + combined with server side defaults. Use the field without the + effective_ prefix to set the value. + effective_stopped: + type: boolean + description: >- + Whether the instance is stopped. This is an output only field that + contains the value computed from the input field combined with + server side defaults. Use the field without the effective_ prefix to + set the value. + effective_usage_policy_id: + type: string + description: >- + The policy that is applied to the instance. This is an output only + field that contains the value computed from the input field combined + with server side defaults. Use the field without the effective_ + prefix to set the value. + enable_pg_native_login: + type: boolean + description: >- + Whether to enable PG native password login on the instance. Defaults + to false. + enable_readable_secondaries: + type: boolean + description: >- + Whether to enable secondaries to serve read-only traffic. Defaults + to false. + node_count: + type: integer + description: >- + The number of nodes in the instance, composed of 1 primary and 0 or + more secondaries. Defaults to 1 primary and 0 secondaries. This + field is input only, see effective_node_count for the output. + parent_instance_ref: + $ref: '#/components/schemas/DatabaseInstanceRef' + description: >- + The ref of the parent instance. This is only available if the + instance is child instance. Input: For specifying the parent + instance to create a child instance. Optional. Output: Only + populated if provided as input to create a child instance. + pg_version: + type: string + description: The version of Postgres running on the instance. + read_only_dns: + type: string + description: >- + The DNS endpoint to connect to the instance for read only access. + This is only available if enable_readable_secondaries is true. + read_write_dns: + type: string + description: The DNS endpoint to connect to the instance for read+write access. + retention_window_in_days: + type: integer + description: >- + The retention window for the instance. This is the time window in + days for which the historical data is retained. The default value is + 7 days. Valid values are 2 to 35 days. + state: + $ref: '#/components/schemas/DatabaseInstanceState' + description: The current state of the instance. + stopped: + type: boolean + description: >- + Whether to stop the instance. An input only param, see + effective_stopped for the output. + uid: + type: string + description: An immutable UUID identifier for the instance. + usage_policy_id: + type: string + description: The desired usage policy to associate with the instance. + required: + - name + description: >- + A DatabaseInstance represents a logical Postgres instance, comprised of + both compute and + storage. + DatabaseInstanceRef: + type: object + properties: + branch_time: + type: string + description: >- + Branch time of the ref database instance. For a parent ref instance, + this is the point in time on the parent instance from which the + instance was created. For a child ref instance, this is the point in + time on the instance from which the child instance was created. + Input: For specifying the point in time to create a child instance. + Optional. Output: Only populated if provided as input to create a + child instance. + effective_lsn: + type: string + description: >- + For a parent ref instance, this is the LSN on the parent instance + from which the instance was created. For a child ref instance, this + is the LSN on the instance from which the child instance was + created. This is an output only field that contains the value + computed from the input field combined with server side defaults. + Use the field without the effective_ prefix to set the value. + lsn: + type: string + description: >- + User-specified WAL LSN of the ref database instance. Input: For + specifying the WAL LSN to create a child instance. Optional. Output: + Only populated if provided as input to create a child instance. + name: + type: string + description: Name of the ref database instance. + uid: + type: string + description: Id of the ref database instance. + description: >- + DatabaseInstanceRef is a reference to a database instance. It is used in + the DatabaseInstance + object to refer to the parent instance of an instance and to refer the child instances of an + instance. To specify as a parent instance during creation of an instance, the lsn and + branch_time fields are optional. If not specified, the child instance will be created from the + latest lsn of the parent. If both lsn and branch_time are specified, the lsn will be used to + create the child instance. + DatabaseInstanceRole: + type: object + properties: + name: + type: string + description: >- + The name of the role. This is the unique identifier for the role in + an instance. + attributes: + $ref: '#/components/schemas/DatabaseInstanceRoleAttributes' + description: >- + The desired API-exposed Postgres role attribute to associate with + the role. Optional. + effective_attributes: + $ref: '#/components/schemas/DatabaseInstanceRoleAttributes' + description: >- + The attributes that are applied to the role. This is an output only + field that contains the value computed from the input field combined + with server side defaults. Use the field without the effective_ + prefix to set the value. + identity_type: + $ref: '#/components/schemas/DatabaseInstanceRoleIdentityType' + description: The type of the role. + instance_name: + type: string + membership_role: + $ref: '#/components/schemas/DatabaseInstanceRoleMembershipRole' + description: An enum value for a standard role that this role is a member of. + required: + - name + description: >- + A DatabaseInstanceRole represents a Postgres role in a database + instance. + DatabaseInstanceRoleAttributes: + type: object + properties: + bypassrls: + type: boolean + createdb: + type: boolean + createrole: + type: boolean + description: >- + Attributes that can be granted to a Postgres role. We are only + implementing a subset for now, + see xref: https://www.postgresql.org/docs/16/sql-createrole.html The values follow Postgres + keyword naming e.g. CREATEDB, BYPASSRLS, etc. which is why they don't include typical + underscores between words. We were requested to make this a nested object/struct representation + since these are knobs from an external spec. + DatabaseTable: + type: object + properties: + name: + type: string + description: Full three-part (catalog, schema, table) name of the table. + database_instance_name: + type: string + description: >- + Name of the target database instance. This is required when creating + database tables in standard catalogs. This is optional when creating + database tables in registered catalogs. If this field is specified + when creating database tables in registered catalogs, the database + instance name MUST match that of the registered catalog (or the + request will be rejected). + logical_database_name: + type: string + description: >- + Target Postgres database object (logical database) name for this + table. When creating a table in a standard catalog, this field is + required. In this scenario, specifying this field will allow + targeting an arbitrary postgres database. Registration of database + tables via /database/tables is currently only supported in standard + catalogs. + required: + - name + description: 'Next field marker: 13' + DeltaTableSyncInfo: + type: object + properties: + delta_commit_timestamp: + type: string + delta_commit_version: + type: integer + description: The Delta Lake commit version that was last successfully synced. + ListDatabaseCatalogsResponse: + type: object + properties: + database_catalogs: + type: array + items: + $ref: '#/components/schemas/DatabaseCatalog' + next_page_token: + type: string + description: Pagination token to request the next page of database catalogs. + ListDatabaseInstanceRolesResponse: + type: object + properties: + database_instance_roles: + type: array + items: + $ref: '#/components/schemas/DatabaseInstanceRole' + next_page_token: + type: string + description: Pagination token to request the next page of instances. + ListDatabaseInstancesResponse: + type: object + properties: + database_instances: + type: array + items: + $ref: '#/components/schemas/DatabaseInstance' + next_page_token: + type: string + description: Pagination token to request the next page of instances. + ListSyncedDatabaseTablesResponse: + type: object + properties: + next_page_token: + type: string + synced_tables: + type: array + items: + $ref: '#/components/schemas/SyncedDatabaseTable' + NewPipelineSpec: + type: object + properties: + budget_policy_id: + type: string + description: Budget policy to set on the newly created pipeline. + storage_catalog: + type: string + description: >- + This field needs to be specified if the destination catalog is a + managed postgres catalog. UC catalog for the pipeline to store + intermediate files (checkpoints, event logs etc). This needs to be a + standard catalog where the user has permissions to create Delta + tables. + storage_schema: + type: string + description: >- + This field needs to be specified if the destination catalog is a + managed postgres catalog. UC schema for the pipeline to store + intermediate files (checkpoints, event logs etc). This needs to be + in the standard catalog where the user has permissions to create + Delta tables. + description: >- + Custom fields that user can set for pipeline while creating + SyncedDatabaseTable. Note that other + fields of pipeline are still inferred by table def internally + RequestedClaims: + type: object + properties: + permission_set: + $ref: '#/components/schemas/RequestedClaimsPermissionSet' + resources: + type: array + items: + $ref: '#/components/schemas/RequestedResource' + RequestedResource: + type: object + properties: + table_name: + type: string + unspecified_resource_name: + type: string + SyncedDatabaseTable: + type: object + properties: + name: + type: string + description: Full three-part (catalog, schema, table) name of the table. + data_synchronization_status: + $ref: '#/components/schemas/SyncedTableStatus' + description: Synced Table data synchronization status + database_instance_name: + type: string + description: >- + Name of the target database instance. This is required when creating + synced database tables in standard catalogs. This is optional when + creating synced database tables in registered catalogs. If this + field is specified when creating synced database tables in + registered catalogs, the database instance name MUST match that of + the registered catalog (or the request will be rejected). + effective_database_instance_name: + type: string + description: >- + The name of the database instance that this table is registered to. + This field is always returned, and for tables inside database + catalogs is inferred database instance associated with the catalog. + This is an output only field that contains the value computed from + the input field combined with server side defaults. Use the field + without the effective_ prefix to set the value. + effective_logical_database_name: + type: string + description: >- + The name of the logical database that this table is registered to. + This is an output only field that contains the value computed from + the input field combined with server side defaults. Use the field + without the effective_ prefix to set the value. + logical_database_name: + type: string + description: >- + Target Postgres database object (logical database) name for this + table. When creating a synced table in a registered Postgres + catalog, the target Postgres database name is inferred to be that of + the registered catalog. If this field is specified in this scenario, + the Postgres database name MUST match that of the registered catalog + (or the request will be rejected). When creating a synced table in a + standard catalog, this field is required. In this scenario, + specifying this field will allow targeting an arbitrary postgres + database. Note that this has implications for the + `create_database_objects_is_missing` field in `spec`. + spec: + $ref: '#/components/schemas/SyncedTableSpec' + unity_catalog_provisioning_state: + $ref: '#/components/schemas/ProvisioningInfoState' + description: >- + The provisioning state of the synced table entity in Unity Catalog. + This is distinct from the state of the data synchronization pipeline + (i.e. the table may be in "ACTIVE" but the pipeline may be in + "PROVISIONING" as it runs asynchronously). + required: + - name + description: 'Next field marker: 18' + SyncedTableContinuousUpdateStatus: + type: object + properties: + initial_pipeline_sync_progress: + $ref: '#/components/schemas/SyncedTablePipelineProgress' + description: Progress of the initial data synchronization. + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was successfully synced to + the synced table. + timestamp: + type: string + description: >- + The end timestamp of the last time any data was synchronized from + the source table to the synced table. This is when the data is + available in the synced table. + description: >- + Detailed status of a synced table. Shown if the synced table is in the + SYNCED_CONTINUOUS_UPDATE + or the SYNCED_UPDATING_PIPELINE_RESOURCES state. + SyncedTableFailedStatus: + type: object + properties: + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was successfully synced to + the synced table. The last source table Delta version that was + synced to the synced table. Only populated if the table is still + synced and available for serving. + timestamp: + type: string + description: >- + The end timestamp of the last time any data was synchronized from + the source table to the synced table. Only populated if the table is + still synced and available for serving. + description: >- + Detailed status of a synced table. Shown if the synced table is in the + OFFLINE_FAILED or the + SYNCED_PIPELINE_FAILED state. + SyncedTablePipelineProgress: + type: object + properties: + estimated_completion_time_seconds: + type: number + description: The estimated time remaining to complete this update in seconds. + latest_version_currently_processing: + type: integer + description: >- + The source table Delta version that was last processed by the + pipeline. The pipeline may not have completely processed this + version yet. + provisioning_phase: + $ref: '#/components/schemas/ProvisioningPhase' + description: The current phase of the data synchronization pipeline. + sync_progress_completion: + type: number + description: >- + The completion ratio of this update. This is a number between 0 and + 1. + synced_row_count: + type: integer + description: The number of rows that have been synced in this update. + total_row_count: + type: integer + description: >- + The total number of rows that need to be synced in this update. This + number may be an estimate. + description: Progress information of the Synced Table data synchronization pipeline. + SyncedTablePosition: + type: object + properties: + delta_table_sync_info: + $ref: '#/components/schemas/DeltaTableSyncInfo' + sync_end_timestamp: + type: string + description: >- + The end timestamp of the most recent successful synchronization. + This is the time when the data is available in the synced table. + sync_start_timestamp: + type: string + description: >- + The starting timestamp of the most recent successful synchronization + from the source table to the destination (synced) table. Note this + is the starting timestamp of the sync operation, not the end time. + E.g., for a batch, this is the time when the sync operation started. + SyncedTableProvisioningStatus: + type: object + properties: + initial_pipeline_sync_progress: + $ref: '#/components/schemas/SyncedTablePipelineProgress' + description: >- + Details about initial data synchronization. Only populated when in + the PROVISIONING_INITIAL_SNAPSHOT state. + description: |- + Detailed status of a synced table. Shown if the synced table is in the + PROVISIONING_PIPELINE_RESOURCES or the PROVISIONING_INITIAL_SNAPSHOT state. + SyncedTableSpec: + type: object + properties: + create_database_objects_if_missing: + type: boolean + description: >- + If true, the synced table's logical database and schema resources in + PG will be created if they do not already exist. + existing_pipeline_id: + type: string + description: >- + At most one of existing_pipeline_id and new_pipeline_spec should be + defined. If existing_pipeline_id is defined, the synced table will + be bin packed into the existing pipeline referenced. This avoids + creating a new pipeline and allows sharing existing compute. In this + case, the scheduling_policy of this synced table must match the + scheduling policy of the existing pipeline. + new_pipeline_spec: + $ref: '#/components/schemas/NewPipelineSpec' + description: >- + At most one of existing_pipeline_id and new_pipeline_spec should be + defined. If new_pipeline_spec is defined, a new pipeline is created + for this synced table. The location pointed to is used to store + intermediate files (checkpoints, event logs etc). The caller must + have write permissions to create Delta tables in the specified + catalog and schema. Again, note this requires write permissions, + whereas the source table only requires read permissions. + primary_key_columns: + type: array + items: + type: string + description: >- + Primary Key columns to be used for data insert/update in the + destination. + scheduling_policy: + $ref: '#/components/schemas/SyncedTableSchedulingPolicy' + description: Scheduling policy of the underlying pipeline. + source_table_full_name: + type: string + description: Three-part (catalog, schema, table) name of the source Delta table. + timeseries_key: + type: string + description: >- + Time series key to deduplicate (tie-break) rows with the same + primary key. + description: Specification of a synced database table. + SyncedTableStatus: + type: object + properties: + continuous_update_status: + $ref: '#/components/schemas/SyncedTableContinuousUpdateStatus' + detailed_state: + $ref: '#/components/schemas/SyncedTableState' + description: The state of the synced table. + failed_status: + $ref: '#/components/schemas/SyncedTableFailedStatus' + last_sync: + $ref: '#/components/schemas/SyncedTablePosition' + description: >- + Summary of the last successful synchronization from source to + destination. Will always be present if there has been a successful + sync. Even if the most recent syncs have failed. Limitation: The + only exception is if the synced table is doing a FULL REFRESH, then + the last sync information will not be available until the full + refresh is complete. This limitation will be addressed in a future + version. This top-level field is a convenience for consumers who + want easy access to last sync information without having to traverse + detailed_status. + message: + type: string + description: A text description of the current state of the synced table. + pipeline_id: + type: string + description: >- + ID of the associated pipeline. The pipeline ID may have been + provided by the client (in the case of bin packing), or generated by + the server (when creating a new pipeline). + provisioning_status: + $ref: '#/components/schemas/SyncedTableProvisioningStatus' + triggered_update_status: + $ref: '#/components/schemas/SyncedTableTriggeredUpdateStatus' + description: Status of a synced table. + SyncedTableTriggeredUpdateStatus: + type: object + properties: + last_processed_commit_version: + type: integer + description: >- + The last source table Delta version that was successfully synced to + the synced table. + timestamp: + type: string + description: >- + The end timestamp of the last time any data was synchronized from + the source table to the synced table. This is when the data is + available in the synced table. + triggered_update_progress: + $ref: '#/components/schemas/SyncedTablePipelineProgress' + description: Progress of the active data synchronization pipeline. + description: >- + Detailed status of a synced table. Shown if the synced table is in the + SYNCED_TRIGGERED_UPDATE + or the SYNCED_NO_PENDING_UPDATE state. + DatabaseInstanceRoleIdentityType: + type: string + enum: + - GROUP + - PG_ONLY + - SERVICE_PRINCIPAL + - USER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DatabaseInstanceRoleMembershipRole: + type: string + enum: + - DATABRICKS_SUPERUSER + description: Roles that the DatabaseInstanceRole can be a member of. + DatabaseInstanceState: + type: string + enum: + - AVAILABLE + - DELETING + - FAILING_OVER + - STARTING + - STOPPED + - UPDATING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ProvisioningInfoState: + type: string + enum: + - ACTIVE + - DEGRADED + - DELETING + - FAILED + - PROVISIONING + - UPDATING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ProvisioningPhase: + type: string + enum: + - PROVISIONING_PHASE_INDEX_SCAN + - PROVISIONING_PHASE_INDEX_SORT + - PROVISIONING_PHASE_MAIN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + RequestedClaimsPermissionSet: + type: string + enum: + - READ_ONLY + description: Might add WRITE in the future + SyncedTableSchedulingPolicy: + type: string + enum: + - CONTINUOUS + - SNAPSHOT + - TRIGGERED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SyncedTableState: + type: string + enum: + - SYNCED_TABLED_OFFLINE + - SYNCED_TABLE_OFFLINE_FAILED + - SYNCED_TABLE_ONLINE + - SYNCED_TABLE_ONLINE_CONTINUOUS_UPDATE + - SYNCED_TABLE_ONLINE_NO_PENDING_UPDATE + - SYNCED_TABLE_ONLINE_PIPELINE_FAILED + - SYNCED_TABLE_ONLINE_TRIGGERED_UPDATE + - SYNCED_TABLE_ONLINE_UPDATING_PIPELINE_RESOURCES + - SYNCED_TABLE_PROVISIONING + - SYNCED_TABLE_PROVISIONING_INITIAL_SNAPSHOT + - SYNCED_TABLE_PROVISIONING_PIPELINE_RESOURCES + description: The state of a synced table. + x-stackQL-resources: + database_catalogs: + id: databricks_workspace.database.database_catalogs + name: database_catalogs + title: Database Catalogs + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1database~1catalogs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1database~1catalogs~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1database~1catalogs~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1database~1catalogs~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1database~1instances~1{instance_name}~1catalogs/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.database_catalogs + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/database_catalogs/methods/get' + - $ref: '#/components/x-stackQL-resources/database_catalogs/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/database_catalogs/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/database_catalogs/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/database_catalogs/methods/delete' + replace: [] + database_instances: + id: databricks_workspace.database.database_instances + name: database_instances + title: Database Instances + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1database~1instances/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1database~1instances/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.database_instances + delete: + operation: + $ref: '#/paths/~1api~12.0~1database~1instances~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1database~1instances~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1database~1instances~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + find_by_uid: + operation: + $ref: '#/paths/~1api~12.0~1database~1instances:findByUid/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/database_instances/methods/get' + - $ref: '#/components/x-stackQL-resources/database_instances/methods/list' + - $ref: >- + #/components/x-stackQL-resources/database_instances/methods/find_by_uid + insert: + - $ref: '#/components/x-stackQL-resources/database_instances/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/database_instances/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/database_instances/methods/delete' + replace: [] + database_instance_roles: + id: databricks_workspace.database.database_instance_roles + name: database_instance_roles + title: Database Instance Roles + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1database~1instances~1{instance_name}~1roles/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1database~1instances~1{instance_name}~1roles/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.database_instance_roles + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1database~1instances~1{instance_name}~1roles~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1database~1instances~1{instance_name}~1roles~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/database_instance_roles/methods/get + - $ref: >- + #/components/x-stackQL-resources/database_instance_roles/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/database_instance_roles/methods/create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/database_instance_roles/methods/delete + replace: [] + database_tables: + id: databricks_workspace.database.database_tables + name: database_tables + title: Database Tables + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1database~1tables/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1database~1tables~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1database~1tables~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/database_tables/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/database_tables/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/database_tables/methods/delete' + replace: [] + synced_database_tables: + id: databricks_workspace.database.synced_database_tables + name: synced_database_tables + title: Synced Database Tables + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1database~1synced_tables/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1database~1synced_tables~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1database~1synced_tables~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1database~1synced_tables~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1database~1instances~1{instance_name}~1synced_tables/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.synced_tables + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/synced_database_tables/methods/get + - $ref: >- + #/components/x-stackQL-resources/synced_database_tables/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/synced_database_tables/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/synced_database_tables/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/synced_database_tables/methods/delete + replace: [] + database_credentials: + id: databricks_workspace.database.database_credentials + name: database_credentials + title: Database Credentials + methods: + generate: + operation: + $ref: '#/paths/~1api~12.0~1database~1credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/dataquality.yaml b/providers/src/databricks_workspace/v00.00.00000/services/dataquality.yaml new file mode 100644 index 00000000..61d24d2b --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/dataquality.yaml @@ -0,0 +1,1582 @@ +openapi: 3.0.0 +info: + title: Databricks Dataquality API (workspace) + description: >- + OpenAPI specification for the Databricks dataquality service + (workspace-level APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.dataquality +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/data-quality/v1/monitors/{object_type}/{object_id}/refreshes/{refresh_id}/cancel: + post: + operationId: data_quality_cancel_refresh + summary: >- + Cancels a data quality monitor refresh. Currently only supported for the + `table` `object_type`. The + tags: + - dataquality + - data_quality + description: >- + Cancels a data quality monitor refresh. Currently only supported for the + `table` `object_type`. The + + call must be made in the same workspace as where the monitor was + created. + + + The caller must have either of the following sets of permissions: 1. + **MANAGE** and **USE_CATALOG** on + + the table's parent catalog. 2. **USE_CATALOG** on the table's parent + catalog, and **MANAGE** and + + **USE_SCHEMA** on the table's parent schema. 3. **USE_CATALOG** on the + table's parent catalog, + + **USE_SCHEMA** on the table's parent schema, and **MANAGE** on the + table. + + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + :param refresh_id: int + Unique id of the refresh operation. + + :returns: :class:`CancelRefreshResponse` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + - name: refresh_id + in: path + required: true + schema: + type: integer + description: Unique id of the refresh operation. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CancelRefreshResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + /api/data-quality/v1/monitors: + post: + operationId: data_quality_create_monitor + summary: >- + Create a data quality monitor on a Unity Catalog object. The caller must + provide either + tags: + - dataquality + - data_quality + description: >- + Create a data quality monitor on a Unity Catalog object. The caller must + provide either + + `anomaly_detection_config` for a schema monitor or + `data_profiling_config` for a table monitor. + + + For the `table` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the table's parent catalog, + **USE_SCHEMA** on the table's parent + + schema, and **SELECT** on the table 2. **USE_CATALOG** on the table's + parent catalog, **MANAGE** and + + **USE_SCHEMA** on the table's parent schema, and **SELECT** on the + table. 3. **USE_CATALOG** on the + + table's parent catalog, **USE_SCHEMA** on the table's parent schema, and + **MANAGE** and **SELECT** on + + the table. + + + Workspace assets, such as the dashboard, will be created in the + workspace where this call was made. + + + For the `schema` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the schema's parent catalog. 2. + **USE_CATALOG** on the schema's + + parent catalog, and **MANAGE** and **USE_SCHEMA** on the schema. + + + :param monitor: :class:`Monitor` + The monitor to create. + + :returns: :class:`Monitor` + requestBody: + content: + application/json: + schema: + type: object + properties: + monitor: + type: string + description: The monitor to create. + required: + - monitor + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Monitor' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + get: + operationId: data_quality_list_monitor + summary: (Unimplemented) List data quality monitors. + tags: + - dataquality + - data_quality + description: |- + (Unimplemented) List data quality monitors. + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`Monitor` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListMonitorResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + /api/data-quality/v1/monitors/{object_type}/{object_id}/refreshes: + post: + operationId: data_quality_create_refresh + summary: >- + Creates a refresh. Currently only supported for the `table` + `object_type`. The call must be made in + tags: + - dataquality + - data_quality + description: >- + Creates a refresh. Currently only supported for the `table` + `object_type`. The call must be made in + + the same workspace as where the monitor was created. + + + The caller must have either of the following sets of permissions: 1. + **MANAGE** and **USE_CATALOG** on + + the table's parent catalog. 2. **USE_CATALOG** on the table's parent + catalog, and **MANAGE** and + + **USE_SCHEMA** on the table's parent schema. 3. **USE_CATALOG** on the + table's parent catalog, + + **USE_SCHEMA** on the table's parent schema, and **MANAGE** on the + table. + + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema`or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + :param refresh: :class:`Refresh` + The refresh to create + + :returns: :class:`Refresh` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema`or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + requestBody: + content: + application/json: + schema: + type: object + properties: + refresh: + type: string + description: The refresh to create + required: + - refresh + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Refresh' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + get: + operationId: data_quality_list_refresh + summary: >- + List data quality monitor refreshes. The call must be made in the same + workspace as where the monitor + tags: + - dataquality + - data_quality + description: >- + List data quality monitor refreshes. The call must be made in the same + workspace as where the monitor + + was created. + + + For the `table` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the table's parent catalog. 2. + **USE_CATALOG** on the table's parent + + catalog, and **MANAGE** and **USE_SCHEMA** on the table's parent schema. + 3. **USE_CATALOG** on the + + table's parent catalog, **USE_SCHEMA** on the table's parent schema, and + **SELECT** on the table. + + + For the `schema` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the schema's parent catalog. 2. + **USE_CATALOG** on the schema's + + parent catalog, and **USE_SCHEMA** on the schema. + + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`Refresh` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListRefreshResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + /api/data-quality/v1/monitors/{object_type}/{object_id}: + delete: + operationId: data_quality_delete_monitor + summary: Delete a data quality monitor on Unity Catalog object. + tags: + - dataquality + - data_quality + description: >- + Delete a data quality monitor on Unity Catalog object. + + + For the `table` `object_type`, the caller must have either of the + following sets of permissions: + + **MANAGE** and **USE_CATALOG** on the table's parent catalog. + **USE_CATALOG** on the table's parent + + catalog, and **MANAGE** and **USE_SCHEMA** on the table's parent schema. + **USE_CATALOG** on the + + table's parent catalog, **USE_SCHEMA** on the table's parent schema, and + **MANAGE** on the table. + + + Note that the metric tables and dashboard will not be deleted as part of + this call; those assets must + + be manually cleaned up (if desired). + + + For the `schema` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the schema's parent catalog. 2. + **USE_CATALOG** on the schema's + + parent catalog, and **MANAGE** and **USE_SCHEMA** on the schema. + + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + get: + operationId: data_quality_get_monitor + summary: Read a data quality monitor on a Unity Catalog object. + tags: + - dataquality + - data_quality + description: >- + Read a data quality monitor on a Unity Catalog object. + + + For the `table` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the table's parent catalog. 2. + **USE_CATALOG** on the table's parent + + catalog, and **MANAGE** and **USE_SCHEMA** on the table's parent schema. + 3. **USE_CATALOG** on the + + table's parent catalog, **USE_SCHEMA** on the table's parent schema, and + **SELECT** on the table. + + + For the `schema` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the schema's parent catalog. 2. + **USE_CATALOG** on the schema's + + parent catalog, and **USE_SCHEMA** on the schema. + + + The returned information includes configuration values on the entity and + parent entity as well as + + information on assets created by the monitor. Some information (e.g. + dashboard) may be filtered out if + + the caller is in a different workspace than where the monitor was + created. + + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + + :returns: :class:`Monitor` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Monitor' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + patch: + operationId: data_quality_update_monitor + summary: Update a data quality monitor on Unity Catalog object. + tags: + - dataquality + - data_quality + description: >- + Update a data quality monitor on Unity Catalog object. + + + For the `table` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the table's parent catalog. 2. + **USE_CATALOG** on the table's parent + + catalog, and **MANAGE** and **USE_SCHEMA** on the table's parent schema. + 3. **USE_CATALOG** on the + + table's parent catalog, **USE_SCHEMA** on the table's parent schema, and + **MANAGE** on the table. + + + For the `schema` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the schema's parent catalog. 2. + **USE_CATALOG** on the schema's + + parent catalog, and **MANAGE** and **USE_SCHEMA** on the schema. + + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + :param monitor: :class:`Monitor` + The monitor to update. + :param update_mask: str + The field mask to specify which fields to update as a comma-separated list. Example value: + `data_profiling_config.custom_metrics,data_profiling_config.schedule.quartz_cron_expression` + + :returns: :class:`Monitor` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + The field mask to specify which fields to update as a + comma-separated list. Example value: + `data_profiling_config.custom_metrics,data_profiling_config.schedule.quartz_cron_expression` + requestBody: + content: + application/json: + schema: + type: object + properties: + monitor: + type: string + description: The monitor to update. + required: + - monitor + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Monitor' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + /api/data-quality/v1/monitors/{object_type}/{object_id}/refreshes/{refresh_id}: + delete: + operationId: data_quality_delete_refresh + summary: (Unimplemented) Delete a refresh + tags: + - dataquality + - data_quality + description: |- + (Unimplemented) Delete a refresh + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + :param refresh_id: int + Unique id of the refresh operation. + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + - name: refresh_id + in: path + required: true + schema: + type: integer + description: Unique id of the refresh operation. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + get: + operationId: data_quality_get_refresh + summary: >- + Get data quality monitor refresh. The call must be made in the same + workspace as where the monitor was + tags: + - dataquality + - data_quality + description: >- + Get data quality monitor refresh. The call must be made in the same + workspace as where the monitor was + + created. + + + For the `table` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the table's parent catalog. 2. + **USE_CATALOG** on the table's parent + + catalog, and **MANAGE** and **USE_SCHEMA** on the table's parent schema. + 3. **USE_CATALOG** on the + + table's parent catalog, **USE_SCHEMA** on the table's parent schema, and + **SELECT** on the table. + + + For the `schema` `object_type`, the caller must have either of the + following sets of permissions: 1. + + **MANAGE** and **USE_CATALOG** on the schema's parent catalog. 2. + **USE_CATALOG** on the schema's + + parent catalog, and **USE_SCHEMA** on the schema. + + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + :param refresh_id: int + Unique id of the refresh operation. + + :returns: :class:`Refresh` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + - name: refresh_id + in: path + required: true + schema: + type: integer + description: Unique id of the refresh operation. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Refresh' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI + patch: + operationId: data_quality_update_refresh + summary: (Unimplemented) Update a refresh + tags: + - dataquality + - data_quality + description: |- + (Unimplemented) Update a refresh + + :param object_type: str + The type of the monitored object. Can be one of the following: `schema` or `table`. + :param object_id: str + The UUID of the request object. It is `schema_id` for `schema`, and `table_id` for `table`. + + Find the `schema_id` from either: 1. The [schema_id] of the `Schemas` resource. 2. In [Catalog + Explorer] > select the `schema` > go to the `Details` tab > the `Schema ID` field. + + Find the `table_id` from either: 1. The [table_id] of the `Tables` resource. 2. In [Catalog + Explorer] > select the `table` > go to the `Details` tab > the `Table ID` field. + + [Catalog Explorer]: https://docs.databricks.com/aws/en/catalog-explorer/ + [schema_id]: https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: https://docs.databricks.com/api/workspace/tables/get#table_id + :param refresh_id: int + Unique id of the refresh operation. + :param refresh: :class:`Refresh` + The refresh to update. + :param update_mask: str + The field mask to specify which fields to update. + + :returns: :class:`Refresh` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + - name: object_id + in: path + required: true + schema: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + - name: refresh_id + in: path + required: true + schema: + type: integer + description: Unique id of the refresh operation. + - name: update_mask + in: query + required: true + schema: + type: string + description: The field mask to specify which fields to update. + requestBody: + content: + application/json: + schema: + type: object + properties: + refresh: + type: string + description: The refresh to update. + required: + - refresh + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Refresh' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataQualityAPI +components: + schemas: + AnomalyDetectionConfig: + type: object + properties: + excluded_table_full_names: + type: array + items: + type: string + description: >- + List of fully qualified table names to exclude from anomaly + detection. + description: Anomaly Detection Configurations. + CancelRefreshResponse: + type: object + properties: + refresh: + $ref: '#/components/schemas/Refresh' + description: The refresh to cancel. + description: Response to cancelling a refresh. + CronSchedule: + type: object + properties: + quartz_cron_expression: + type: string + description: >- + The expression that determines when to run the monitor. See + [examples]. [examples]: + https://www.quartz-scheduler.org/documentation/quartz-2.3.0/tutorials/crontrigger.html + timezone_id: + type: string + description: >- + A Java timezone id. The schedule for a job will be resolved with + respect to this timezone. See `Java TimeZone + `_ + for details. The timezone id (e.g., ``America/Los_Angeles``) in + which to evaluate the quartz expression. + pause_status: + $ref: '#/components/schemas/CronSchedulePauseStatus' + description: >- + Read only field that indicates whether the schedule is paused or + not. + required: + - quartz_cron_expression + - timezone_id + description: The data quality monitoring workflow cron schedule. + DataProfilingConfig: + type: object + properties: + output_schema_id: + type: string + description: ID of the schema where output tables are created. + assets_dir: + type: string + description: >- + Field for specifying the absolute path to a custom directory to + store data-monitoring assets. Normally prepopulated to a default + user location via UI and Python APIs. + baseline_table_name: + type: string + description: >- + Baseline table name. Baseline data is used to compute drift from the + data in the monitored `table_name`. The baseline table and the + monitored table shall have the same schema. + custom_metrics: + type: array + items: + $ref: '#/components/schemas/DataProfilingCustomMetric' + description: Custom metrics. + dashboard_id: + type: string + description: >- + Id of dashboard that visualizes the computed metrics. This can be + empty if the monitor is in PENDING state. + drift_metrics_table_name: + type: string + description: >- + Table that stores drift metrics data. Format: + `catalog.schema.table_name`. + effective_warehouse_id: + type: string + description: The warehouse for dashboard creation + inference_log: + $ref: '#/components/schemas/InferenceLogConfig' + description: '`Analysis Configuration` for monitoring inference log tables.' + latest_monitor_failure_message: + type: string + description: The latest error message for a monitor failure. + monitor_version: + type: integer + description: >- + Represents the current monitor configuration version in use. The + version will be represented in a numeric fashion (1,2,3...). The + field has flexibility to take on negative values, which can indicate + corrupted monitor_version numbers. + monitored_table_name: + type: string + description: 'Unity Catalog table to monitor. Format: `catalog.schema.table_name`' + notification_settings: + $ref: '#/components/schemas/NotificationSettings' + description: Field for specifying notification settings. + profile_metrics_table_name: + type: string + description: >- + Table that stores profile metrics data. Format: + `catalog.schema.table_name`. + schedule: + $ref: '#/components/schemas/CronSchedule' + description: The cron schedule. + skip_builtin_dashboard: + type: boolean + description: >- + Whether to skip creating a default dashboard summarizing data + quality metrics. + slicing_exprs: + type: array + items: + type: string + description: >- + List of column expressions to slice data with for targeted analysis. + The data is grouped by each expression independently, resulting in a + separate slice for each predicate and its complements. For example + `slicing_exprs=[“col_1”, “col_2 > 10”]` will generate the following + slices: two slices for `col_2 > 10` (True and False), and one slice + per unique value in `col1`. For high-cardinality columns, only the + top 100 unique values by frequency will generate slices. + snapshot: + $ref: '#/components/schemas/SnapshotConfig' + description: '`Analysis Configuration` for monitoring snapshot tables.' + status: + $ref: '#/components/schemas/DataProfilingStatus' + description: The data profiling monitor status. + time_series: + $ref: '#/components/schemas/TimeSeriesConfig' + description: '`Analysis Configuration` for monitoring time series tables.' + warehouse_id: + type: string + description: >- + Optional argument to specify the warehouse for dashboard creation. + If not specified, the first running warehouse will be used. + required: + - output_schema_id + description: Data Profiling Configurations. + DataProfilingCustomMetric: + type: object + properties: + name: + type: string + description: Name of the metric in the output tables. + definition: + type: string + description: >- + Jinja template for a SQL expression that specifies how to compute + the metric. See [create metric definition]. [create metric + definition]: + https://docs.databricks.com/en/lakehouse-monitoring/custom-metrics.html#create-definition + input_columns: + type: array + items: + type: string + description: >- + A list of column names in the input table the metric should be + computed for. Can use ``":table"`` to indicate that the metric needs + information from multiple columns. + output_data_type: + type: string + description: The output type of the custom metric. + type: + $ref: '#/components/schemas/DataProfilingCustomMetricType' + description: The type of the custom metric. + required: + - name + - definition + - input_columns + - output_data_type + - type + description: Custom metric definition. + InferenceLogConfig: + type: object + properties: + problem_type: + $ref: '#/components/schemas/InferenceProblemType' + description: Problem type the model aims to solve. + timestamp_column: + type: string + description: Column for the timestamp. + granularities: + type: array + items: + $ref: '#/components/schemas/AggregationGranularity' + description: >- + List of granularities to use when aggregating data into time windows + based on their timestamp. + prediction_column: + type: string + description: Column for the prediction. + model_id_column: + type: string + description: Column for the model identifier. + label_column: + type: string + description: Column for the label. + required: + - problem_type + - timestamp_column + - granularities + - prediction_column + - model_id_column + description: Inference log configuration. + ListMonitorResponse: + type: object + properties: + monitors: + type: array + items: + $ref: '#/components/schemas/Monitor' + next_page_token: + type: string + description: Response for listing Monitors. + ListRefreshResponse: + type: object + properties: + next_page_token: + type: string + refreshes: + type: array + items: + $ref: '#/components/schemas/Refresh' + description: Response for listing refreshes. + Monitor: + type: object + properties: + object_type: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema` or `table`. + object_id: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + anomaly_detection_config: + $ref: '#/components/schemas/AnomalyDetectionConfig' + description: >- + Anomaly Detection Configuration, applicable to `schema` object + types. + data_profiling_config: + $ref: '#/components/schemas/DataProfilingConfig' + description: >- + Data Profiling Configuration, applicable to `table` object types. + Exactly one `Analysis Configuration` must be present. + required: + - object_type + - object_id + description: >- + Monitor for the data quality of unity catalog entities such as schema or + table. + NotificationDestination: + type: object + properties: + email_addresses: + type: array + items: + type: string + description: >- + The list of email addresses to send the notification to. A maximum + of 5 email addresses is supported. + description: Destination of the data quality monitoring notification. + NotificationSettings: + type: object + properties: + on_failure: + $ref: '#/components/schemas/NotificationDestination' + description: Destinations to send notifications on failure/timeout. + description: Settings for sending notifications on the data quality monitoring. + Refresh: + type: object + properties: + object_type: + type: string + description: >- + The type of the monitored object. Can be one of the following: + `schema`or `table`. + object_id: + type: string + description: >- + The UUID of the request object. It is `schema_id` for `schema`, and + `table_id` for `table`. Find the `schema_id` from either: 1. The + [schema_id] of the `Schemas` resource. 2. In [Catalog Explorer] > + select the `schema` > go to the `Details` tab > the `Schema ID` + field. Find the `table_id` from either: 1. The [table_id] of the + `Tables` resource. 2. In [Catalog Explorer] > select the `table` > + go to the `Details` tab > the `Table ID` field. [Catalog Explorer]: + https://docs.databricks.com/aws/en/catalog-explorer/ [schema_id]: + https://docs.databricks.com/api/workspace/schemas/get#schema_id + [table_id]: + https://docs.databricks.com/api/workspace/tables/get#table_id + end_time_ms: + type: integer + description: Time when the refresh ended (milliseconds since 1/1/1970 UTC). + message: + type: string + description: >- + An optional message to give insight into the current state of the + refresh (e.g. FAILURE messages). + refresh_id: + type: integer + description: Unique id of the refresh operation. + start_time_ms: + type: integer + description: Time when the refresh started (milliseconds since 1/1/1970 UTC). + state: + $ref: '#/components/schemas/RefreshState' + description: The current state of the refresh. + trigger: + $ref: '#/components/schemas/RefreshTrigger' + description: What triggered the refresh. + required: + - object_type + - object_id + description: >- + The Refresh object gives information on a refresh of the data quality + monitoring pipeline. + SnapshotConfig: + type: object + properties: {} + description: Snapshot analysis configuration. + TimeSeriesConfig: + type: object + properties: + timestamp_column: + type: string + description: Column for the timestamp. + granularities: + type: array + items: + $ref: '#/components/schemas/AggregationGranularity' + description: >- + List of granularities to use when aggregating data into time windows + based on their timestamp. + required: + - timestamp_column + - granularities + description: Time series analysis configuration. + AggregationGranularity: + type: string + enum: + - AGGREGATION_GRANULARITY_1_DAY + - AGGREGATION_GRANULARITY_1_HOUR + - AGGREGATION_GRANULARITY_1_MONTH + - AGGREGATION_GRANULARITY_1_WEEK + - AGGREGATION_GRANULARITY_1_YEAR + - AGGREGATION_GRANULARITY_2_WEEKS + - AGGREGATION_GRANULARITY_30_MINUTES + - AGGREGATION_GRANULARITY_3_WEEKS + - AGGREGATION_GRANULARITY_4_WEEKS + - AGGREGATION_GRANULARITY_5_MINUTES + description: >- + The granularity for aggregating data into time windows based on their + timestamp. + CronSchedulePauseStatus: + type: string + enum: + - CRON_SCHEDULE_PAUSE_STATUS_PAUSED + - CRON_SCHEDULE_PAUSE_STATUS_UNPAUSED + description: The data quality monitoring workflow cron schedule pause status. + DataProfilingCustomMetricType: + type: string + enum: + - DATA_PROFILING_CUSTOM_METRIC_TYPE_AGGREGATE + - DATA_PROFILING_CUSTOM_METRIC_TYPE_DERIVED + - DATA_PROFILING_CUSTOM_METRIC_TYPE_DRIFT + description: The custom metric type. + DataProfilingStatus: + type: string + enum: + - DATA_PROFILING_STATUS_ACTIVE + - DATA_PROFILING_STATUS_DELETE_PENDING + - DATA_PROFILING_STATUS_ERROR + - DATA_PROFILING_STATUS_FAILED + - DATA_PROFILING_STATUS_PENDING + description: The status of the data profiling monitor. + InferenceProblemType: + type: string + enum: + - INFERENCE_PROBLEM_TYPE_CLASSIFICATION + - INFERENCE_PROBLEM_TYPE_REGRESSION + description: Inference problem type the model aims to solve. + RefreshState: + type: string + enum: + - MONITOR_REFRESH_STATE_CANCELED + - MONITOR_REFRESH_STATE_FAILED + - MONITOR_REFRESH_STATE_PENDING + - MONITOR_REFRESH_STATE_RUNNING + - MONITOR_REFRESH_STATE_SUCCESS + - MONITOR_REFRESH_STATE_UNKNOWN + description: The state of the refresh. + RefreshTrigger: + type: string + enum: + - MONITOR_REFRESH_TRIGGER_DATA_CHANGE + - MONITOR_REFRESH_TRIGGER_MANUAL + - MONITOR_REFRESH_TRIGGER_SCHEDULE + - MONITOR_REFRESH_TRIGGER_UNKNOWN + description: The trigger of the refresh. + x-stackQL-resources: + data_quality_refreshes: + id: databricks_workspace.dataquality.data_quality_refreshes + name: data_quality_refreshes + title: Data Quality Refreshes + methods: + cancel: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}~1refreshes~1{refresh_id}~1cancel/post + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}~1refreshes/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}~1refreshes/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.refreshes + delete: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}~1refreshes~1{refresh_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}~1refreshes~1{refresh_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}~1refreshes~1{refresh_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/data_quality_refreshes/methods/get + - $ref: >- + #/components/x-stackQL-resources/data_quality_refreshes/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/data_quality_refreshes/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/data_quality_refreshes/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/data_quality_refreshes/methods/delete + replace: [] + data_quality: + id: databricks_workspace.dataquality.data_quality + name: data_quality + title: Data Quality + methods: + create: + operation: + $ref: '#/paths/~1api~1data-quality~1v1~1monitors/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~1data-quality~1v1~1monitors/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.monitors + delete: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~1data-quality~1v1~1monitors~1{object_type}~1{object_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/data_quality/methods/get' + - $ref: '#/components/x-stackQL-resources/data_quality/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/data_quality/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/data_quality/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/data_quality/methods/delete' + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/dbsql.yaml b/providers/src/databricks_workspace/v00.00.00000/services/dbsql.yaml deleted file mode 100644 index 343fbd17..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/dbsql.yaml +++ /dev/null @@ -1,5452 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Dbsql API - description: Dbsql -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/sql/alerts: - post: - operationId: alerts-create - externalDocs: - url: https://docs.databricks.com/api/workspace/alerts/create - x-stackQL-resource: alerts - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - state: - type: string - seconds_to_retrigger: - type: integer - owner_user_name: - type: string - display_name: - type: string - id: - type: string - condition: - type: object - properties: - op: - type: string - operand: - type: object - properties: - column: - type: object - properties: - name: - type: string - threshold: - type: object - properties: - value: - type: object - properties: - double_value: - type: integer - query_id: - type: string - parent_path: - type: string - lifecycle_state: - type: string - update_time: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - alert: - type: object - example: - alert: - seconds_to_retrigger: 0 - display_name: Test alert - condition: - op: GREATER_THAN - operand: - column: - name: foo - threshold: - value: - double_value: 1 - query_id: dee5cca8-1c79-4b5e-a71m1-e7f9d241bdf6 - parent_path: /Workspace/Users/user.name@acme.com - description: Creates an alert. - get: - operationId: alerts-list - externalDocs: - url: https://docs.databricks.com/api/workspace/alerts/list - x-stackQL-resource: alerts - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - next_page_token: - type: string - results: - type: array - items: - type: object - properties: - state: - type: string - seconds_to_retrigger: - type: integer - owner_user_name: - type: string - display_name: - type: string - id: - type: string - condition: - type: object - properties: - op: - type: string - operand: - type: object - properties: - column: - type: object - properties: - name: - type: string - threshold: - type: object - properties: - value: - type: object - properties: - double_value: - type: integer - query_id: - type: string - lifecycle_state: - type: string - update_time: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.results - description: Gets a list of alerts accessible to the user, ordered by creation - time. - /api/2.0/sql/alerts/{id}: - delete: - operationId: alerts-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/alerts/delete - x-stackQL-resource: alerts - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Moves an alert to the trash. Trashed alerts immediately disappear - from searches and list views, and can no longer trigger. You can restore a - trashed alert through the UI. A trashed alert is permanently deleted after - 30 days. - get: - operationId: alerts-get - externalDocs: - url: https://docs.databricks.com/api/workspace/alerts/get - x-stackQL-resource: alerts - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - state: - type: string - seconds_to_retrigger: - type: integer - owner_user_name: - type: string - display_name: - type: string - id: - type: string - condition: - type: object - properties: - op: - type: string - operand: - type: object - properties: - column: - type: object - properties: - name: - type: string - threshold: - type: object - properties: - value: - type: object - properties: - string_value: - type: string - query_id: - type: string - parent_path: - type: string - lifecycle_state: - type: string - update_time: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets an alert. - patch: - operationId: alerts-update - externalDocs: - url: https://docs.databricks.com/api/workspace/alerts/update - x-stackQL-resource: alerts - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - state: - type: string - seconds_to_retrigger: - type: integer - owner_user_name: - type: string - display_name: - type: string - id: - type: string - condition: - type: object - properties: - op: - type: string - operand: - type: object - properties: - column: - type: object - properties: - name: - type: string - threshold: - type: object - properties: - value: - type: object - properties: - bool_value: - type: boolean - query_id: - type: string - parent_path: - type: string - lifecycle_state: - type: string - update_time: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - update_mask: - type: required - alert: - type: object - example: - alert: - condition: - op: EQUAL - operand: - column: - name: bar - threshold: - value: - bool_value: true - display_name: Renamed alert - id: fe25e731-92f2-4838-9fb2-1ca364320a3d - update_mask: display_name,condition - description: Updates an alert. - /api/2.0/preview/sql/alerts: - post: - operationId: alerts-legacy-create - externalDocs: - url: https://docs.databricks.com/api/workspace/alertslegacy/create - x-stackQL-resource: alerts_legacy - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: An alert was successfully created - content: - application/json: - schema: - type: object - properties: - created_at: - type: string - id: - type: string - parent: - type: string - last_triggered_at: - type: string - name: - type: string - options: - type: object - properties: - column: - type: string - custom_body: - type: string - custom_subject: - type: string - muted: - type: boolean - op: - type: string - value: - type: 'null' - empty_result_state: - type: string - query: - type: object - properties: - created_at: - type: string - data_source_id: - type: string - description: - type: string - id: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - is_safe: - type: boolean - name: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - query: - type: string - updated_at: - type: string - tags: - type: array - items: - type: string - user_id: - type: integer - rearm: - type: integer - state: - type: string - updated_at: - type: string - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - query_id: - type: string - rearm: - type: required - parent: - type: UUID - options: - type: object - example: - name: Total Trips > 300 - options: - column: total_trips - custom_body: Total trips exceeded - custom_subject: The total trips has exceeded 300. - muted: false - op: '>' - value: null - empty_result_state: unknown - query_id: dee5cca8-1c79-4b5e-a711-e7f9d241bdf6 - rearm: 0 - parent: folders/2025532471912059 - description: Creates an alert. An alert is a Databricks SQL object that periodically - runs a query, evaluates a condition of its result, and notifies users or notification - destinations if the condition was met. - get: - operationId: alerts-legacy-list - externalDocs: - url: https://docs.databricks.com/api/workspace/alertslegacy/list - x-stackQL-resource: alerts_legacy - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: List of alerts was successfully fetched. - content: - application/json: - schema: - type: array - items: - type: object - properties: - created_at: - type: string - id: - type: string - parent: - type: string - last_triggered_at: - type: string - name: - type: string - options: - type: object - properties: - column: - type: string - custom_body: - type: string - custom_subject: - type: string - muted: - type: boolean - op: - type: string - value: - type: 'null' - empty_result_state: - type: string - query: - type: object - properties: - created_at: - type: string - data_source_id: - type: string - description: - type: string - id: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - is_safe: - type: boolean - name: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - query: - type: string - updated_at: - type: string - tags: - type: array - items: - type: string - user_id: - type: integer - rearm: - type: integer - state: - type: string - updated_at: - type: string - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - description: Gets a list of alerts. - /api/2.0/preview/sql/alerts/{alert_id}: - delete: - operationId: alerts-legacy-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/alertslegacy/delete - x-stackQL-resource: alerts_legacy - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: alert_id - required: true - description: string - in: path - responses: - '200': - description: Alert was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes an alert. Deleted alerts are no longer accessible and cannot - be restored. - get: - operationId: alerts-legacy-get - externalDocs: - url: https://docs.databricks.com/api/workspace/alertslegacy/get - x-stackQL-resource: alerts_legacy - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: alert_id - required: true - description: string - in: path - responses: - '200': - description: Alert was successfully fetched. - content: - application/json: - schema: - type: object - properties: - created_at: - type: string - id: - type: string - parent: - type: string - last_triggered_at: - type: string - name: - type: string - options: - type: object - properties: - column: - type: string - custom_body: - type: string - custom_subject: - type: string - muted: - type: boolean - op: - type: string - value: - type: 'null' - empty_result_state: - type: string - query: - type: object - properties: - created_at: - type: string - data_source_id: - type: string - description: - type: string - id: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - is_safe: - type: boolean - name: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - query: - type: string - updated_at: - type: string - tags: - type: array - items: - type: string - user_id: - type: integer - rearm: - type: integer - state: - type: string - updated_at: - type: string - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets an alert. - put: - operationId: alerts-legacy-update - externalDocs: - url: https://docs.databricks.com/api/workspace/alertslegacy/update - x-stackQL-resource: alerts_legacy - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: alert_id - required: true - description: string - in: path - responses: - '200': - description: Alert was successfully updated. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - query_id: - type: string - rearm: - type: required - options: - type: object - example: - name: Total Trips > 300 - options: - column: total_trips - custom_body: Total trips exceeded - custom_subject: The total trips has exceeded 300. - muted: false - op: '>' - value: null - empty_result_state: unknown - query_id: dee5cca8-1c79-4b5e-a711-e7f9d241bdf6 - rearm: 0 - description: Updates an alert. - /api/2.0/preview/sql/dashboards: - post: - operationId: dashboards-create - externalDocs: - url: https://docs.databricks.com/api/workspace/dashboards/create - x-stackQL-resource: dashboards - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: A dashboard object was successfully created. - content: - application/json: - schema: - type: object - properties: - id: - type: string - slug: - type: string - parent: - type: string - name: - type: string - user_id: - type: integer - dashboard_filters_enabled: - type: boolean - widgets: - type: array - items: - type: object - properties: - id: - type: string - width: - type: integer - options: - type: object - properties: - parameterMappings: - type: object - properties: - param: - type: object - properties: - name: - type: string - mapTo: - type: string - title: - type: string - type: - type: string - value: - type: 'null' - isHidden: - type: boolean - position: - type: object - properties: - autoHeight: - type: boolean - sizeX: - type: integer - sizeY: - type: integer - col: - type: integer - row: - type: integer - title: - type: string - description: - type: string - updated_at: - type: string - created_at: - type: string - visualization: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - tags: - type: array - items: - type: string - updated_at: - type: string - created_at: - type: string - is_favorite: - type: boolean - can_edit: - type: boolean - permission_tier: - type: string - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - parent: - type: string - tags: - type: string - is_favorite: - type: Array of string - run_as_role: - type: boolean - dashboard_filters_enabled: - type: string - example: - name: Sales Dashboard - parent: folders/2025532471912059 - tags: - - Payroll - is_favorite: true - run_as_role: viewer - dashboard_filters_enabled: true - description: Creates a new dashboard object. Only the - get: - operationId: dashboards-list - externalDocs: - url: https://docs.databricks.com/api/workspace/dashboards/list - x-stackQL-resource: dashboards - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_size - in: query - - name: page - in: query - - name: order - in: query - - name: q - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - count: - type: integer - page: - type: integer - page_size: - type: integer - results: - type: array - items: - type: object - properties: - id: - type: string - slug: - type: string - parent: - type: string - name: - type: string - user_id: - type: integer - dashboard_filters_enabled: - type: boolean - widgets: - type: array - items: - type: object - properties: - id: - type: string - width: - type: integer - options: - type: object - properties: - parameterMappings: - type: object - properties: - param: - type: object - properties: - name: - type: string - mapTo: - type: string - title: - type: string - type: - type: string - value: - type: 'null' - isHidden: - type: boolean - position: - type: object - properties: - autoHeight: - type: boolean - sizeX: - type: integer - sizeY: - type: integer - col: - type: integer - row: - type: integer - title: - type: string - description: - type: string - updated_at: - type: string - created_at: - type: string - visualization: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - tags: - type: array - items: - type: string - updated_at: - type: string - created_at: - type: string - is_favorite: - type: boolean - can_edit: - type: boolean - permission_tier: - type: string - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.results - description: Fetch a paginated list of dashboard objects. - /api/2.0/preview/sql/dashboards/{dashboard_id}: - delete: - operationId: dashboards-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/dashboards/delete - x-stackQL-resource: dashboards - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Dashboard successfully moved to trash. - content: - application/json: - schema: - type: object - properties: {} - description: Moves a dashboard to the trash. Trashed dashboards do not appear - in list views or searches, and cannot be shared. - get: - operationId: dashboards-get - externalDocs: - url: https://docs.databricks.com/api/workspace/dashboards/get - x-stackQL-resource: dashboards - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Dashboard fetched successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - slug: - type: string - parent: - type: string - name: - type: string - user_id: - type: integer - dashboard_filters_enabled: - type: boolean - widgets: - type: array - items: - type: object - properties: - id: - type: string - width: - type: integer - options: - type: object - properties: - parameterMappings: - type: object - properties: - param: - type: object - properties: - name: - type: string - mapTo: - type: string - title: - type: string - type: - type: string - value: - type: 'null' - isHidden: - type: boolean - position: - type: object - properties: - autoHeight: - type: boolean - sizeX: - type: integer - sizeY: - type: integer - col: - type: integer - row: - type: integer - title: - type: string - description: - type: string - updated_at: - type: string - created_at: - type: string - visualization: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - tags: - type: array - items: - type: string - updated_at: - type: string - created_at: - type: string - is_favorite: - type: boolean - can_edit: - type: boolean - permission_tier: - type: string - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - description: Returns a JSON representation of a dashboard object, including - its visualization and query objects. - post: - operationId: dashboards-update - externalDocs: - url: https://docs.databricks.com/api/workspace/dashboards/update - x-stackQL-resource: dashboards - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Dashboard updated successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - slug: - type: string - parent: - type: string - name: - type: string - user_id: - type: integer - dashboard_filters_enabled: - type: boolean - widgets: - type: array - items: - type: object - properties: - id: - type: string - width: - type: integer - options: - type: object - properties: - parameterMappings: - type: object - properties: - param: - type: object - properties: - name: - type: string - mapTo: - type: string - title: - type: string - type: - type: string - value: - type: 'null' - isHidden: - type: boolean - position: - type: object - properties: - autoHeight: - type: boolean - sizeX: - type: integer - sizeY: - type: integer - col: - type: integer - row: - type: integer - title: - type: string - description: - type: string - updated_at: - type: string - created_at: - type: string - visualization: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - tags: - type: array - items: - type: string - updated_at: - type: string - created_at: - type: string - is_favorite: - type: boolean - can_edit: - type: boolean - permission_tier: - type: string - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - run_as_role: - type: string - tags: - type: Array of string - example: - name: Sales Dashboard - run_as_role: viewer - tags: - - Payroll - description: The dashboard definition that replaces the current definition - for this - description: Modify this dashboard definition. This operation only affects attributes - of the dashboard object. It does not add, modify, or remove widgets. - /api/2.0/preview/sql/dashboards/trash/{dashboard_id}: - post: - operationId: dashboards-restore - externalDocs: - url: https://docs.databricks.com/api/workspace/dashboards/restore - x-stackQL-resource: dashboards - x-stackQL-method: restore - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Dashboard restored successfully. - content: - application/json: - schema: - type: object - properties: {} - description: A restored dashboard appears in list views and searches and can - be shared. - /api/2.0/preview/sql/permissions/{objectType}/{objectId}: - get: - operationId: permissions-get - externalDocs: - url: https://docs.databricks.com/api/workspace/dbsqlpermissions/get - x-stackQL-resource: permissions - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: objectType - required: true - description: string - in: path - - name: objectId - required: true - description: uuid - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - permission_level: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a JSON representation of the access control list (ACL) for - a specified object. - post: - operationId: permissions-set - externalDocs: - url: https://docs.databricks.com/api/workspace/dbsqlpermissions/set - x-stackQL-resource: permissions - x-stackQL-method: set - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: objectType - required: true - description: string - in: path - - name: objectId - required: true - description: uuid - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - permission_level: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: user@example.com - group_name: admins - permission_level: CAN_VIEW - description: An ACL list to be applied to the object specified in the URL. - description: Sets the access control list (ACL) for a specified object. This - operation will complete rewrite the ACL. - /api/2.0/preview/sql/permissions/{objectType}/{objectId}/transfer: - post: - operationId: permissions-transferownership - externalDocs: - url: https://docs.databricks.com/api/workspace/dbsqlpermissions/transferownership - x-stackQL-resource: permissions - x-stackQL-method: transferownership - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: objectType - required: true - description: string - in: path - - name: null - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - message: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - new_owner: - type: email - example: - new_owner: user@example.com - description: Email address for the new owner, who must exist in the workspace. - description: Transfer ownership of a single object. - /api/2.0/sql/queries: - post: - operationId: queries-create - externalDocs: - url: https://docs.databricks.com/api/workspace/queries/create - x-stackQL-resource: queries - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - description: - type: string - owner_user_name: - type: string - tags: - type: array - items: - type: string - display_name: - type: string - id: - type: string - parent_path: - type: string - lifecycle_state: - type: string - last_modifier_user_name: - type: string - query_text: - type: string - parameters: - type: array - items: - type: object - properties: - name: - type: string - text_value: - type: object - properties: - value: - type: string - title: - type: string - warehouse_id: - type: string - run_as_mode: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - query: - type: object - example: - query: - description: Example description - tags: - - Tag 1 - display_name: Example query - parent_path: /Workspace/Users/user@acme.com - query_text: SELECT 1 - parameters: - - name: foo - text_value: - value: bar - title: foo - warehouse_id: a7066a8ef796be84 - run_as_mode: OWNER - description: Creates a query. - get: - operationId: queries-list - externalDocs: - url: https://docs.databricks.com/api/workspace/queries/list - x-stackQL-resource: queries - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - next_page_token: - type: string - results: - type: array - items: - type: object - properties: - description: - type: string - owner_user_name: - type: string - tags: - type: array - items: - type: string - display_name: - type: string - id: - type: string - lifecycle_state: - type: string - last_modifier_user_name: - type: string - query_text: - type: string - parameters: - type: array - items: - type: object - properties: - name: - type: string - text_value: - type: object - properties: - value: - type: string - title: - type: string - update_time: - type: string - warehouse_id: - type: string - run_as_mode: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.results - description: Gets a list of queries accessible to the user, ordered by creation - time. - /api/2.0/sql/queries/{id}: - delete: - operationId: queries-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/queries/delete - x-stackQL-resource: queries - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Moves a query to the trash. Trashed queries immediately disappear - from searches and list views, and cannot be used for alerts. You can restore - a trashed query through the UI. A trashed query is permanently deleted after - 30 days. - get: - operationId: queries-get - externalDocs: - url: https://docs.databricks.com/api/workspace/queries/get - x-stackQL-resource: queries - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - description: - type: string - owner_user_name: - type: string - tags: - type: array - items: - type: string - display_name: - type: string - id: - type: string - parent_path: - type: string - lifecycle_state: - type: string - last_modifier_user_name: - type: string - query_text: - type: string - parameters: - type: array - items: - type: object - properties: - name: - type: string - text_value: - type: object - properties: - value: - type: string - title: - type: string - update_time: - type: string - warehouse_id: - type: string - run_as_mode: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a query. - patch: - operationId: queries-update - externalDocs: - url: https://docs.databricks.com/api/workspace/queries/update - x-stackQL-resource: queries - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - description: - type: string - owner_user_name: - type: string - tags: - type: array - items: - type: string - display_name: - type: string - id: - type: string - parent_path: - type: string - lifecycle_state: - type: string - last_modifier_user_name: - type: string - query_text: - type: string - parameters: - type: array - items: - type: object - properties: - name: - type: string - text_value: - type: object - properties: - value: - type: string - title: - type: string - update_time: - type: string - warehouse_id: - type: string - run_as_mode: - type: string - create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - update_mask: - type: required - query: - type: object - example: - id: fe25e731-92f2-4838-9fb2-1ca364320a3d - query: - description: Example description updated - display_name: Example query updated - query_text: SELECT 2 - update_mask: display_name,description,query_text - description: Updates a query. - /api/2.0/preview/sql/queries: - post: - operationId: queries-legacy-create - externalDocs: - url: https://docs.databricks.com/api/workspace/querieslegacy/create - x-stackQL-resource: queries_legacy - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Query created successfully - content: - application/json: - schema: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - data_source_id: - type: UUID - query: - type: string - name: - type: string - parent: - type: string - description: - type: string - options: - type: object - run_as_role: - type: string - tags: - type: Array of string - example: - data_source_id: 0c205e24-5db2-4940-adb1-fb13c7ce960b - query: SELECT field FROM table WHERE field = {{ param }} - name: Orders by month by customer - parent: folders/2025532471912059 - description: Summarizes total order dollars for customers in the Europe/Asia - region. - options: - parameters: - - title: customer - name: param - type: text - value: acme - run_as_role: viewer - tags: - - Payroll - description: A query object to be created. - description: Creates a new query definition. Queries created with this endpoint - belong to the authenticated user making the request. - get: - operationId: queries-legacy-list - externalDocs: - url: https://docs.databricks.com/api/workspace/querieslegacy/list - x-stackQL-resource: queries_legacy - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_size - in: query - - name: page - in: query - - name: order - in: query - - name: q - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - count: - type: integer - page: - type: integer - page_size: - type: integer - results: - type: array - items: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - x-stackQL-objectKey: $.results - description: Gets a list of queries. Optionally, this list can be filtered by - a search term. - /api/2.0/preview/sql/queries/{query_id}: - delete: - operationId: queries-legacy-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/querieslegacy/delete - x-stackQL-resource: queries_legacy - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: query_id - required: true - description: UUID - in: path - responses: - '200': - description: Query moved to trash. - content: - application/json: - schema: - type: object - properties: {} - description: Moves a query to the trash. Trashed queries immediately disappear - from searches and list views, and they cannot be used for alerts. The trash - is deleted after 30 days. - get: - operationId: queries-legacy-get - externalDocs: - url: https://docs.databricks.com/api/workspace/querieslegacy/get - x-stackQL-resource: queries_legacy - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: query_id - required: true - description: UUID - in: path - responses: - '200': - description: Query fetched successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - description: Retrieve a query object definition along with contextual permissions - information about the currently authenticated user. - post: - operationId: queries-legacy-update - externalDocs: - url: https://docs.databricks.com/api/workspace/querieslegacy/update - x-stackQL-resource: queries_legacy - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: query_id - required: true - description: UUID - in: path - responses: - '200': - description: Query updated successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: - type: object - properties: - id: - type: string - type: - type: string - name: - type: string - description: - type: string - options: - type: 'null' - updated_at: - type: string - created_at: - type: string - query: - type: object - properties: - id: - type: string - parent: - type: string - latest_query_data_id: - type: string - name: - type: string - description: - type: string - query: - type: string - query_hash: - type: string - is_archived: - type: boolean - is_draft: - type: boolean - updated_at: - type: string - created_at: - type: string - data_source_id: - type: string - options: - type: object - properties: - moved_to_trash_at: - type: string - catalog: - type: string - schema: - type: string - parameters: - type: array - items: - type: object - properties: - title: - type: string - name: - type: string - type: - type: string - enumOptions: - type: string - queryId: - type: string - multiValuesOptions: - type: object - properties: - prefix: - type: string - separator: - type: string - suffix: - type: string - value: - type: 'null' - tags: - type: array - items: - type: string - is_safe: - type: boolean - user_id: - type: integer - last_modified_by_id: - type: integer - visualizations: - type: array - items: {} - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - is_favorite: - type: boolean - user: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - last_modified_by: - type: object - properties: - id: - type: integer - name: - type: string - email: - type: string - can_edit: - type: boolean - permission_tier: - type: string - run_as_role: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - data_source_id: - type: UUID - query: - type: string - name: - type: string - description: - type: string - options: - type: object - run_as_role: - type: string - tags: - type: Array of string - example: - data_source_id: 0c205e24-5db2-4940-adb1-fb13c7ce960b - query: SELECT field FROM table WHERE field = {{ param }} - name: Orders by month by customer - description: Summarizes total order dollars for customers in the Europe/Asia - region. - options: - parameters: - - title: customer - name: param - type: text - value: acme - run_as_role: viewer - tags: - - Payroll - description: The query definition that will replace the current definition - for this - description: Modify this query definition. - /api/2.0/preview/sql/queries/trash/{query_id}: - post: - operationId: queries-legacy-restore - externalDocs: - url: https://docs.databricks.com/api/workspace/querieslegacy/restore - x-stackQL-resource: queries_legacy - x-stackQL-method: restore - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: query_id - required: true - description: UUID - in: path - responses: - '200': - description: Query restored successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Restore a query that has been moved to the trash. A restored query - appears in list views and searches. You can use restored queries for alerts. - /api/2.0/sql/history/queries: - get: - operationId: query-history-list - externalDocs: - url: https://docs.databricks.com/api/workspace/queryhistory/list - x-stackQL-resource: query_history - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: null - in: query - - name: max_results - in: query - - name: page_token - in: query - - name: include_metrics - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - next_page_token: - type: string - has_next_page: - type: boolean - res: - type: array - items: - type: object - properties: - query_id: - type: string - status: - type: string - query_text: - type: string - query_start_time_ms: - type: integer - execution_end_time_ms: - type: integer - query_end_time_ms: - type: integer - user_id: - type: integer - user_name: - type: string - spark_ui_url: - type: string - endpoint_id: - type: string - rows_produced: - type: integer - error_message: - type: string - lookup_key: - type: string - metrics: - type: object - properties: - total_time_ms: - type: integer - read_bytes: - type: integer - rows_produced_count: - type: integer - compilation_time_ms: - type: integer - execution_time_ms: - type: integer - read_remote_bytes: - type: integer - write_remote_bytes: - type: integer - read_cache_bytes: - type: integer - spill_to_disk_bytes: - type: integer - task_total_time_ms: - type: integer - read_files_count: - type: integer - read_partitions_count: - type: integer - photon_total_time_ms: - type: integer - rows_read_count: - type: integer - result_fetch_time_ms: - type: integer - network_sent_bytes: - type: integer - result_from_cache: - type: boolean - pruned_bytes: - type: integer - pruned_files_count: - type: integer - provisioning_queue_start_timestamp: - type: integer - overloading_queue_start_timestamp: - type: integer - query_compilation_start_timestamp: - type: integer - executed_as_user_id: - type: integer - executed_as_user_name: - type: string - is_final: - type: boolean - channel_used: - type: object - properties: - name: - type: string - dbsql_version: - type: string - plans_state: - type: string - statement_type: - type: string - warehouse_id: - type: string - duration: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '499': - description: Operation was canceled by the caller. - '500': - description: Internal error. - '503': - description: Service is currently unavailable. - x-stackQL-objectKey: $.res - description: List the history of queries through SQL warehouses, and serverless - compute. - /api/2.0/sql/statements/{statement_id}/cancel: - post: - operationId: statement-execution-cancelexecution - externalDocs: - url: https://docs.databricks.com/api/workspace/statementexecution/cancelexecution - x-stackQL-resource: statement_execution - x-stackQL-method: cancelexecution - x-stackQL-verb: exec - x-numReqParams: 2 - parameters: - - name: statement_id - required: true - description: string - in: path - - name: statement_id - required: true - description: string - in: path - responses: - '200': - description: Cancel response is empty; receiving response indicates successful - receipt. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - '503': - description: Service is currently unavailable. - description: Requests that an executing statement be canceled. Callers must - poll for status to see the terminal state. - /api/2.0/sql/statements/: - post: - operationId: statement-execution-executestatement - externalDocs: - url: https://docs.databricks.com/api/workspace/statementexecution/executestatement - x-stackQL-resource: statement_execution - x-stackQL-method: executestatement - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: StatementResponse contains `statement_id` and `status`; other - fields might be absent or present depending on context. If the SQL warehouse - fails to execute the provided statement, a 200 response is returned with - `status.state` set to `FAILED` (in constract to a failure when accepting - the request, which results in a non-200 response). Details of the error - can be found at `status.error` in case of execution failures. - content: - application/json: - schema: - type: object - properties: - statement_id: - type: string - status: - type: object - properties: - state: - type: string - manifest: - type: object - properties: - format: - type: string - schema: - type: object - properties: - column_count: - type: integer - columns: - type: array - items: - type: object - properties: - name: - type: string - position: - type: integer - type_name: - type: string - type_text: - type: string - result: - type: object - properties: - chunk_index: - type: integer - row_offset: - type: integer - row_count: - type: integer - data_array: - type: array - items: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - '503': - description: Service is currently unavailable. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - statement: - type: required - warehouse_id: - type: string - catalog: - type: required - schema: - type: string - row_limit: - type: string - byte_limit: - type: string - disposition: - type: integer - format: int64 - format: - type: integer - format: int64 - wait_timeout: - type: string - on_wait_timeout: - type: string - parameters: - type: Array of object - example: - warehouse_id: abcdef0123456789 - statement: SELECT * FROM range(100) - wait_timeout: 0s - description: Execute a SQL statement and optionally await its results for a - specified time. - /api/2.0/sql/statements/{statement_id}: - get: - operationId: statement-execution-getstatement - externalDocs: - url: https://docs.databricks.com/api/workspace/statementexecution/getstatement - x-stackQL-resource: statement_execution - x-stackQL-method: getstatement - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: statement_id - required: true - description: string - in: path - - name: statement_id - required: true - description: string - in: path - responses: - '200': - description: StatementResponse contains `statement_id` and `status`; other - fields might be absent or present depending on context. In case of an - error during execution of the SQL statement -- as opposed to an error - while processing the request -- a 200 response is returned with error - details in the `status` field. - content: - application/json: - schema: - type: object - properties: - statement_id: - type: string - status: - type: object - properties: - state: - type: string - manifest: - type: object - properties: - format: - type: string - schema: - type: object - properties: - column_count: - type: integer - columns: - type: array - items: - type: object - properties: - name: - type: string - position: - type: integer - type_name: - type: string - type_text: - type: string - result: - type: object - properties: - chunk_index: - type: integer - row_offset: - type: integer - row_count: - type: integer - data_array: - type: array - items: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - '503': - description: Service is currently unavailable. - description: This request can be used to poll for the statement's status. When - the - /api/2.0/sql/statements/{statement_id}/result/chunks/{chunk_index}: - get: - operationId: statement-execution-results-getstatementresultchunkn - externalDocs: - url: https://docs.databricks.com/api/workspace/statementexecution/getstatementresultchunkn - x-stackQL-resource: statement_execution_results - x-stackQL-method: getstatementresultchunkn - x-stackQL-verb: select - x-numReqParams: 4 - parameters: - - name: statement_id - required: true - description: string - in: path - - name: chunk_index - required: true - description: int32 - in: path - - name: statement_id - required: true - description: string - in: path - - name: chunk_index - required: true - description: string - in: path - responses: - '200': - description: Successful return; depending on `disposition` returns chunks - of data either inline, or as links. - content: - application/json: - schema: - type: object - properties: - chunk_index: - type: integer - row_offset: - type: integer - row_count: - type: integer - next_chunk_index: - type: integer - next_chunk_internal_link: - type: string - data_array: - type: array - items: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - '503': - description: Service is currently unavailable. - description: After the statement execution has - /api/2.0/sql/warehouses: - post: - operationId: warehouses-create - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/create - x-stackQL-resource: warehouses - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - cluster_size: - type: string - min_num_clusters: - type: int32 - max_num_clusters: - type: int32 - auto_stop_mins: - type: int32 - creator_name: - type: string - instance_profile_arn: - type: string - spot_instance_policy: - type: string - enable_photon: - type: boolean - enable_serverless_compute: - type: boolean - warehouse_type: - type: string - tags: - type: object - channel: - type: object - example: - name: string - cluster_size: string - min_num_clusters: '1' - max_num_clusters: 0 - auto_stop_mins: '120' - creator_name: string - instance_profile_arn: string - tags: - custom_tags: - - key: string - value: string - spot_instance_policy: POLICY_UNSPECIFIED - enable_photon: true - channel: - name: CHANNEL_NAME_PREVIEW - dbsql_version: string - enable_serverless_compute: true - warehouse_type: TYPE_UNSPECIFIED - description: Creates a new SQL warehouse. - get: - operationId: warehouses-list - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/list - x-stackQL-resource: warehouses - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: run_as_user_id - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - warehouses: - type: array - items: - type: object - properties: - id: - type: string - name: - type: string - cluster_size: - type: string - min_num_clusters: - type: string - max_num_clusters: - type: integer - auto_stop_mins: - type: string - creator_name: - type: string - instance_profile_arn: - type: string - tags: - type: object - properties: - custom_tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - spot_instance_policy: - type: string - enable_photon: - type: boolean - channel: - type: object - properties: - name: - type: string - dbsql_version: - type: string - enable_serverless_compute: - type: boolean - warehouse_type: - type: string - num_clusters: - type: integer - num_active_sessions: - type: integer - state: - type: string - jdbc_url: - type: string - odbc_params: - type: object - properties: - hostname: - type: string - path: - type: string - protocol: - type: string - port: - type: integer - health: - type: object - properties: - status: - type: string - message: - type: string - failure_reason: - type: object - properties: - code: - type: string - type: - type: string - parameters: - type: object - properties: - property1: - type: string - property2: - type: string - summary: - type: string - details: - type: string - x-stackQL-objectKey: $.warehouses - description: Lists all SQL warehouses that a user has manager permissions on. - /api/2.0/sql/warehouses/{id}: - delete: - operationId: warehouses-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/delete - x-stackQL-resource: warehouses - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a SQL warehouse. - get: - operationId: warehouses-get - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/get - x-stackQL-resource: warehouses - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - id: - type: string - name: - type: string - cluster_size: - type: string - min_num_clusters: - type: string - max_num_clusters: - type: integer - auto_stop_mins: - type: string - creator_name: - type: string - instance_profile_arn: - type: string - tags: - type: object - properties: - custom_tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - spot_instance_policy: - type: string - enable_photon: - type: boolean - channel: - type: object - properties: - name: - type: string - dbsql_version: - type: string - enable_serverless_compute: - type: boolean - warehouse_type: - type: string - num_clusters: - type: integer - num_active_sessions: - type: integer - state: - type: string - jdbc_url: - type: string - odbc_params: - type: object - properties: - hostname: - type: string - path: - type: string - protocol: - type: string - port: - type: integer - health: - type: object - properties: - status: - type: string - message: - type: string - failure_reason: - type: object - properties: - code: - type: string - type: - type: string - parameters: - type: object - properties: - property1: - type: string - property2: - type: string - summary: - type: string - details: - type: string - description: Gets the information for a single SQL warehouse. - /api/2.0/sql/warehouses/{id}/edit: - post: - operationId: warehouses-edit - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/edit - x-stackQL-resource: warehouses - x-stackQL-method: edit - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - cluster_size: - type: string - min_num_clusters: - type: int32 - max_num_clusters: - type: int32 - auto_stop_mins: - type: int32 - creator_name: - type: string - instance_profile_arn: - type: string - spot_instance_policy: - type: string - enable_photon: - type: boolean - enable_serverless_compute: - type: boolean - warehouse_type: - type: string - tags: - type: object - channel: - type: object - example: - name: string - cluster_size: string - min_num_clusters: '1' - max_num_clusters: 0 - auto_stop_mins: '120' - creator_name: string - instance_profile_arn: string - tags: - custom_tags: - - key: string - value: string - spot_instance_policy: POLICY_UNSPECIFIED - enable_photon: true - channel: - name: CHANNEL_NAME_PREVIEW - dbsql_version: string - enable_serverless_compute: true - warehouse_type: TYPE_UNSPECIFIED - description: Updates the configuration for a SQL warehouse. - /api/2.0/sql/warehouses/{id}/start: - post: - operationId: warehouses-start - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/start - x-stackQL-resource: warehouses - x-stackQL-method: start - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - description: Starts a SQL warehouse. - /api/2.0/sql/warehouses/{id}/stop: - post: - operationId: warehouses-stop - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/stop - x-stackQL-resource: warehouses - x-stackQL-method: stop - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - description: Stops a SQL warehouse. - /api/2.0/sql/config/warehouses: - get: - operationId: warehouse-config-getworkspacewarehouseconfig - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/getworkspacewarehouseconfig - x-stackQL-resource: warehouse_config - x-stackQL-method: getworkspacewarehouseconfig - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - security_policy: - type: string - data_access_config: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - instance_profile_arn: - type: string - channel: - type: object - properties: - name: - type: string - dbsql_version: - type: string - global_param: - type: object - properties: - config_pair: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - configuration_pairs: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - config_param: - type: object - properties: - config_pair: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - configuration_pairs: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - sql_configuration_parameters: - type: object - properties: - config_pair: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - configuration_pairs: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - google_service_account: - type: string - enabled_warehouse_types: - type: array - items: - type: object - properties: - warehouse_type: - type: string - enabled: - type: boolean - description: Gets the workspace level configuration that is shared by all SQL - warehouses in a workspace. - put: - operationId: warehouse-config-setworkspacewarehouseconfig - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/setworkspacewarehouseconfig - x-stackQL-resource: warehouse_config - x-stackQL-method: setworkspacewarehouseconfig - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - security_policy: - type: string - instance_profile_arn: - type: string - google_service_account: - type: string - data_access_config: - type: Array of object - channel: - type: object - global_param: - type: object - config_param: - type: object - sql_configuration_parameters: - type: object - enabled_warehouse_types: - type: Array of object - example: - security_policy: NONE - data_access_config: - - key: string - value: string - instance_profile_arn: string - channel: - name: CHANNEL_NAME_PREVIEW - dbsql_version: string - global_param: - config_pair: - - key: string - value: string - configuration_pairs: - - key: string - value: string - config_param: - config_pair: - - key: string - value: string - configuration_pairs: - - key: string - value: string - sql_configuration_parameters: - config_pair: - - key: string - value: string - configuration_pairs: - - key: string - value: string - google_service_account: string - enabled_warehouse_types: - - warehouse_type: TYPE_UNSPECIFIED - enabled: true - description: Sets the workspace level configuration that is shared by all SQL - warehouses in a workspace. - /api/2.0/permissions/warehouses/{warehouse_id}: - get: - operationId: warehouse-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/getpermissions - x-stackQL-resource: warehouse_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: warehouse_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a SQL warehouse. SQL warehouses can inherit - permissions from their root object. - put: - operationId: warehouse-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/setpermissions - x-stackQL-resource: warehouse_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: warehouse_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: warehouse-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/updatepermissions - x-stackQL-resource: warehouse_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: warehouse_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a SQL warehouse. SQL warehouses can - inherit permissions from their root object. - /api/2.0/permissions/warehouses/{warehouse_id}/permissionLevels: - get: - operationId: warehouse-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/warehouses/getpermissionlevels - x-stackQL-resource: warehouse_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: warehouse_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. -components: - x-stackQL-resources: - alerts: - id: databricks_workspace.dbsql.alerts - name: alerts - title: Alerts - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1sql~1alerts/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1sql~1alerts/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.results - delete: - operation: - $ref: '#/paths/~1api~12.0~1sql~1alerts~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1sql~1alerts~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1sql~1alerts~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/alerts/methods/get' - - $ref: '#/components/x-stackQL-resources/alerts/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/alerts/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/alerts/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/alerts/methods/delete' - exec: [] - alerts_legacy: - id: databricks_workspace.dbsql.alerts_legacy - name: alerts_legacy - title: Alerts_legacy - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts~1{alert_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts~1{alert_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts~1{alert_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/get' - - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/delete' - exec: [] - dashboards: - id: databricks_workspace.dbsql.dashboards - name: dashboards - title: Dashboards - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.results - delete: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards~1{dashboard_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards~1{dashboard_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards~1{dashboard_id}/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - restore: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards~1trash~1{dashboard_id}/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/get' - - $ref: '#/components/x-stackQL-resources/dashboards/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/restore' - permissions: - id: databricks_workspace.dbsql.permissions - name: permissions - title: Permissions - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1permissions~1{objectType}~1{objectId}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - set: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1permissions~1{objectType}~1{objectId}/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - transferownership: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1permissions~1{objectType}~1{objectId}~1transfer/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permissions/methods/get' - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/permissions/methods/set' - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/permissions/methods/transferownership' - queries: - id: databricks_workspace.dbsql.queries - name: queries - title: Queries - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1sql~1queries/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1sql~1queries/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.results - delete: - operation: - $ref: '#/paths/~1api~12.0~1sql~1queries~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1sql~1queries~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1sql~1queries~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/queries/methods/get' - - $ref: '#/components/x-stackQL-resources/queries/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/queries/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/queries/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/queries/methods/delete' - exec: [] - queries_legacy: - id: databricks_workspace.dbsql.queries_legacy - name: queries_legacy - title: Queries_legacy - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1queries/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1queries/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.results - delete: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1{query_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1{query_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1{query_id}/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - restore: - operation: - $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1trash~1{query_id}/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/get' - - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/restore' - query_history: - id: databricks_workspace.dbsql.query_history - name: query_history - title: Query_history - methods: - list: - operation: - $ref: '#/paths/~1api~12.0~1sql~1history~1queries/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.res - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/query_history/methods/list' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - statement_execution: - id: databricks_workspace.dbsql.statement_execution - name: statement_execution - title: Statement_execution - methods: - cancelexecution: - operation: - $ref: '#/paths/~1api~12.0~1sql~1statements~1{statement_id}~1cancel/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - executestatement: - operation: - $ref: '#/paths/~1api~12.0~1sql~1statements~1/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getstatement: - operation: - $ref: '#/paths/~1api~12.0~1sql~1statements~1{statement_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/statement_execution/methods/getstatement' - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/statement_execution/methods/executestatement' - - $ref: '#/components/x-stackQL-resources/statement_execution/methods/cancelexecution' - statement_execution_results: - id: databricks_workspace.dbsql.statement_execution_results - name: statement_execution_results - title: Statement_execution_results - methods: - getstatementresultchunkn: - operation: - $ref: '#/paths/~1api~12.0~1sql~1statements~1{statement_id}~1result~1chunks~1{chunk_index}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/statement_execution_results/methods/getstatementresultchunkn' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - warehouses: - id: databricks_workspace.dbsql.warehouses - name: warehouses - title: Warehouses - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1sql~1warehouses/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1sql~1warehouses/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.warehouses - delete: - operation: - $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - edit: - operation: - $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}~1edit/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - start: - operation: - $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}~1start/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - stop: - operation: - $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}~1stop/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/warehouses/methods/get' - - $ref: '#/components/x-stackQL-resources/warehouses/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/warehouses/methods/create' - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/warehouses/methods/edit' - delete: - - $ref: '#/components/x-stackQL-resources/warehouses/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/warehouses/methods/stop' - - $ref: '#/components/x-stackQL-resources/warehouses/methods/start' - warehouse_config: - id: databricks_workspace.dbsql.warehouse_config - name: warehouse_config - title: Warehouse_config - methods: - getworkspacewarehouseconfig: - operation: - $ref: '#/paths/~1api~12.0~1sql~1config~1warehouses/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setworkspacewarehouseconfig: - operation: - $ref: '#/paths/~1api~12.0~1sql~1config~1warehouses/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/warehouse_config/methods/getworkspacewarehouseconfig' - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/warehouse_config/methods/setworkspacewarehouseconfig' - delete: [] - exec: [] - warehouse_permissions: - id: databricks_workspace.dbsql.warehouse_permissions - name: warehouse_permissions - title: Warehouse_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/warehouse_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/warehouse_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/warehouse_permissions/methods/setpermissions' - delete: [] - exec: [] - warehouse_permission_levels: - id: databricks_workspace.dbsql.warehouse_permission_levels - name: warehouse_permission_levels - title: Warehouse_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/warehouse_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/deltalivetables.yaml b/providers/src/databricks_workspace/v00.00.00000/services/deltalivetables.yaml deleted file mode 100644 index b9e362d1..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/deltalivetables.yaml +++ /dev/null @@ -1,1619 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Deltalivetables API - description: Deltalivetables -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/pipelines: - post: - operationId: pipelines-create - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/create - x-stackQL-resource: pipelines - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - pipeline_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: string - name: - type: string - storage: - type: string - target: - type: string - schema: - type: string - continuous: - type: boolean - development: - type: boolean - photon: - type: boolean - edition: - type: string - channel: - type: string - catalog: - type: string - serverless: - type: boolean - allow_duplicate_names: - type: boolean - dry_run: - type: boolean - configuration: - type: object - clusters: - type: Array of object - libraries: - type: Array of object - trigger: - type: object - filters: - type: object - notifications: - type: Array of object - deployment: - type: object - ingestion_definition: - type: object - example: - continuous: false - name: Wikipedia pipeline (SQL) - clusters: - - label: default - autoscale: - min_workers: 1 - max_workers: 5 - mode: ENHANCED - libraries: - - notebook: - path: /Users/username/DLT Notebooks/Delta Live Tables quickstart - (SQL) - storage: /Users/username/data - description: Creates a new data processing pipeline based on the requested configuration. - If successful, this method returns the ID of the new pipeline. - get: - operationId: pipelines-listpipelines - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/listpipelines - x-stackQL-resource: pipelines - x-stackQL-method: listpipelines - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - - name: order_by - in: query - - name: filter - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - statuses: - type: array - items: - type: object - properties: - name: - type: string - pipeline_id: - type: string - state: - type: string - latest_updates: - type: array - items: - type: object - properties: - update_id: - type: string - state: - type: string - creation_time: - type: string - creator_user_name: - type: string - next_page_token: - type: string - prev_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.statuses - description: Lists pipelines defined in the Delta Live Tables system. - /api/2.0/pipelines/{pipeline_id}: - delete: - operationId: pipelines-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/delete - x-stackQL-resource: pipelines - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a pipeline. - get: - operationId: pipelines-get - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/get - x-stackQL-resource: pipelines - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - name: - type: string - pipeline_id: - type: string - state: - type: string - latest_updates: - type: array - items: - type: object - properties: - update_id: - type: string - state: - type: string - creation_time: - type: string - creator_user_name: - type: string - spec: - type: object - properties: - continuous: - type: boolean - name: - type: string - deployment: - type: object - properties: - kind: - type: string - metadata_file_path: - type: string - clusters: - type: array - items: - type: object - properties: - label: - type: string - autoscale: - type: object - properties: - min_workers: - type: integer - max_workers: - type: integer - mode: - type: string - id: - type: string - libraries: - type: array - items: - type: object - properties: - notebook: - type: object - properties: - path: - type: string - storage: - type: string - target: - type: string - run_as_user_name: - type: string - cluster_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - put: - operationId: pipelines-update - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/update - x-stackQL-resource: pipelines - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: string - name: - type: string - storage: - type: string - target: - type: string - schema: - type: string - continuous: - type: boolean - development: - type: boolean - photon: - type: boolean - edition: - type: string - channel: - type: string - catalog: - type: string - serverless: - type: boolean - pipeline_id: - type: string - allow_duplicate_names: - type: boolean - expected_last_modified: - type: integer - format: int64 - configuration: - type: object - clusters: - type: Array of object - libraries: - type: Array of object - trigger: - type: object - filters: - type: object - notifications: - type: Array of object - deployment: - type: object - ingestion_definition: - type: object - example: - continuous: false - name: Wikipedia pipeline (SQL) - clusters: - - label: default - autoscale: - min_workers: 1 - max_workers: 5 - mode: ENHANCED - id: a12cd3e4-0ab1-1abc-1a2b-1a2bcd3e4fg5 - libraries: - - notebook: - path: /Users/username/DLT Notebooks/Delta Live Tables quickstart - (SQL) - storage: /Users/username/data - target: wikipedia_quickstart_data - description: Updates a pipeline with the supplied configuration. - /api/2.0/pipelines/{pipeline_id}/stop: - post: - operationId: pipelines-stop - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/stop - x-stackQL-resource: pipelines - x-stackQL-method: stop - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Stops the pipeline by canceling the active update. If there is - no active update for the pipeline, this request is a no-op. - /api/2.0/pipelines/{pipeline_id}/events: - get: - operationId: pipeline-events-listpipelineevents - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/listpipelineevents - x-stackQL-resource: pipeline_events - x-stackQL-method: listpipelineevents - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - - name: max_results - in: query - - name: order_by - in: query - - name: filter - in: query - - name: page_token - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - next_page_token: - type: string - prev_page_token: - type: string - events: - type: array - items: - type: object - properties: - id: - type: string - sequence: - type: object - properties: - data_plane_id: - type: object - properties: - instance: - type: string - seq_no: - type: integer - control_plane_seq_no: - type: integer - origin: - type: object - properties: - cloud: - type: string - region: - type: string - org_id: - type: integer - pipeline_id: - type: string - pipeline_name: - type: string - cluster_id: - type: string - update_id: - type: string - maintenance_id: - type: string - table_id: - type: string - dataset_name: - type: string - flow_id: - type: string - flow_name: - type: string - batch_id: - type: integer - request_id: - type: string - uc_resource_id: - type: string - host: - type: string - materialization_name: - type: string - timestamp: - type: string - message: - type: string - level: - type: string - error: - type: object - properties: - fatal: - type: boolean - exceptions: - type: array - items: - type: object - properties: - class_name: - type: string - message: - type: string - stack: - type: array - items: - type: object - properties: - declaring_class: - type: string - method_name: - type: string - file_name: - type: string - line_number: - type: integer - event_type: - type: string - maturity_level: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.events - description: Retrieves events for a pipeline. - /api/2.0/permissions/pipelines/{pipeline_id}: - get: - operationId: pipeline-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/getpermissions - x-stackQL-resource: pipeline_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a pipeline. Pipelines can inherit permissions - from their root object. - put: - operationId: pipeline-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/setpermissions - x-stackQL-resource: pipeline_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: pipeline-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/updatepermissions - x-stackQL-resource: pipeline_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a pipeline. Pipelines can inherit permissions - from their root object. - /api/2.0/permissions/pipelines/{pipeline_id}/permissionLevels: - get: - operationId: pipeline-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/getpermissionlevels - x-stackQL-resource: pipeline_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/pipelines/{pipeline_id}/updates/{update_id}: - get: - operationId: pipeline-updates-getupdate - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/getupdate - x-stackQL-resource: pipeline_updates - x-stackQL-method: getupdate - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: pipeline_id - required: true - description: string - in: path - - name: update_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - update: - type: object - properties: - creation_time: - type: integer - pipeline_id: - type: string - state: - type: string - update_id: - type: string - config: - type: object - properties: - continuous: - type: boolean - name: - type: string - clusters: - type: array - items: - type: object - properties: - label: - type: string - autoscale: - type: object - properties: - min_workers: - type: integer - max_workers: - type: integer - mode: - type: string - id: - type: string - libraries: - type: array - items: - type: object - properties: - notebook: - type: object - properties: - path: - type: string - configuration: - type: object - properties: - pipelines.numStreamRetryAttempts: - type: string - storage: - type: string - target: - type: string - development: - type: boolean - full_refresh: - type: boolean - cause: - type: string - request_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.update - description: Gets an update from an active pipeline. - /api/2.0/pipelines/{pipeline_id}/updates: - get: - operationId: pipeline-updates-listupdates - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/listupdates - x-stackQL-resource: pipeline_updates - x-stackQL-method: listupdates - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - - name: page_token - in: query - - name: max_results - in: query - - name: until_update_id - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - updates: - type: array - items: - type: object - properties: - pipeline_id: - type: string - update_id: - type: string - config: - type: object - properties: - id: - type: string - name: - type: string - storage: - type: string - configuration: - type: object - properties: - property1: - type: string - property2: - type: string - clusters: - type: array - items: - type: object - properties: - label: - type: string - spark_conf: - type: object - properties: - spark.speculation: - type: boolean - spark.streaming.ui.retainedBatches: - type: integer - aws_attributes: - type: object - properties: - first_on_demand: - type: string - availability: - type: string - zone_id: - type: string - instance_profile_arn: - type: string - spot_bid_price_percent: - type: string - ebs_volume_type: - type: string - ebs_volume_count: - type: string - ebs_volume_size: - type: integer - ebs_volume_iops: - type: integer - ebs_volume_throughput: - type: integer - node_type_id: - type: string - driver_node_type_id: - type: string - ssh_public_keys: - type: array - items: - type: string - custom_tags: - type: object - properties: - property1: - type: string - property2: - type: string - cluster_log_conf: - type: object - properties: - dbfs: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - init_scripts: - type: array - items: - type: object - properties: - workspace: - type: object - properties: - destination: - type: string - volumes: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - file: - type: object - properties: - destination: - type: string - dbfs: - type: object - properties: - destination: - type: string - abfss: - type: object - properties: - destination: - type: string - gcs: - type: object - properties: - destination: - type: string - spark_env_vars: - type: object - properties: - property1: - type: string - property2: - type: string - instance_pool_id: - type: string - driver_instance_pool_id: - type: string - policy_id: - type: string - num_workers: - type: integer - autoscale: - type: object - properties: - min_workers: - type: integer - max_workers: - type: integer - mode: - type: string - apply_policy_default_values: - type: boolean - enable_local_disk_encryption: - type: boolean - libraries: - type: array - items: - type: object - properties: - notebook: - type: object - properties: - path: - type: string - file: - type: object - properties: - path: - type: string - whl: - type: string - trigger: - type: object - properties: - manual: - type: object - properties: {} - cron: - type: object - properties: - quartz_cron_schedule: - type: string - timezone_id: - type: string - target: - type: string - schema: - type: string - filters: - type: object - properties: - include: - type: array - items: - type: string - exclude: - type: array - items: - type: string - continuous: - type: boolean - development: - type: boolean - photon: - type: boolean - edition: - type: string - channel: - type: string - catalog: - type: string - notifications: - type: array - items: - type: object - properties: - email_recipients: - type: array - items: - type: string - alerts: - type: array - items: - type: string - serverless: - type: boolean - deployment: - type: object - properties: - kind: - type: string - metadata_file_path: - type: string - ingestion_definition: - type: object - properties: - connection_name: - type: string - ingestion_gateway_id: - type: string - objects: - type: array - items: - type: object - properties: - schema: - type: object - properties: - source_catalog: - type: string - source_schema: - type: string - destination_catalog: - type: string - destination_schema: - type: string - table_configuration: - type: object - properties: - primary_keys: - type: array - items: - type: string - sequence_by: - type: array - items: - type: string - table: - type: object - properties: - source_catalog: - type: string - source_schema: - type: string - source_table: - type: string - destination_catalog: - type: string - destination_schema: - type: string - destination_table: - type: string - table_configuration: - type: object - properties: - primary_keys: - type: array - items: - type: string - sequence_by: - type: array - items: - type: string - report: - type: object - properties: - source_url: - type: string - destination_catalog: - type: string - destination_schema: - type: string - destination_table: - type: string - table_configuration: - type: object - properties: - primary_keys: - type: array - items: - type: string - sequence_by: - type: array - items: - type: string - table_configuration: - type: object - properties: - primary_keys: - type: array - items: - type: string - sequence_by: - type: array - items: - type: string - cause: - type: string - state: - type: string - cluster_id: - type: string - creation_time: - type: integer - validate_only: - type: boolean - full_refresh: - type: boolean - refresh_selection: - type: array - items: - type: string - full_refresh_selection: - type: array - items: - type: string - next_page_token: - type: string - prev_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.updates - description: List updates for an active pipeline. - post: - operationId: pipeline-updates-startupdate - externalDocs: - url: https://docs.databricks.com/api/workspace/pipelines/startupdate - x-stackQL-resource: pipeline_updates - x-stackQL-method: startupdate - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: pipeline_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - update_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - full_refresh: - type: boolean - validate_only: - type: boolean - cause: - type: string - refresh_selection: - type: Array of string - full_refresh_selection: - type: Array of string - example: - full_refresh: true - description: Starts a new update for the pipeline. If there is already an active - update for the pipeline, the request will fail and the active update will - remain running. -components: - x-stackQL-resources: - pipelines: - id: databricks_workspace.deltalivetables.pipelines - name: pipelines - title: Pipelines - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1pipelines/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listpipelines: - operation: - $ref: '#/paths/~1api~12.0~1pipelines/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.statuses - delete: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - stop: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1stop/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/pipelines/methods/get' - - $ref: '#/components/x-stackQL-resources/pipelines/methods/listpipelines' - insert: - - $ref: '#/components/x-stackQL-resources/pipelines/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/pipelines/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/pipelines/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/pipelines/methods/stop' - pipeline_events: - id: databricks_workspace.deltalivetables.pipeline_events - name: pipeline_events - title: Pipeline_events - methods: - listpipelineevents: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1events/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.events - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/pipeline_events/methods/listpipelineevents' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - pipeline_permissions: - id: databricks_workspace.deltalivetables.pipeline_permissions - name: pipeline_permissions - title: Pipeline_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/pipeline_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/pipeline_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/pipeline_permissions/methods/setpermissions' - delete: [] - exec: [] - pipeline_permission_levels: - id: databricks_workspace.deltalivetables.pipeline_permission_levels - name: pipeline_permission_levels - title: Pipeline_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/pipeline_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - pipeline_updates: - id: databricks_workspace.deltalivetables.pipeline_updates - name: pipeline_updates - title: Pipeline_updates - methods: - getupdate: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1updates~1{update_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.update - listupdates: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1updates/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.updates - startupdate: - operation: - $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1updates/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/pipeline_updates/methods/listupdates' - - $ref: '#/components/x-stackQL-resources/pipeline_updates/methods/getupdate' - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/pipeline_updates/methods/startupdate' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/deltasharing.yaml b/providers/src/databricks_workspace/v00.00.00000/services/deltasharing.yaml deleted file mode 100644 index 79240229..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/deltasharing.yaml +++ /dev/null @@ -1,1768 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Deltasharing API - description: Deltasharing -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.1/unity-catalog/providers: - post: - operationId: providers-create - externalDocs: - url: https://docs.databricks.com/api/workspace/providers/create - x-stackQL-resource: providers - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new provider was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - authentication_type: - type: string - recipient_profile_str: - type: string - comment: - type: string - owner: - type: string - recipient_profile: - type: object - properties: - share_credentials_version: - type: integer - endpoint: - type: string - bearer_token: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - data_provider_global_metastore_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - authentication_type: - type: string - comment: - type: required - recipient_profile_str: - type: string - example: - name: string - authentication_type: TOKEN - comment: string - recipient_profile_str: string - description: Creates a new authentication provider minimally based on a name - and authentication type. The caller must be an admin on the metastore. - get: - operationId: providers-list - externalDocs: - url: https://docs.databricks.com/api/workspace/providers/list - x-stackQL-resource: providers - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: data_provider_global_metastore_id - in: query - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The provider list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - providers: - type: array - items: - type: object - properties: - name: - type: string - authentication_type: - type: string - recipient_profile_str: - type: string - comment: - type: string - owner: - type: string - recipient_profile: - type: object - properties: - share_credentials_version: - type: integer - endpoint: - type: string - bearer_token: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - data_provider_global_metastore_id: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.providers - description: Gets an array of available authentication providers. The caller - must either be a metastore admin or the owner of the providers. Providers - not owned by the caller are not included in the response. There is no guarantee - of a specific ordering of the elements in the array. - /api/2.1/unity-catalog/providers/{name}: - delete: - operationId: providers-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/providers/delete - x-stackQL-resource: providers - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The provider was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes an authentication provider, if the caller is a metastore - admin or is the owner of the provider. - get: - operationId: providers-get - externalDocs: - url: https://docs.databricks.com/api/workspace/providers/get - x-stackQL-resource: providers - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The provider was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - authentication_type: - type: string - recipient_profile_str: - type: string - comment: - type: string - owner: - type: string - recipient_profile: - type: object - properties: - share_credentials_version: - type: integer - endpoint: - type: string - bearer_token: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - data_provider_global_metastore_id: - type: string - description: Gets a specific authentication provider. The caller must supply - the name of the provider, and must either be a metastore admin or the owner - of the provider. - patch: - operationId: providers-update - externalDocs: - url: https://docs.databricks.com/api/workspace/providers/update - x-stackQL-resource: providers - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The provider was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - authentication_type: - type: string - recipient_profile_str: - type: string - comment: - type: string - owner: - type: string - recipient_profile: - type: object - properties: - share_credentials_version: - type: integer - endpoint: - type: string - bearer_token: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - data_provider_global_metastore_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - recipient_profile_str: - type: string - comment: - type: string - owner: - type: string - new_name: - type: string - example: - recipient_profile_str: string - comment: string - owner: string - new_name: string - description: Updates the information for an authentication provider, if the - caller is a metastore admin or is the owner of the provider. If the update - changes the provider name, the caller must be both a metastore admin and the - owner of the provider. - /api/2.1/unity-catalog/providers/{name}/shares: - get: - operationId: provider-shares-listshares - externalDocs: - url: https://docs.databricks.com/api/workspace/providers/listshares - x-stackQL-resource: provider_shares - x-stackQL-method: listshares - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The provider shares list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - shares: - type: array - items: - type: object - properties: - name: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.shares - description: 'Gets an array of a specified provider''s shares within the metastore - where:' - /api/2.1/unity-catalog/recipients: - post: - operationId: recipients-create - externalDocs: - url: https://docs.databricks.com/api/workspace/recipients/create - x-stackQL-resource: recipients - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new recipient was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - authentication_type: - type: string - sharing_code: - type: string - data_recipient_global_metastore_id: - type: string - owner: - type: string - comment: - type: string - ip_access_list: - type: object - properties: - allowed_ip_addresses: - type: array - items: - type: string - properties_kvpairs: - type: object - properties: - properties: - type: object - properties: - property1: - type: string - property2: - type: string - activation_url: - type: string - activated: - type: boolean - created_at: - type: integer - created_by: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - authentication_type: - type: string - owner: - type: required - comment: - type: string - data_recipient_global_metastore_id: - type: string - sharing_code: - type: string - expiration_time: - type: string - ip_access_list: - type: object - properties_kvpairs: - type: object - example: - name: string - authentication_type: TOKEN - owner: string - comment: string - ip_access_list: - allowed_ip_addresses: - - string - properties_kvpairs: - properties: - property1: string - property2: string - data_recipient_global_metastore_id: string - sharing_code: string - expiration_time: 0 - description: Creates a new recipient with the delta sharing authentication type - in the metastore. The caller must be a metastore admin or has the - get: - operationId: recipients-list - externalDocs: - url: https://docs.databricks.com/api/workspace/recipients/list - x-stackQL-resource: recipients - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: data_recipient_global_metastore_id - in: query - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The list of recipients was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - recipients: - type: array - items: - type: object - properties: - name: - type: string - authentication_type: - type: string - sharing_code: - type: string - data_recipient_global_metastore_id: - type: string - owner: - type: string - comment: - type: string - ip_access_list: - type: object - properties: - allowed_ip_addresses: - type: array - items: - type: string - properties_kvpairs: - type: object - properties: - properties: - type: object - properties: - property1: - type: string - property2: - type: string - activation_url: - type: string - activated: - type: boolean - created_at: - type: integer - created_by: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.recipients - description: 'Gets an array of all share recipients within the current metastore - where:' - /api/2.1/unity-catalog/recipients/{name}: - delete: - operationId: recipients-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/recipients/delete - x-stackQL-resource: recipients - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The recipient was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes the specified recipient from the metastore. The caller - must be the owner of the recipient. - get: - operationId: recipients-get - externalDocs: - url: https://docs.databricks.com/api/workspace/recipients/get - x-stackQL-resource: recipients - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The recipient was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - authentication_type: - type: string - sharing_code: - type: string - data_recipient_global_metastore_id: - type: string - owner: - type: string - comment: - type: string - ip_access_list: - type: object - properties: - allowed_ip_addresses: - type: array - items: - type: string - properties_kvpairs: - type: object - properties: - properties: - type: object - properties: - property1: - type: string - property2: - type: string - activation_url: - type: string - activated: - type: boolean - created_at: - type: integer - created_by: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - description: 'Gets a share recipient from the metastore if:' - patch: - operationId: recipients-update - externalDocs: - url: https://docs.databricks.com/api/workspace/recipients/update - x-stackQL-resource: recipients - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The recipient was successfully updated. - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - comment: - type: string - new_name: - type: string - expiration_time: - type: integer - format: int64 - ip_access_list: - type: object - properties_kvpairs: - type: object - example: - owner: string - comment: string - ip_access_list: - allowed_ip_addresses: - - string - new_name: string - properties_kvpairs: - properties: - property1: string - property2: string - expiration_time: 0 - description: Updates an existing recipient in the metastore. The caller must - be a metastore admin or the owner of the recipient. If the recipient name - will be updated, the user must be both a metastore admin and the owner of - the recipient. - /api/2.1/unity-catalog/recipients/{name}/rotate-token: - post: - operationId: recipients-rotatetoken - externalDocs: - url: https://docs.databricks.com/api/workspace/recipients/rotatetoken - x-stackQL-resource: recipients - x-stackQL-method: rotatetoken - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The recipient token was successfully rotated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - authentication_type: - type: string - sharing_code: - type: string - data_recipient_global_metastore_id: - type: string - owner: - type: string - comment: - type: string - ip_access_list: - type: object - properties: - allowed_ip_addresses: - type: array - items: - type: string - properties_kvpairs: - type: object - properties: - properties: - type: object - properties: - property1: - type: string - property2: - type: string - activation_url: - type: string - activated: - type: boolean - created_at: - type: integer - created_by: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - updated_at: - type: integer - updated_by: - type: string - cloud: - type: string - region: - type: string - metastore_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - existing_token_expire_in_seconds: - type: required - example: - existing_token_expire_in_seconds: 0 - description: Refreshes the specified recipient's delta sharing authentication - token with the provided token info. The caller must be the owner of the recipient. - /api/2.1/unity-catalog/public/data_sharing_activation/{activation_url}: - get: - operationId: recipient-activation-token-retrievetoken - externalDocs: - url: https://docs.databricks.com/api/workspace/recipientactivation/retrievetoken - x-stackQL-resource: recipient_activation_token - x-stackQL-method: retrievetoken - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: activation_url - required: true - description: string - in: path - responses: - '200': - description: The access token was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - shareCredentialsVersion: - type: integer - bearerToken: - type: string - endpoint: - type: string - expirationTime: - type: string - description: Retrieve access token with an activation url. This is a public - API without any authentication. - /api/2.1/unity-catalog/public/data_sharing_activation_info/{activation_url}: - get: - operationId: recipient-activation-url-getactivationurlinfo - externalDocs: - url: https://docs.databricks.com/api/workspace/recipientactivation/getactivationurlinfo - x-stackQL-resource: recipient_activation_url - x-stackQL-method: getactivationurlinfo - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: activation_url - required: true - description: string - in: path - responses: - '200': - description: The activation URL fetch was successful. - content: - application/json: - schema: - type: object - properties: {} - description: Gets an activation URL for a share. - /api/2.1/unity-catalog/recipients/{name}/share-permissions: - get: - operationId: recipient-share-permissions-sharepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/recipients/sharepermissions - x-stackQL-resource: recipient_share_permissions - x-stackQL-method: sharepermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The recipient share permissions list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - permissions_out: - type: array - items: - type: object - properties: - share_name: - type: string - privilege_assignments: - type: array - items: - type: object - properties: - principal: - type: string - privileges: - type: array - items: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.permissions_out - description: Gets the share permissions for the specified Recipient. The caller - must be a metastore admin or the owner of the Recipient. - /api/2.1/unity-catalog/shares: - post: - operationId: shares-create - externalDocs: - url: https://docs.databricks.com/api/workspace/shares/create - x-stackQL-resource: shares - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new share was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - storage_root: - type: string - objects: - type: array - items: - type: object - properties: - name: - type: string - comment: - type: string - shared_as: - type: string - string_shared_as: - type: string - content: - type: string - history_data_sharing_status: - type: string - partitions: - type: array - items: - type: object - properties: - values: - type: array - items: - type: object - properties: - name: - type: string - value: - type: string - recipient_property_key: - type: string - op: - type: string - cdf_enabled: - type: boolean - start_version: - type: integer - data_object_type: - type: string - added_at: - type: integer - added_by: - type: string - status: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - comment: - type: string - storage_root: - type: string - example: - name: string - comment: string - storage_root: string - description: Creates a new share for data objects. Data objects can be added - after creation with - get: - operationId: shares-list - externalDocs: - url: https://docs.databricks.com/api/workspace/shares/list - x-stackQL-resource: shares - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The shares list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - shares: - type: array - items: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - storage_root: - type: string - objects: - type: array - items: - type: object - properties: - name: - type: string - comment: - type: string - shared_as: - type: string - string_shared_as: - type: string - content: - type: string - history_data_sharing_status: - type: string - partitions: - type: array - items: - type: object - properties: - values: - type: array - items: - type: object - properties: - name: - type: string - value: - type: string - recipient_property_key: - type: string - op: - type: string - cdf_enabled: - type: boolean - start_version: - type: integer - data_object_type: - type: string - added_at: - type: integer - added_by: - type: string - status: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.shares - description: Gets an array of data object shares from the metastore. The caller - must be a metastore admin or the owner of the share. There is no guarantee - of a specific ordering of the elements in the array. - /api/2.1/unity-catalog/shares/{name}: - delete: - operationId: shares-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/shares/delete - x-stackQL-resource: shares - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The share was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a data object share from the metastore. The caller must - be an owner of the share. - get: - operationId: shares-get - externalDocs: - url: https://docs.databricks.com/api/workspace/shares/get - x-stackQL-resource: shares - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: include_shared_data - in: query - responses: - '200': - description: The share was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - storage_root: - type: string - objects: - type: array - items: - type: object - properties: - name: - type: string - comment: - type: string - shared_as: - type: string - string_shared_as: - type: string - content: - type: string - history_data_sharing_status: - type: string - partitions: - type: array - items: - type: object - properties: - values: - type: array - items: - type: object - properties: - name: - type: string - value: - type: string - recipient_property_key: - type: string - op: - type: string - cdf_enabled: - type: boolean - start_version: - type: integer - data_object_type: - type: string - added_at: - type: integer - added_by: - type: string - status: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - description: Gets a data object share from the metastore. The caller must be - a metastore admin or the owner of the share. - patch: - operationId: shares-update - externalDocs: - url: https://docs.databricks.com/api/workspace/shares/update - x-stackQL-resource: shares - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The share was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - storage_root: - type: string - objects: - type: array - items: - type: object - properties: - name: - type: string - comment: - type: string - shared_as: - type: string - string_shared_as: - type: string - content: - type: string - history_data_sharing_status: - type: string - partitions: - type: array - items: - type: object - properties: - values: - type: array - items: - type: object - properties: - name: - type: string - value: - type: string - recipient_property_key: - type: string - op: - type: string - cdf_enabled: - type: boolean - start_version: - type: integer - data_object_type: - type: string - added_at: - type: integer - added_by: - type: string - status: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - comment: - type: string - storage_root: - type: string - new_name: - type: string - updates: - type: Array of object - example: - owner: string - comment: string - storage_root: string - new_name: string - updates: - - action: ADD - data_object: - name: string - comment: string - shared_as: string - string_shared_as: string - content: string - history_data_sharing_status: ENABLED - partitions: - - values: - - name: string - value: string - recipient_property_key: string - op: EQUAL - cdf_enabled: true - start_version: 0 - data_object_type: TABLE - added_at: 0 - added_by: string - status: ACTIVE - description: Updates the share with the changes and data objects in the request. - The caller must be the owner of the share or a metastore admin. - /api/2.1/unity-catalog/shares/{name}/permissions: - get: - operationId: share-permissions-sharepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/shares/sharepermissions - x-stackQL-resource: share_permissions - x-stackQL-method: sharepermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The share permissions list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - privilege_assignments: - type: array - items: - type: object - properties: - principal: - type: string - privileges: - type: array - items: - type: string - x-stackQL-objectKey: $.privilege_assignments - description: Gets the permissions for a data share from the metastore. The caller - must be a metastore admin or the owner of the share. - patch: - operationId: share-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/shares/updatepermissions - x-stackQL-resource: share_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The share permissions were successfully updated. - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - changes: - type: Array of object - example: - changes: - - principal: string - add: - - SELECT - remove: - - SELECT - description: Updates the permissions for a data share in the metastore. The - caller must be a metastore admin or an owner of the share. -components: - x-stackQL-resources: - providers: - id: databricks_workspace.deltasharing.providers - name: providers - title: Providers - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1providers/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1providers/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.providers - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/providers/methods/get' - - $ref: '#/components/x-stackQL-resources/providers/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/providers/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/providers/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/providers/methods/delete' - exec: [] - provider_shares: - id: databricks_workspace.deltasharing.provider_shares - name: provider_shares - title: Provider_shares - methods: - listshares: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}~1shares/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.shares - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_shares/methods/listshares' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - recipients: - id: databricks_workspace.deltasharing.recipients - name: recipients - title: Recipients - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.recipients - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - rotatetoken: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}~1rotate-token/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/recipients/methods/get' - - $ref: '#/components/x-stackQL-resources/recipients/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/recipients/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/recipients/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/recipients/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/recipients/methods/rotatetoken' - recipient_activation_token: - id: databricks_workspace.deltasharing.recipient_activation_token - name: recipient_activation_token - title: Recipient_activation_token - methods: - retrievetoken: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1public~1data_sharing_activation~1{activation_url}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/recipient_activation_token/methods/retrievetoken' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - recipient_activation_url: - id: databricks_workspace.deltasharing.recipient_activation_url - name: recipient_activation_url - title: Recipient_activation_url - methods: - getactivationurlinfo: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1public~1data_sharing_activation_info~1{activation_url}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/recipient_activation_url/methods/getactivationurlinfo' - recipient_share_permissions: - id: databricks_workspace.deltasharing.recipient_share_permissions - name: recipient_share_permissions - title: Recipient_share_permissions - methods: - sharepermissions: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}~1share-permissions/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permissions_out - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/recipient_share_permissions/methods/sharepermissions' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - shares: - id: databricks_workspace.deltasharing.shares - name: shares - title: Shares - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1shares/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1shares/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.shares - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/shares/methods/get' - - $ref: '#/components/x-stackQL-resources/shares/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/shares/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/shares/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/shares/methods/delete' - exec: [] - share_permissions: - id: databricks_workspace.deltasharing.share_permissions - name: share_permissions - title: Share_permissions - methods: - sharepermissions: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}~1permissions/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.privilege_assignments - updatepermissions: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}~1permissions/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/share_permissions/methods/sharepermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/share_permissions/methods/updatepermissions' - replace: [] - delete: [] - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/filemanagement.yaml b/providers/src/databricks_workspace/v00.00.00000/services/filemanagement.yaml deleted file mode 100644 index e3fcf603..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/filemanagement.yaml +++ /dev/null @@ -1,865 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Filemanagement API - description: Filemanagement -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/dbfs/list: - get: - operationId: dbfs-directories-list - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/list - x-stackQL-resource: dbfs_directories - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: path - required: true - description: string - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - files: - type: array - items: - type: object - properties: - path: - type: string - is_dir: - type: boolean - file_size: - type: integer - modification_time: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.files - description: List the contents of a directory, or details of the file. If the - file or directory does not exist, this call throws an exception with - /api/2.0/dbfs/mkdirs: - post: - operationId: dbfs-directories-mkdirs - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/mkdirs - x-stackQL-resource: dbfs_directories - x-stackQL-method: mkdirs - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - path: - type: required - example: - path: /mnt/foo - description: Creates the given directory and necessary parent directories if - they do not exist. If a file (not a directory) exists at any prefix of the - input path, this call throws an exception with - /api/2.0/dbfs/delete: - post: - operationId: dbfs-objects-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/delete - x-stackQL-resource: dbfs_objects - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - path: - type: required - recursive: - type: string - example: - path: /mnt/foo - recursive: false - description: Delete the file or directory (optionally recursively delete all - files in the directory). This call throws an exception with - /api/2.0/dbfs/move: - post: - operationId: dbfs-files-move - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/move - x-stackQL-resource: dbfs_files - x-stackQL-method: move - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - source_path: - type: required - destination_path: - type: string - example: - source_path: /mnt/foo - destination_path: /mnt/bar - description: Moves a file from one location to another location within DBFS. - If the source file does not exist, this call throws an exception with - /api/2.0/dbfs/put: - post: - operationId: dbfs-files-put - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/put - x-stackQL-resource: dbfs_files - x-stackQL-method: put - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - path: - type: required - contents: - type: string - overwrite: - type: string - example: - path: /mnt/foo - contents: string - overwrite: false - description: Uploads a file through the use of multipart form post. It is mainly - used for streaming uploads, but can also be used as a convenient single call - for data upload. - /api/2.0/dbfs/read: - get: - operationId: dbfs-files-read - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/read - x-stackQL-resource: dbfs_files - x-stackQL-method: read - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: path - required: true - description: string - in: query - - name: offset - in: query - - name: length - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - bytes_read: - type: integer - data: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Returns the contents of a file. If the file does not exist, this - call throws an exception with - /api/2.0/dbfs/get-status: - get: - operationId: dbfs-objects-getstatus - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/getstatus - x-stackQL-resource: dbfs_objects - x-stackQL-method: getstatus - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: path - required: true - description: string - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - path: - type: string - is_dir: - type: boolean - file_size: - type: integer - modification_time: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the file information for a file or directory. If the file - or directory does not exist, this call throws an exception with - /api/2.0/dbfs/add-block: - post: - operationId: dbfs-streams-addblock - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/addblock - x-stackQL-resource: dbfs_streams - x-stackQL-method: addblock - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - handle: - type: required - data: - type: integer - format: int64 - example: - handle: 0 - data: string - description: Appends a block of data to the stream specified by the input handle. - If the handle does not exist, this call will throw an exception with - /api/2.0/dbfs/close: - post: - operationId: dbfs-streams-close - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/close - x-stackQL-resource: dbfs_streams - x-stackQL-method: close - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - handle: - type: required - example: - handle: 0 - description: Closes the stream specified by the input handle. If the handle - does not exist, this call throws an exception with - /api/2.0/dbfs/create: - post: - operationId: dbfs-streams-create - externalDocs: - url: https://docs.databricks.com/api/workspace/dbfs/create - x-stackQL-resource: dbfs_streams - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - handle: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - path: - type: required - overwrite: - type: string - example: - path: /mnt/foo - overwrite: false - description: Opens a stream to write to a file and returns a handle to this - stream. There is a 10 minute idle timeout on this handle. If a file or directory - already exists on the given path and - /api/2.0/fs/directories{directory_path}: - put: - operationId: directories-createdirectory - externalDocs: - url: https://docs.databricks.com/api/workspace/files/createdirectory - x-stackQL-resource: directories - x-stackQL-method: createdirectory - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: directory_path - required: true - description: string - in: path - responses: - '204': - description: The directory was successfully created or already exists. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Creates an empty directory. If necessary, also creates any parent - directories of the new, empty directory (like the shell command - delete: - operationId: directories-deletedirectory - externalDocs: - url: https://docs.databricks.com/api/workspace/files/deletedirectory - x-stackQL-resource: directories - x-stackQL-method: deletedirectory - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: directory_path - required: true - description: string - in: path - responses: - '204': - description: The directory was successfully deleted. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes an empty directory. - get: - operationId: directory-contents-listdirectorycontents - externalDocs: - url: https://docs.databricks.com/api/workspace/files/listdirectorycontents - x-stackQL-resource: directory_contents - x-stackQL-method: listdirectorycontents - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: directory_path - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: An array of DirectoryEntry for the contents of the directory. - If `next_page_token` is set, there may be more entries in the directory. - content: - application/json: - schema: - type: object - properties: - contents: - type: array - items: - type: object - properties: - path: - type: string - is_directory: - type: boolean - file_size: - type: integer - last_modified: - type: integer - name: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - x-stackQL-objectKey: $.contents - description: Returns the contents of a directory. If there is no directory at - the specified path, the API returns a HTTP 404 error. - /api/2.0/fs/files{file_path}: - delete: - operationId: files-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/files/delete - x-stackQL-resource: files - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: file_path - required: true - description: string - in: path - responses: - '204': - description: The file was successfully deleted. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a file. If the request is successful, there is no response - body. - get: - operationId: files-download - externalDocs: - url: https://docs.databricks.com/api/workspace/files/download - x-stackQL-resource: files - x-stackQL-method: download - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: file_path - required: true - description: string - in: path - responses: - '200': - description: '' - description: Downloads a file. The file contents are the response body. This - is a standard HTTP file download, not a JSON RPC. It supports the Range and - If-Unmodified-Since HTTP headers. - put: - operationId: files-upload - externalDocs: - url: https://docs.databricks.com/api/workspace/files/upload - x-stackQL-resource: files - x-stackQL-method: upload - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: file_path - required: true - description: string - in: path - - name: overwrite - in: query - responses: - '204': - description: The file was successfully uploaded. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - description: Uploads a file of up to 5 GiB. The file contents should be sent - as the request body as raw bytes (an octet stream); do not encode or otherwise - modify the bytes before sending. The contents of the resulting file will be - exactly the bytes sent in the request body. If the request is successful, - there is no response body. -components: - x-stackQL-resources: - dbfs_directories: - id: databricks_workspace.filemanagement.dbfs_directories - name: dbfs_directories - title: Dbfs_directories - methods: - list: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.files - mkdirs: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1mkdirs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dbfs_directories/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/dbfs_directories/methods/mkdirs' - update: [] - replace: [] - delete: [] - exec: [] - dbfs_objects: - id: databricks_workspace.filemanagement.dbfs_objects - name: dbfs_objects - title: Dbfs_objects - methods: - delete: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getstatus: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1get-status/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dbfs_objects/methods/getstatus' - insert: [] - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dbfs_objects/methods/delete' - exec: [] - dbfs_files: - id: databricks_workspace.filemanagement.dbfs_files - name: dbfs_files - title: Dbfs_files - methods: - move: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1move/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - put: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1put/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - read: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1read/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dbfs_files/methods/read' - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/dbfs_files/methods/put' - - $ref: '#/components/x-stackQL-resources/dbfs_files/methods/move' - dbfs_streams: - id: databricks_workspace.filemanagement.dbfs_streams - name: dbfs_streams - title: Dbfs_streams - methods: - addblock: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1add-block/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - close: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1close/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - create: - operation: - $ref: '#/paths/~1api~12.0~1dbfs~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: - - $ref: '#/components/x-stackQL-resources/dbfs_streams/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dbfs_streams/methods/close' - exec: - - $ref: '#/components/x-stackQL-resources/dbfs_streams/methods/addblock' - directories: - id: databricks_workspace.filemanagement.directories - name: directories - title: Directories - methods: - createdirectory: - operation: - $ref: '#/paths/~1api~12.0~1fs~1directories{directory_path}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletedirectory: - operation: - $ref: '#/paths/~1api~12.0~1fs~1directories{directory_path}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: - - $ref: '#/components/x-stackQL-resources/directories/methods/createdirectory' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/directories/methods/deletedirectory' - exec: [] - directory_contents: - id: databricks_workspace.filemanagement.directory_contents - name: directory_contents - title: Directory_contents - methods: - listdirectorycontents: - operation: - $ref: '#/paths/~1api~12.0~1fs~1directories{directory_path}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.contents - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/directory_contents/methods/listdirectorycontents' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - files: - id: databricks_workspace.filemanagement.files - name: files - title: Files - methods: - delete: - operation: - $ref: '#/paths/~1api~12.0~1fs~1files{file_path}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - download: - operation: - $ref: '#/paths/~1api~12.0~1fs~1files{file_path}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - upload: - operation: - $ref: '#/paths/~1api~12.0~1fs~1files{file_path}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/files/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/files/methods/upload' - - $ref: '#/components/x-stackQL-resources/files/methods/download' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/files.yaml b/providers/src/databricks_workspace/v00.00.00000/services/files.yaml new file mode 100644 index 00000000..b4969473 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/files.yaml @@ -0,0 +1,1305 @@ +openapi: 3.0.0 +info: + title: Databricks Files API (workspace) + description: >- + OpenAPI specification for the Databricks files service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.files +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/dbfs/add-block: + post: + operationId: dbfs_add_block + summary: >- + Appends a block of data to the stream specified by the input handle. If + the handle does not exist, + tags: + - files + - dbfs + description: >- + Appends a block of data to the stream specified by the input handle. If + the handle does not exist, + + this call will throw an exception with ``RESOURCE_DOES_NOT_EXIST``. + + + If the block of data exceeds 1 MB, this call will throw an exception + with ``MAX_BLOCK_SIZE_EXCEEDED``. + + + :param handle: int + The handle on an open stream. + :param data: str + The base64-encoded data to append to the stream. This has a limit of 1 MB. + requestBody: + content: + application/json: + schema: + type: object + properties: + handle: + type: integer + description: The handle on an open stream. + data: + type: string + description: >- + The base64-encoded data to append to the stream. This has a + limit of 1 MB. + required: + - handle + - data + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/close: + post: + operationId: dbfs_close + summary: >- + Closes the stream specified by the input handle. If the handle does not + exist, this call throws an + tags: + - files + - dbfs + description: >- + Closes the stream specified by the input handle. If the handle does not + exist, this call throws an + + exception with ``RESOURCE_DOES_NOT_EXIST``. + + + :param handle: int + The handle on an open stream. + requestBody: + content: + application/json: + schema: + type: object + properties: + handle: + type: integer + description: The handle on an open stream. + required: + - handle + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/create: + post: + operationId: dbfs_create + summary: >- + Opens a stream to write to a file and returns a handle to this stream. + There is a 10 minute idle + tags: + - files + - dbfs + description: >- + Opens a stream to write to a file and returns a handle to this stream. + There is a 10 minute idle + + timeout on this handle. If a file or directory already exists on the + given path and __overwrite__ is + + set to false, this call will throw an exception with + ``RESOURCE_ALREADY_EXISTS``. + + + A typical workflow for file upload would be: + + + 1. Issue a ``create`` call and get a handle. 2. Issue one or more + ``add-block`` calls with the handle + + you have. 3. Issue a ``close`` call with the handle you have. + + + :param path: str + The path of the new file. The path should be the absolute DBFS path. + :param overwrite: bool (optional) + The flag that specifies whether to overwrite existing file/files. + + :returns: :class:`CreateResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + path: + type: string + description: >- + The path of the new file. The path should be the absolute + DBFS path. + overwrite: + type: string + description: >- + The flag that specifies whether to overwrite existing + file/files. + required: + - path + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/delete: + post: + operationId: dbfs_delete + summary: >- + Delete the file or directory (optionally recursively delete all files in + the directory). This call + tags: + - files + - dbfs + description: >- + Delete the file or directory (optionally recursively delete all files in + the directory). This call + + throws an exception with `IO_ERROR` if the path is a non-empty directory + and `recursive` is set to + + `false` or on other similar errors. + + + When you delete a large number of files, the delete operation is done in + increments. The call returns + + a response after approximately 45 seconds with an error message (503 + Service Unavailable) asking you + + to re-invoke the delete operation until the directory structure is fully + deleted. + + + For operations that delete more than 10K files, we discourage using the + DBFS REST API, but advise you + + to perform such operations in the context of a cluster, using the [File + system utility + + (dbutils.fs)](/dev-tools/databricks-utils.html#dbutils-fs). `dbutils.fs` + covers the functional scope + + of the DBFS REST API, but from notebooks. Running such operations using + notebooks provides better + + control and manageability, such as selective deletes, and the + possibility to automate periodic delete + + jobs. + + + :param path: str + The path of the file or directory to delete. The path should be the absolute DBFS path. + :param recursive: bool (optional) + Whether or not to recursively delete the directory's contents. Deleting empty directories can be + done without providing the recursive flag. + requestBody: + content: + application/json: + schema: + type: object + properties: + path: + type: string + description: >- + The path of the file or directory to delete. The path should + be the absolute DBFS path. + recursive: + type: string + description: >- + Whether or not to recursively delete the directory's + contents. Deleting empty directories can be done without + providing the recursive flag. + required: + - path + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/get-status: + get: + operationId: dbfs_get_status + summary: >- + Gets the file information for a file or directory. If the file or + directory does not exist, this call + tags: + - files + - dbfs + description: >- + Gets the file information for a file or directory. If the file or + directory does not exist, this call + + throws an exception with `RESOURCE_DOES_NOT_EXIST`. + + + :param path: str + The path of the file or directory. The path should be the absolute DBFS path. + + :returns: :class:`FileInfo` + parameters: + - name: path + in: query + required: true + schema: + type: string + description: >- + The path of the file or directory. The path should be the absolute + DBFS path. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FileInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/list: + get: + operationId: dbfs_list + summary: >- + List the contents of a directory, or details of the file. If the file or + directory does not exist, + tags: + - files + - dbfs + description: >- + List the contents of a directory, or details of the file. If the file or + directory does not exist, + + this call throws an exception with `RESOURCE_DOES_NOT_EXIST`. + + + When calling list on a large directory, the list operation will time out + after approximately 60 + + seconds. We strongly recommend using list only on directories containing + less than 10K files and + + discourage using the DBFS REST API for operations that list more than + 10K files. Instead, we recommend + + that you perform such operations in the context of a cluster, using the + [File system utility + + (dbutils.fs)](/dev-tools/databricks-utils.html#dbutils-fs), which + provides the same functionality + + without timing out. + + + :param path: str + The path of the file or directory. The path should be the absolute DBFS path. + + :returns: Iterator over :class:`FileInfo` + parameters: + - name: path + in: query + required: true + schema: + type: string + description: >- + The path of the file or directory. The path should be the absolute + DBFS path. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListStatusResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/mkdirs: + post: + operationId: dbfs_mkdirs + summary: >- + Creates the given directory and necessary parent directories if they do + not exist. If a file (not a + tags: + - files + - dbfs + description: >- + Creates the given directory and necessary parent directories if they do + not exist. If a file (not a + + directory) exists at any prefix of the input path, this call throws an + exception with + + `RESOURCE_ALREADY_EXISTS`. **Note**: If this operation fails, it might + have succeeded in creating some + + of the necessary parent directories. + + + :param path: str + The path of the new directory. The path should be the absolute DBFS path. + requestBody: + content: + application/json: + schema: + type: object + properties: + path: + type: string + description: >- + The path of the new directory. The path should be the + absolute DBFS path. + required: + - path + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/move: + post: + operationId: dbfs_move + summary: >- + Moves a file from one location to another location within DBFS. If the + source file does not exist, + tags: + - files + - dbfs + description: >- + Moves a file from one location to another location within DBFS. If the + source file does not exist, + + this call throws an exception with `RESOURCE_DOES_NOT_EXIST`. If a file + already exists in the + + destination path, this call throws an exception with + `RESOURCE_ALREADY_EXISTS`. If the given source + + path is a directory, this call always recursively moves all files. + + + :param source_path: str + The source path of the file or directory. The path should be the absolute DBFS path. + :param destination_path: str + The destination path of the file or directory. The path should be the absolute DBFS path. + requestBody: + content: + application/json: + schema: + type: object + properties: + source_path: + type: string + description: >- + The source path of the file or directory. The path should be + the absolute DBFS path. + destination_path: + type: string + description: >- + The destination path of the file or directory. The path + should be the absolute DBFS path. + required: + - source_path + - destination_path + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/put: + post: + operationId: dbfs_put + summary: >- + Uploads a file through the use of multipart form post. It is mainly used + for streaming uploads, but + tags: + - files + - dbfs + description: >- + Uploads a file through the use of multipart form post. It is mainly used + for streaming uploads, but + + can also be used as a convenient single call for data upload. + + + Alternatively you can pass contents as base64 string. + + + The amount of data that can be passed (when not streaming) using the + __contents__ parameter is limited + + to 1 MB. `MAX_BLOCK_SIZE_EXCEEDED` will be thrown if this limit is + exceeded. + + + If you want to upload large files, use the streaming upload. For + details, see :method:dbfs/create, + + :method:dbfs/addBlock, :method:dbfs/close. + + + :param path: str + The path of the new file. The path should be the absolute DBFS path. + :param contents: str (optional) + This parameter might be absent, and instead a posted file will be used. + :param overwrite: bool (optional) + The flag that specifies whether to overwrite existing file/files. + requestBody: + content: + application/json: + schema: + type: object + properties: + path: + type: string + description: >- + The path of the new file. The path should be the absolute + DBFS path. + contents: + type: string + description: >- + This parameter might be absent, and instead a posted file + will be used. + overwrite: + type: string + description: >- + The flag that specifies whether to overwrite existing + file/files. + required: + - path + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/dbfs/read: + get: + operationId: dbfs_read + summary: >- + Returns the contents of a file. If the file does not exist, this call + throws an exception with + tags: + - files + - dbfs + description: >- + Returns the contents of a file. If the file does not exist, this call + throws an exception with + + `RESOURCE_DOES_NOT_EXIST`. If the path is a directory, the read length + is negative, or if the offset + + is negative, this call throws an exception with + `INVALID_PARAMETER_VALUE`. If the read length exceeds + + 1 MB, this call throws an exception with `MAX_READ_SIZE_EXCEEDED`. + + + If `offset + length` exceeds the number of bytes in a file, it reads the + contents until the end of + + file. + + + :param path: str + The path of the file to read. The path should be the absolute DBFS path. + :param length: int (optional) + The number of bytes to read starting from the offset. This has a limit of 1 MB, and a default value + of 0.5 MB. + :param offset: int (optional) + The offset to read from in bytes. + + :returns: :class:`ReadResponse` + parameters: + - name: path + in: query + required: true + schema: + type: string + description: >- + The path of the file to read. The path should be the absolute DBFS + path. + - name: length + in: query + required: false + schema: + type: string + description: >- + The number of bytes to read starting from the offset. This has a + limit of 1 MB, and a default value of 0.5 MB. + - name: offset + in: query + required: false + schema: + type: string + description: The offset to read from in bytes. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ReadResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbfsAPI + /api/2.0/fs/directories{directory_path}: + put: + operationId: files_create_directory + summary: >- + Creates an empty directory. If necessary, also creates any parent + directories of the new, empty + tags: + - files + description: >- + Creates an empty directory. If necessary, also creates any parent + directories of the new, empty + + directory (like the shell command `mkdir -p`). If called on an existing + directory, returns a success + + response; this method is idempotent (it will succeed if the directory + already exists). + + + :param directory_path: str + The absolute path of a directory. + parameters: + - name: directory_path + in: path + required: true + schema: + type: string + description: The absolute path of a directory. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI + delete: + operationId: files_delete_directory + summary: Deletes an empty directory. + tags: + - files + description: >- + Deletes an empty directory. + + + To delete a non-empty directory, first delete all of its contents. This + can be done by listing the + + directory contents and deleting each file and subdirectory recursively. + + + :param directory_path: str + The absolute path of a directory. + parameters: + - name: directory_path + in: path + required: true + schema: + type: string + description: The absolute path of a directory. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI + head: + operationId: files_get_directory_metadata + summary: >- + Get the metadata of a directory. The response HTTP headers contain the + metadata. There is no response + tags: + - files + description: >- + Get the metadata of a directory. The response HTTP headers contain the + metadata. There is no response + + body. + + + This method is useful to check if a directory exists and the caller has + access to it. + + + If you wish to ensure the directory exists, you can instead use `PUT`, + which will create the directory + + if it does not exist, and is idempotent (it will succeed if the + directory already exists). + + + :param directory_path: str + The absolute path of a directory. + parameters: + - name: directory_path + in: path + required: true + schema: + type: string + description: The absolute path of a directory. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI + get: + operationId: files_list_directory_contents + summary: >- + Returns the contents of a directory. If there is no directory at the + specified path, the API returns a + tags: + - files + description: >- + Returns the contents of a directory. If there is no directory at the + specified path, the API returns a + + HTTP 404 error. + + + :param directory_path: str + The absolute path of a directory. + :param page_size: int (optional) + The maximum number of directory entries to return. The response may contain fewer entries. If the + response contains a `next_page_token`, there may be more entries, even if fewer than `page_size` + entries are in the response. + + We recommend not to set this value unless you are intentionally listing less than the complete + directory contents. + + If unspecified, at most 1000 directory entries will be returned. The maximum value is 1000. Values + above 1000 will be coerced to 1000. + :param page_token: str (optional) + An opaque page token which was the `next_page_token` in the response of the previous request to list + the contents of this directory. Provide this token to retrieve the next page of directory entries. + When providing a `page_token`, all other parameters provided to the request must match the previous + request. To list all of the entries in a directory, it is necessary to continue requesting pages of + entries until the response contains no `next_page_token`. Note that the number of entries returned + must not be used to determine when the listing is complete. + + :returns: Iterator over :class:`DirectoryEntry` + parameters: + - name: directory_path + in: path + required: true + schema: + type: string + description: The absolute path of a directory. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of directory entries to return. The response may + contain fewer entries. If the response contains a `next_page_token`, + there may be more entries, even if fewer than `page_size` entries + are in the response. We recommend not to set this value unless you + are intentionally listing less than the complete directory contents. + If unspecified, at most 1000 directory entries will be returned. The + maximum value is 1000. Values above 1000 will be coerced to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + An opaque page token which was the `next_page_token` in the response + of the previous request to list the contents of this directory. + Provide this token to retrieve the next page of directory entries. + When providing a `page_token`, all other parameters provided to the + request must match the previous request. To list all of the entries + in a directory, it is necessary to continue requesting pages of + entries until the response contains no `next_page_token`. Note that + the number of entries returned must not be used to determine when + the listing is complete. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListDirectoryResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI + /api/2.0/fs/files{file_path}: + delete: + operationId: files_delete + summary: Deletes a file. If the request is successful, there is no response body. + tags: + - files + description: |- + Deletes a file. If the request is successful, there is no response body. + + :param file_path: str + The absolute path of the file. + parameters: + - name: file_path + in: path + required: true + schema: + type: string + description: The absolute path of the file. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI + get: + operationId: files_download + summary: >- + Downloads a file. The file contents are the response body. This is a + standard HTTP file download, not + tags: + - files + description: >- + Downloads a file. The file contents are the response body. This is a + standard HTTP file download, not + + a JSON RPC. It supports the Range and If-Unmodified-Since HTTP headers. + + + :param file_path: str + The absolute path of the file. + + :returns: :class:`DownloadResponse` + parameters: + - name: file_path + in: path + required: true + schema: + type: string + description: The absolute path of the file. + responses: + '200': + description: Success + content: + application/octet-stream: + schema: + type: string + format: binary + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI + head: + operationId: files_get_metadata + summary: >- + Get the metadata of a file. The response HTTP headers contain the + metadata. There is no response body. + tags: + - files + description: >- + Get the metadata of a file. The response HTTP headers contain the + metadata. There is no response body. + + + :param file_path: str + The absolute path of the file. + + :returns: :class:`GetMetadataResponse` + parameters: + - name: file_path + in: path + required: true + schema: + type: string + description: The absolute path of the file. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetMetadataResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI + put: + operationId: files_upload + summary: >- + Uploads a file of up to 5 GiB. The file contents should be sent as the + request body as raw bytes (an + tags: + - files + description: >- + Uploads a file of up to 5 GiB. The file contents should be sent as the + request body as raw bytes (an + + octet stream); do not encode or otherwise modify the bytes before + sending. The contents of the + + resulting file will be exactly the bytes sent in the request body. If + the request is successful, there + + is no response body. + + + :param file_path: str + The absolute path of the file. + :param contents: BinaryIO + + :param overwrite: bool (optional) + If true or unspecified, an existing file will be overwritten. If false, an error will be returned if + the path points to an existing file. + parameters: + - name: file_path + in: path + required: true + schema: + type: string + description: The absolute path of the file. + - name: overwrite + in: query + required: false + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + contents: + type: string + description: >- + :param overwrite: bool (optional) If true or unspecified, an + existing file will be overwritten. If false, an error will + be returned if the path points to an existing file. + required: + - contents + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FilesAPI +components: + schemas: + AddBlockResponse: + type: object + properties: {} + CloseResponse: + type: object + properties: {} + CreateResponse: + type: object + properties: + handle: + type: integer + DeleteResponse: + type: object + properties: {} + DirectoryEntry: + type: object + properties: + file_size: + type: integer + is_directory: + type: boolean + description: True if the path is a directory. + last_modified: + type: integer + description: >- + Last modification time of given file in milliseconds since unix + epoch. + name: + type: string + description: >- + The name of the file or directory. This is the last component of the + path. + path: + type: string + description: The absolute path of the file or directory. + DownloadResponse: + type: object + properties: + content_length: + type: integer + content-type: + type: string + contents: + type: string + format: binary + last-modified: + type: string + description: The last modified time of the file in HTTP-date (RFC 7231) format. + FileInfo: + type: object + properties: + file_size: + type: integer + is_dir: + type: boolean + description: True if the path is a directory. + modification_time: + type: integer + description: Last modification time of given file in milliseconds since epoch. + path: + type: string + description: The absolute path of the file or directory. + GetMetadataResponse: + type: object + properties: + content_length: + type: integer + content-type: + type: string + last-modified: + type: string + description: The last modified time of the file in HTTP-date (RFC 7231) format. + ListDirectoryResponse: + type: object + properties: + contents: + type: array + items: + $ref: '#/components/schemas/DirectoryEntry' + next_page_token: + type: string + description: >- + A token, which can be sent as `page_token` to retrieve the next + page. + ListStatusResponse: + type: object + properties: + files: + type: array + items: + $ref: '#/components/schemas/FileInfo' + MkDirsResponse: + type: object + properties: {} + MoveResponse: + type: object + properties: {} + PutResponse: + type: object + properties: {} + ReadResponse: + type: object + properties: + bytes_read: + type: integer + data: + type: string + description: The base64-encoded contents of the file read. + x-stackQL-resources: + dbfs: + id: databricks_workspace.files.dbfs + name: dbfs + title: Dbfs + methods: + add_block: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1add-block/post' + response: + mediaType: application/json + openAPIDocKey: '200' + close: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1close/post' + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get_status: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1get-status/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.files + mkdirs: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1mkdirs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + move: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1move/post' + response: + mediaType: application/json + openAPIDocKey: '200' + put: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1put/post' + response: + mediaType: application/json + openAPIDocKey: '200' + read: + operation: + $ref: '#/paths/~1api~12.0~1dbfs~1read/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/dbfs/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/dbfs/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/dbfs/methods/delete' + replace: [] + files: + id: databricks_workspace.files.files + name: files + title: Files + methods: + create_directory: + operation: + $ref: '#/paths/~1api~12.0~1fs~1directories{directory_path}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + delete_directory: + operation: + $ref: '#/paths/~1api~12.0~1fs~1directories{directory_path}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + list_directory_contents: + operation: + $ref: '#/paths/~1api~12.0~1fs~1directories{directory_path}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.contents + delete: + operation: + $ref: '#/paths/~1api~12.0~1fs~1files{file_path}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + download: + operation: + $ref: '#/paths/~1api~12.0~1fs~1files{file_path}/get' + response: + mediaType: application/octet-stream + openAPIDocKey: '200' + upload: + operation: + $ref: '#/paths/~1api~12.0~1fs~1files{file_path}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/files/methods/list_directory_contents + - $ref: '#/components/x-stackQL-resources/files/methods/download' + insert: [] + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/files/methods/delete_directory' + - $ref: '#/components/x-stackQL-resources/files/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/files/methods/create_directory' + - $ref: '#/components/x-stackQL-resources/files/methods/upload' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/iam.yaml b/providers/src/databricks_workspace/v00.00.00000/services/iam.yaml index b52ac9ed..baa6a328 100644 --- a/providers/src/databricks_workspace/v00.00.00000/services/iam.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/services/iam.yaml @@ -1,2891 +1,3629 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Iam API - description: Iam -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/preview/accounts/access-control/assignable-roles: - get: - operationId: assignable-roles-getassignablerolesforresource - externalDocs: - url: https://docs.databricks.com/api/workspace/accountaccesscontrolproxy/getassignablerolesforresource - x-stackQL-resource: assignable_roles - x-stackQL-method: getassignablerolesforresource - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: resource - required: true - description: string - in: query - responses: - '200': - description: Assignable roles were returned successfully. - content: - application/json: - schema: - type: object - properties: - roles: - type: array - items: - type: object - properties: - name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.roles - description: Gets all the roles that can be granted on an account-level resource. - A role is grantable if the rule set on the resource can contain an access - rule of the role. - /api/2.0/preview/scim/v2/Me: - get: - operationId: currentuser-me - externalDocs: - url: https://docs.databricks.com/api/workspace/currentuser/me - x-stackQL-resource: currentuser - x-stackQL-method: me - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: User information was returned successfully. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - userName: - type: string - emails: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - name: - type: object - properties: - givenName: - type: string - familyName: - type: string - displayName: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get details about the current method caller's identity. - /api/2.0/preview/scim/v2/Groups: - post: - operationId: groups-create - externalDocs: - url: https://docs.databricks.com/api/workspace/groups/create - x-stackQL-resource: groups - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '201': - description: The group creation request succeeded. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - displayName: - type: string - members: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - id: - type: integer - format: int64 - displayName: - type: string - externalId: - type: string - members: - type: Array of object - groups: - type: Array of object - roles: - type: Array of object - entitlements: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:schemas:core:2.0:Group - id: string - displayName: string - members: - - $ref: string - value: string - display: string - primary: true - type: string - groups: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - entitlements: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - description: Properties of the new group. - description: Creates a group in the Databricks workspace with a unique name, - using the supplied group details. - get: - operationId: groups-list - externalDocs: - url: https://docs.databricks.com/api/workspace/groups/list - x-stackQL-resource: groups - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: List groups operation was succesful. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - totalResults: - type: integer - startIndex: - type: integer - itemsPerPage: - type: integer - Resources: - type: array - items: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - displayName: - type: string - members: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.Resources - description: Gets all details of the groups associated with the Databricks workspace. - /api/2.0/preview/scim/v2/Groups/{id}: - delete: - operationId: groups-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/groups/delete - x-stackQL-resource: groups - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '204': - description: Group was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a group from the Databricks workspace. - get: - operationId: groups-get - externalDocs: - url: https://docs.databricks.com/api/workspace/groups/get - x-stackQL-resource: groups - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Group information was returned successfully. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - displayName: - type: string - members: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the information for a specific group in the Databricks workspace. - patch: - operationId: groups-patch - externalDocs: - url: https://docs.databricks.com/api/workspace/groups/patch - x-stackQL-resource: groups - x-stackQL-method: patch - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Group information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - Operations: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:api:messages:2.0:PatchOp - Operations: - - op: add - path: string - value: null - description: Operations to be applied on group information. - description: Partially updates the details of a group. - put: - operationId: groups-update - externalDocs: - url: https://docs.databricks.com/api/workspace/groups/update - x-stackQL-resource: groups - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Group information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - id: - type: integer - format: int64 - displayName: - type: string - externalId: - type: string - members: - type: Array of object - groups: - type: Array of object - roles: - type: Array of object - entitlements: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:schemas:core:2.0:Group - id: string - displayName: string - members: - - $ref: string - value: string - display: string - primary: true - type: string - groups: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - entitlements: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - description: Operations to be applied on group information. - description: Updates the details of a group by replacing the entire group entity. - /api/2.0/permissions/{request_object_type}/{request_object_id}: - get: - operationId: permissions-get - externalDocs: - url: https://docs.databricks.com/api/workspace/permissions/get - x-stackQL-resource: permissions - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: request_object_type - required: true - description: string - in: path - - name: request_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of an object. Objects can inherit permissions - from their parent objects or root object. - put: - operationId: permissions-set - externalDocs: - url: https://docs.databricks.com/api/workspace/permissions/set - x-stackQL-resource: permissions - x-stackQL-method: set - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: request_object_type - required: true - description: string - in: path - - name: request_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their parent objects or root object. - patch: - operationId: permissions-update - externalDocs: - url: https://docs.databricks.com/api/workspace/permissions/update - x-stackQL-resource: permissions - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: request_object_type - required: true - description: string - in: path - - name: request_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on an object. Objects can inherit permissions - from their parent objects or root object. - /api/2.0/permissions/{request_object_type}/{request_object_id}/permissionLevels: - get: - operationId: permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/permissions/getpermissionlevels - x-stackQL-resource: permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: request_object_type - required: true - description: string - in: path - - name: request_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/preview/accounts/access-control/rule-sets: - get: - operationId: rule-sets-getruleset - externalDocs: - url: https://docs.databricks.com/api/workspace/accountaccesscontrolproxy/getruleset - x-stackQL-resource: rule_sets - x-stackQL-method: getruleset - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: query - - name: etag - required: true - description: string - in: query - responses: - '200': - description: The rule set was returned successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - grant_rules: - type: array - items: - type: object - properties: - principals: - type: array - items: - type: string - role: - type: string - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - description: Get a rule set by its name. A rule set is always attached to a - resource and contains a list of access rules on the said resource. Currently - only a default rule set for each resource is supported. - put: - operationId: rule-sets-updateruleset - externalDocs: - url: https://docs.databricks.com/api/workspace/accountaccesscontrolproxy/updateruleset - x-stackQL-resource: rule_sets - x-stackQL-method: updateruleset - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Rule sets on the resource was updated successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - grant_rules: - type: array - items: - type: object - properties: - role: - type: string - principals: - type: array - items: - type: string - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - rule_set: - type: object - example: - name: string - rule_set: - name: string - grant_rules: - - principals: - - string - role: string - etag: string - description: The rule set to update to. - description: Replace the rules of a rule set. First, use a GET rule set request - to read the current version of the rule set before modifying it. This pattern - helps prevent conflicts between concurrent updates. - /api/2.0/preview/scim/v2/ServicePrincipals: - post: - operationId: service-principals-create - externalDocs: - url: https://docs.databricks.com/api/workspace/serviceprincipals/create - x-stackQL-resource: service_principals - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '201': - description: The user creation request succeeded. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - applicationId: - type: string - displayName: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - id: - type: integer - format: int64 - applicationId: - type: uuid - displayName: - type: string - externalId: - type: string - active: - type: boolean - groups: - type: Array of object - roles: - type: Array of object - entitlements: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:schemas:core:2.0:ServicePrincipal - id: string - applicationId: 97ab27fa-30e2-43e3-92a3-160e80f4c0d5 - displayName: etl-service - groups: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - entitlements: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Properties of the new service principal. - description: Creates a new service principal in the Databricks workspace. - get: - operationId: service-principals-list - externalDocs: - url: https://docs.databricks.com/api/workspace/serviceprincipals/list - x-stackQL-resource: service_principals - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: List service principals operation was succesful. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - totalResults: - type: integer - startIndex: - type: integer - itemsPerPage: - type: integer - Resources: - type: array - items: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - applicationId: - type: string - displayName: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.Resources - description: Gets the set of service principals associated with a Databricks - workspace. - /api/2.0/preview/scim/v2/ServicePrincipals/{id}: - delete: - operationId: service-principals-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/serviceprincipals/delete - x-stackQL-resource: service_principals - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '204': - description: Service principal was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Delete a single service principal in the Databricks workspace. - get: - operationId: service-principals-get - externalDocs: - url: https://docs.databricks.com/api/workspace/serviceprincipals/get - x-stackQL-resource: service_principals - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Service principal information was returned successfully. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - applicationId: - type: string - displayName: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the details for a single service principal define in the Databricks - workspace. - patch: - operationId: service-principals-patch - externalDocs: - url: https://docs.databricks.com/api/workspace/serviceprincipals/patch - x-stackQL-resource: service_principals - x-stackQL-method: patch - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Service principal information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - Operations: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:api:messages:2.0:PatchOp - Operations: - - op: add - path: string - value: null - description: Operations to be applied on service principal information. - description: Partially updates the details of a single service principal in - the Databricks workspace. - put: - operationId: service-principals-update - externalDocs: - url: https://docs.databricks.com/api/workspace/serviceprincipals/update - x-stackQL-resource: service_principals - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: Service principal information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - id: - type: integer - format: int64 - applicationId: - type: uuid - displayName: - type: string - externalId: - type: string - active: - type: boolean - groups: - type: Array of object - roles: - type: Array of object - entitlements: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:schemas:core:2.0:ServicePrincipal - id: string - applicationId: 97ab27fa-30e2-43e3-92a3-160e80f4c0d5 - displayName: etl-service - groups: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - entitlements: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Updates the details of a single service principal. - /api/2.0/preview/scim/v2/Users: - post: - operationId: users-create - externalDocs: - url: https://docs.databricks.com/api/workspace/users/create - x-stackQL-resource: users - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '201': - description: The user creation request succeeded. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - userName: - type: string - emails: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - name: - type: object - properties: - givenName: - type: string - familyName: - type: string - displayName: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - id: - type: integer - format: int64 - userName: - type: email - displayName: - type: string - externalId: - type: string - active: - type: boolean - emails: - type: Array of object - name: - type: object - groups: - type: Array of object - roles: - type: Array of object - entitlements: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:schemas:core:2.0:User - id: string - userName: user@example.com - emails: - - $ref: string - value: string - display: string - primary: true - type: string - name: - givenName: string - familyName: string - displayName: string - groups: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - entitlements: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Properties of the new user. - description: Creates a new user in the Databricks workspace. This new user will - also be added to the Databricks account. - get: - operationId: users-list - externalDocs: - url: https://docs.databricks.com/api/workspace/users/list - x-stackQL-resource: users - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: List users operation was succesful. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - totalResults: - type: integer - startIndex: - type: integer - itemsPerPage: - type: integer - Resources: - type: array - items: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - userName: - type: string - emails: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - name: - type: object - properties: - givenName: - type: string - familyName: - type: string - displayName: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.Resources - description: Gets details for all the users associated with a Databricks workspace. - /api/2.0/preview/scim/v2/Users/{id}: - delete: - operationId: users-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/users/delete - x-stackQL-resource: users - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '204': - description: User was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a user. Deleting a user from a Databricks workspace also - removes objects associated with the user. - get: - operationId: users-get - externalDocs: - url: https://docs.databricks.com/api/workspace/users/get - x-stackQL-resource: users - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - - name: filter - in: query - - name: attributes - in: query - - name: excludedAttributes - in: query - - name: startIndex - in: query - - name: count - in: query - - name: sortBy - in: query - - name: sortOrder - in: query - responses: - '200': - description: User information was returned successfully. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: string - id: - type: string - userName: - type: string - emails: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - name: - type: object - properties: - givenName: - type: string - familyName: - type: string - displayName: - type: string - groups: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - roles: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - entitlements: - type: array - items: - type: object - properties: - $ref: - type: string - value: - type: string - display: - type: string - primary: - type: boolean - type: - type: string - externalId: - type: string - active: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets information for a specific user in Databricks workspace. - patch: - operationId: users-patch - externalDocs: - url: https://docs.databricks.com/api/workspace/users/patch - x-stackQL-resource: users - x-stackQL-method: patch - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: User information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - Operations: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:api:messages:2.0:PatchOp - Operations: - - op: add - path: string - value: null - description: Operations to be applied on user information. - description: Partially updates a user resource by applying the supplied operations - on specific user attributes. - put: - operationId: users-update - externalDocs: - url: https://docs.databricks.com/api/workspace/users/update - x-stackQL-resource: users - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: int64 - in: path - responses: - '200': - description: User information was updated successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - schemas: - type: Array of string - id: - type: integer - format: int64 - userName: - type: email - displayName: - type: string - externalId: - type: string - active: - type: boolean - emails: - type: Array of object - name: - type: object - groups: - type: Array of object - roles: - type: Array of object - entitlements: - type: Array of object - example: - schemas: - - urn:ietf:params:scim:schemas:core:2.0:User - id: string - userName: user@example.com - emails: - - $ref: string - value: string - display: string - primary: true - type: string - name: - givenName: string - familyName: string - displayName: string - groups: - - $ref: string - value: string - display: string - primary: true - type: string - roles: - - $ref: string - value: string - display: string - primary: true - type: string - entitlements: - - $ref: string - value: string - display: string - primary: true - type: string - externalId: string - active: true - description: Replaces a user's information with the data supplied in request. - /api/2.0/permissions/authorization/passwords: - get: - operationId: user-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/users/getpermissions - x-stackQL-resource: user_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of all passwords. Passwords can inherit permissions - from their root object. - put: - operationId: user-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/users/setpermissions - x-stackQL-resource: user_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_USE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: user-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/users/updatepermissions - x-stackQL-resource: user_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_USE - description: Updates the permissions on all passwords. Passwords can inherit - permissions from their root object. - /api/2.0/permissions/authorization/passwords/permissionLevels: - get: - operationId: user-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/users/getpermissionlevels - x-stackQL-resource: user_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. -components: - x-stackQL-resources: - assignable_roles: - id: databricks_workspace.iam.assignable_roles - name: assignable_roles - title: Assignable_roles - methods: - getassignablerolesforresource: - operation: - $ref: '#/paths/~1api~12.0~1preview~1accounts~1access-control~1assignable-roles/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.roles - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/assignable_roles/methods/getassignablerolesforresource' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - currentuser: - id: databricks_workspace.iam.currentuser - name: currentuser - title: Currentuser - methods: - me: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Me/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/currentuser/methods/me' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - groups: - id: databricks_workspace.iam.groups - name: groups - title: Groups - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.Resources - delete: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - patch: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/groups/methods/get' - - $ref: '#/components/x-stackQL-resources/groups/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/groups/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/groups/methods/update' - - $ref: '#/components/x-stackQL-resources/groups/methods/patch' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/groups/methods/delete' - exec: [] - permissions: - id: databricks_workspace.iam.permissions - name: permissions - title: Permissions - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - set: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permissions/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/permissions/methods/update' - replace: - - $ref: '#/components/x-stackQL-resources/permissions/methods/set' - delete: [] - exec: [] - permission_levels: - id: databricks_workspace.iam.permission_levels - name: permission_levels - title: Permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - rule_sets: - id: databricks_workspace.iam.rule_sets - name: rule_sets - title: Rule_sets - methods: - getruleset: - operation: - $ref: '#/paths/~1api~12.0~1preview~1accounts~1access-control~1rule-sets/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updateruleset: - operation: - $ref: '#/paths/~1api~12.0~1preview~1accounts~1access-control~1rule-sets/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/rule_sets/methods/getruleset' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/rule_sets/methods/updateruleset' - replace: [] - delete: [] - exec: [] - service_principals: - id: databricks_workspace.iam.service_principals - name: service_principals - title: Service_principals - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.Resources - delete: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - patch: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/get' - - $ref: '#/components/x-stackQL-resources/service_principals/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/update' - - $ref: '#/components/x-stackQL-resources/service_principals/methods/patch' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/service_principals/methods/delete' - exec: [] - users: - id: databricks_workspace.iam.users - name: users - title: Users - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.Resources - delete: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - patch: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/users/methods/get' - - $ref: '#/components/x-stackQL-resources/users/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/users/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/users/methods/update' - - $ref: '#/components/x-stackQL-resources/users/methods/patch' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/users/methods/delete' - exec: [] - user_permissions: - id: databricks_workspace.iam.user_permissions - name: user_permissions - title: User_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1passwords/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1passwords/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1passwords/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/user_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/user_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/user_permissions/methods/setpermissions' - delete: [] - exec: [] - user_permission_levels: - id: databricks_workspace.iam.user_permission_levels - name: user_permission_levels - title: User_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1passwords~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/user_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] +openapi: 3.0.0 +info: + title: Databricks Iam API (workspace) + description: >- + OpenAPI specification for the Databricks iam service (workspace-level APIs), + generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.iam +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/access-control/check-policy-v2: + get: + operationId: access_control_check_policy + summary: Check access policy to a resource. + tags: + - iam + - access_control + description: |- + Check access policy to a resource. + + :param actor: :class:`Actor` + :param permission: str + :param resource: str + Ex: (servicePrincipal/use, accounts//servicePrincipals/) Ex: + (servicePrincipal.ruleSet/update, accounts//servicePrincipals//ruleSets/default) + :param consistency_token: :class:`ConsistencyToken` + :param authz_identity: :class:`RequestAuthzIdentity` + :param resource_info: :class:`ResourceInfo` (optional) + + :returns: :class:`CheckPolicyResponse` + parameters: + - name: actor + in: query + required: true + schema: + type: string + description: ':param permission: str' + - name: permission + in: query + required: true + schema: + type: string + - name: resource + in: query + required: true + schema: + type: string + description: >- + Ex: (servicePrincipal/use, + accounts//servicePrincipals/) Ex: + (servicePrincipal.ruleSet/update, + accounts//servicePrincipals//ruleSets/default) + - name: consistency_token + in: query + required: true + schema: + type: string + description: ':param authz_identity: :class:`RequestAuthzIdentity`' + - name: authz_identity + in: query + required: true + schema: + type: string + - name: resource_info + in: query + required: false + schema: + type: string + description: ':returns: :class:`CheckPolicyResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CheckPolicyResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccessControlAPI + /api/2.0/preview/accounts/access-control/assignable-roles: + get: + operationId: account_access_control_proxy_get_assignable_roles_for_resource + summary: >- + Gets all the roles that can be granted on an account level resource. A + role is grantable if the rule + tags: + - iam + - account_access_control_proxy + description: >- + Gets all the roles that can be granted on an account level resource. A + role is grantable if the rule + + set on the resource can contain an access rule of the role. + + + :param resource: str + The resource name for which assignable roles will be listed. + + Examples | Summary :--- | :--- `resource=accounts/` | A resource name for the account. + `resource=accounts//groups/` | A resource name for the group. + `resource=accounts//servicePrincipals/` | A resource name for the service + principal. `resource=accounts//tagPolicies/` | A resource name for the + tag policy. + + :returns: :class:`GetAssignableRolesForResourceResponse` + parameters: + - name: resource + in: query + required: true + schema: + type: string + description: >- + The resource name for which assignable roles will be listed. + Examples | Summary :--- | :--- `resource=accounts/` | A + resource name for the account. + `resource=accounts//groups/` | A resource name + for the group. + `resource=accounts//servicePrincipals/` | A + resource name for the service principal. + `resource=accounts//tagPolicies/` | A + resource name for the tag policy. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetAssignableRolesForResourceResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountAccessControlProxyAPI + /api/2.0/preview/accounts/access-control/rule-sets: + get: + operationId: account_access_control_proxy_get_rule_set + summary: >- + Get a rule set by its name. A rule set is always attached to a resource + and contains a list of access + tags: + - iam + - account_access_control_proxy + description: >- + Get a rule set by its name. A rule set is always attached to a resource + and contains a list of access + + rules on the said resource. Currently only a default rule set for each + resource is supported. + + + :param name: str + The ruleset name associated with the request. + + Examples | Summary :--- | :--- `name=accounts//ruleSets/default` | A name for a rule set + on the account. `name=accounts//groups//ruleSets/default` | A name for a rule + set on the group. + `name=accounts//servicePrincipals//ruleSets/default` | + A name for a rule set on the service principal. + `name=accounts//tagPolicies//ruleSets/default` | A name for a rule set on + the tag policy. + :param etag: str + Etag used for versioning. The response is at least as fresh as the eTag provided. Etag is used for + optimistic concurrency control as a way to help prevent simultaneous updates of a rule set from + overwriting each other. It is strongly suggested that systems make use of the etag in the read -> + modify -> write pattern to perform rule set updates in order to avoid race conditions that is get an + etag from a GET rule set request, and pass it with the PUT update request to identify the rule set + version you are updating. + + Examples | Summary :--- | :--- `etag=` | An empty etag can only be used in GET to indicate no + freshness requirements. `etag=RENUAAABhSweA4NvVmmUYdiU717H3Tgy0UJdor3gE4a+mq/oj9NjAf8ZsQ==` | An + etag encoded a specific version of the rule set to get or to be updated. + + :returns: :class:`RuleSetResponse` + parameters: + - name: name + in: query + required: true + schema: + type: string + description: >- + The ruleset name associated with the request. Examples | Summary + :--- | :--- `name=accounts//ruleSets/default` | A name + for a rule set on the account. + `name=accounts//groups//ruleSets/default` | A + name for a rule set on the group. + `name=accounts//servicePrincipals//ruleSets/default` + | A name for a rule set on the service principal. + `name=accounts//tagPolicies//ruleSets/default` + | A name for a rule set on the tag policy. + - name: etag + in: query + required: true + schema: + type: string + description: >- + Etag used for versioning. The response is at least as fresh as the + eTag provided. Etag is used for optimistic concurrency control as a + way to help prevent simultaneous updates of a rule set from + overwriting each other. It is strongly suggested that systems make + use of the etag in the read -> modify -> write pattern to perform + rule set updates in order to avoid race conditions that is get an + etag from a GET rule set request, and pass it with the PUT update + request to identify the rule set version you are updating. Examples + | Summary :--- | :--- `etag=` | An empty etag can only be used in + GET to indicate no freshness requirements. + `etag=RENUAAABhSweA4NvVmmUYdiU717H3Tgy0UJdor3gE4a+mq/oj9NjAf8ZsQ==` + | An etag encoded a specific version of the rule set to get or to be + updated. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RuleSetResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountAccessControlProxyAPI + put: + operationId: account_access_control_proxy_update_rule_set + summary: >- + Replace the rules of a rule set. First, use get to read the current + version of the rule set before + tags: + - iam + - account_access_control_proxy + description: >- + Replace the rules of a rule set. First, use get to read the current + version of the rule set before + + modifying it. This pattern helps prevent conflicts between concurrent + updates. + + + :param name: str + Name of the rule set. + :param rule_set: :class:`RuleSetUpdateRequest` + + + :returns: :class:`RuleSetResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the rule set. + rule_set: + type: string + description: ':returns: :class:`RuleSetResponse`' + required: + - name + - rule_set + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RuleSetResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AccountAccessControlProxyAPI + /api/2.0/preview/scim/v2/Me: + get: + operationId: current_user_me + summary: Get details about the current method caller's identity. + tags: + - iam + - current_user + description: |- + Get details about the current method caller's identity. + + + :returns: :class:`User` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/User' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CurrentUserAPI + /api/2.0/preview/scim/v2/Groups: + post: + operationId: groups_v2_create + summary: >- + Creates a group in the Databricks workspace with a unique name, using + the supplied group details. + tags: + - iam + - groups_v2 + description: >- + Creates a group in the Databricks workspace with a unique name, using + the supplied group details. + + + :param display_name: str (optional) + String that represents a human-readable group name + :param entitlements: List[:class:`ComplexValue`] (optional) + Entitlements assigned to the group. See [assigning entitlements] for a full list of supported + values. + + [assigning entitlements]: https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + :param external_id: str (optional) + + :param groups: List[:class:`ComplexValue`] (optional) + + :param id: str (optional) + Databricks group ID + :param members: List[:class:`ComplexValue`] (optional) + + :param meta: :class:`ResourceMeta` (optional) + Container for the group identifier. Workspace local versus account. + :param roles: List[:class:`ComplexValue`] (optional) + Corresponds to AWS instance profile/arn role. + :param schemas: List[:class:`GroupSchema`] (optional) + The schema of the group. + + :returns: :class:`Group` + requestBody: + content: + application/json: + schema: + type: object + properties: + display_name: + type: string + description: String that represents a human-readable group name + entitlements: + type: string + description: >- + Entitlements assigned to the group. See [assigning + entitlements] for a full list of supported values. + [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + external_id: + type: string + description: ':param groups: List[:class:`ComplexValue`] (optional)' + groups: + type: string + id: + type: string + description: Databricks group ID + members: + type: string + description: >- + :param meta: :class:`ResourceMeta` (optional) Container for + the group identifier. Workspace local versus account. + meta: + type: string + roles: + type: string + description: Corresponds to AWS instance profile/arn role. + schemas: + type: string + description: The schema of the group. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Group' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GroupsV2API + get: + operationId: groups_v2_list + summary: Gets all details of the groups associated with the Databricks workspace. + tags: + - iam + - groups_v2 + description: |- + Gets all details of the groups associated with the Databricks workspace. + + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. + :param sort_order: :class:`ListSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: Iterator over :class:`Group` + parameters: + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: Attribute to sort the results. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListGroupsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GroupsV2API + /api/2.0/preview/scim/v2/Groups/{id}: + delete: + operationId: groups_v2_delete + summary: Deletes a group from the Databricks workspace. + tags: + - iam + - groups_v2 + description: |- + Deletes a group from the Databricks workspace. + + :param id: str + Unique ID for a group in the Databricks workspace. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a group in the Databricks workspace. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GroupsV2API + get: + operationId: groups_v2_get + summary: Gets the information for a specific group in the Databricks workspace. + tags: + - iam + - groups_v2 + description: |- + Gets the information for a specific group in the Databricks workspace. + + :param id: str + Unique ID for a group in the Databricks workspace. + + :returns: :class:`Group` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a group in the Databricks workspace. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Group' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GroupsV2API + patch: + operationId: groups_v2_patch + summary: Partially updates the details of a group. + tags: + - iam + - groups_v2 + description: |- + Partially updates the details of a group. + + :param id: str + Unique ID in the Databricks workspace. + :param operations: List[:class:`Patch`] (optional) + :param schemas: List[:class:`PatchSchema`] (optional) + The schema of the patch request. Must be ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID in the Databricks workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + operations: + type: string + description: >- + :param schemas: List[:class:`PatchSchema`] (optional) The + schema of the patch request. Must be + ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + schemas: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GroupsV2API + put: + operationId: groups_v2_update + summary: Updates the details of a group by replacing the entire group entity. + tags: + - iam + - groups_v2 + description: |- + Updates the details of a group by replacing the entire group entity. + + :param id: str + Databricks group ID + :param display_name: str (optional) + String that represents a human-readable group name + :param entitlements: List[:class:`ComplexValue`] (optional) + Entitlements assigned to the group. See [assigning entitlements] for a full list of supported + values. + + [assigning entitlements]: https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + :param external_id: str (optional) + :param groups: List[:class:`ComplexValue`] (optional) + :param members: List[:class:`ComplexValue`] (optional) + :param meta: :class:`ResourceMeta` (optional) + Container for the group identifier. Workspace local versus account. + :param roles: List[:class:`ComplexValue`] (optional) + Corresponds to AWS instance profile/arn role. + :param schemas: List[:class:`GroupSchema`] (optional) + The schema of the group. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Databricks group ID + requestBody: + content: + application/json: + schema: + type: object + properties: + display_name: + type: string + description: String that represents a human-readable group name + entitlements: + type: string + description: >- + Entitlements assigned to the group. See [assigning + entitlements] for a full list of supported values. + [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + external_id: + type: string + description: ':param groups: List[:class:`ComplexValue`] (optional)' + groups: + type: string + members: + type: string + description: >- + :param meta: :class:`ResourceMeta` (optional) Container for + the group identifier. Workspace local versus account. + meta: + type: string + roles: + type: string + description: Corresponds to AWS instance profile/arn role. + schemas: + type: string + description: The schema of the group. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GroupsV2API + /api/2.0/permissionmigration: + post: + operationId: permission_migration_migrate_permissions + summary: Migrate Permissions. + tags: + - iam + - permission_migration + description: |- + Migrate Permissions. + + :param workspace_id: int + WorkspaceId of the associated workspace where the permission migration will occur. + :param from_workspace_group_name: str + The name of the workspace group that permissions will be migrated from. + :param to_account_group_name: str + The name of the account group that permissions will be migrated to. + :param size: int (optional) + The maximum number of permissions that will be migrated. + + :returns: :class:`MigratePermissionsResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + workspace_id: + type: integer + description: >- + WorkspaceId of the associated workspace where the permission + migration will occur. + from_workspace_group_name: + type: string + description: >- + The name of the workspace group that permissions will be + migrated from. + to_account_group_name: + type: string + description: >- + The name of the account group that permissions will be + migrated to. + size: + type: string + description: The maximum number of permissions that will be migrated. + required: + - workspace_id + - from_workspace_group_name + - to_account_group_name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MigratePermissionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PermissionMigrationAPI + /api/2.0/permissions/{request_object_type}/{request_object_id}: + get: + operationId: permissions_get + summary: >- + Gets the permissions of an object. Objects can inherit permissions from + their parent objects or root + tags: + - iam + - permissions + description: >- + Gets the permissions of an object. Objects can inherit permissions from + their parent objects or root + + object. + + + :param request_object_type: str + The type of the request object. Can be one of the following: alerts, alertsv2, authorization, + clusters, cluster-policies, dashboards, dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, registered-models, repos, serving-endpoints, or + warehouses. + :param request_object_id: str + The id of the request object. + + :returns: :class:`ObjectPermissions` + parameters: + - name: request_object_type + in: path + required: true + schema: + type: string + description: >- + The type of the request object. Can be one of the following: alerts, + alertsv2, authorization, clusters, cluster-policies, dashboards, + dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, + registered-models, repos, serving-endpoints, or warehouses. + - name: request_object_id + in: path + required: true + schema: + type: string + description: The id of the request object. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ObjectPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PermissionsAPI + put: + operationId: permissions_set + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - iam + - permissions + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their parent objects or root + + object. + + + :param request_object_type: str + The type of the request object. Can be one of the following: alerts, alertsv2, authorization, + clusters, cluster-policies, dashboards, dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, registered-models, repos, serving-endpoints, or + warehouses. + :param request_object_id: str + The id of the request object. + :param access_control_list: List[:class:`AccessControlRequest`] + (optional) + + + :returns: :class:`ObjectPermissions` + parameters: + - name: request_object_type + in: path + required: true + schema: + type: string + description: >- + The type of the request object. Can be one of the following: alerts, + alertsv2, authorization, clusters, cluster-policies, dashboards, + dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, + registered-models, repos, serving-endpoints, or warehouses. + - name: request_object_id + in: path + required: true + schema: + type: string + description: The id of the request object. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ObjectPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ObjectPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PermissionsAPI + patch: + operationId: permissions_update + summary: >- + Updates the permissions on an object. Objects can inherit permissions + from their parent objects or + tags: + - iam + - permissions + description: >- + Updates the permissions on an object. Objects can inherit permissions + from their parent objects or + + root object. + + + :param request_object_type: str + The type of the request object. Can be one of the following: alerts, alertsv2, authorization, + clusters, cluster-policies, dashboards, dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, registered-models, repos, serving-endpoints, or + warehouses. + :param request_object_id: str + The id of the request object. + :param access_control_list: List[:class:`AccessControlRequest`] + (optional) + + + :returns: :class:`ObjectPermissions` + parameters: + - name: request_object_type + in: path + required: true + schema: + type: string + description: >- + The type of the request object. Can be one of the following: alerts, + alertsv2, authorization, clusters, cluster-policies, dashboards, + dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, + registered-models, repos, serving-endpoints, or warehouses. + - name: request_object_id + in: path + required: true + schema: + type: string + description: The id of the request object. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ObjectPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ObjectPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PermissionsAPI + /api/2.0/permissions/{request_object_type}/{request_object_id}/permissionLevels: + get: + operationId: permissions_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - iam + - permissions + description: |- + Gets the permission levels that a user can have on an object. + + :param request_object_type: str + The type of the request object. Can be one of the following: alerts, alertsv2, authorization, + clusters, cluster-policies, dashboards, dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, registered-models, repos, serving-endpoints, or + warehouses. + :param request_object_id: str + + :returns: :class:`GetPermissionLevelsResponse` + parameters: + - name: request_object_type + in: path + required: true + schema: + type: string + description: >- + The type of the request object. Can be one of the following: alerts, + alertsv2, authorization, clusters, cluster-policies, dashboards, + dbsql-dashboards, directories, experiments, files, genie, + instance-pools, jobs, notebooks, pipelines, queries, + registered-models, repos, serving-endpoints, or warehouses. + - name: request_object_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetPermissionLevelsResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PermissionsAPI + /api/2.0/preview/scim/v2/ServicePrincipals: + post: + operationId: service_principals_v2_create + summary: Creates a new service principal in the Databricks workspace. + tags: + - iam + - service_principals_v2 + description: |- + Creates a new service principal in the Databricks workspace. + + :param active: bool (optional) + If this user is active + :param application_id: str (optional) + UUID relating to the service principal + :param display_name: str (optional) + String that represents a concatenation of given and family names. + :param entitlements: List[:class:`ComplexValue`] (optional) + Entitlements assigned to the service principal. See [assigning entitlements] for a full list of + supported values. + + [assigning entitlements]: https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + :param external_id: str (optional) + :param groups: List[:class:`ComplexValue`] (optional) + :param id: str (optional) + Databricks service principal ID. + :param roles: List[:class:`ComplexValue`] (optional) + Corresponds to AWS instance profile/arn role. + :param schemas: List[:class:`ServicePrincipalSchema`] (optional) + The schema of the List response. + + :returns: :class:`ServicePrincipal` + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + application_id: + type: string + description: UUID relating to the service principal + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. + entitlements: + type: string + description: >- + Entitlements assigned to the service principal. See + [assigning entitlements] for a full list of supported + values. [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + external_id: + type: string + description: ':param groups: List[:class:`ComplexValue`] (optional)' + groups: + type: string + id: + type: string + description: Databricks service principal ID. + roles: + type: string + description: Corresponds to AWS instance profile/arn role. + schemas: + type: string + description: The schema of the List response. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServicePrincipal' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalsV2API + get: + operationId: service_principals_v2_list + summary: >- + Gets the set of service principals associated with a Databricks + workspace. + tags: + - iam + - service_principals_v2 + description: >- + Gets the set of service principals associated with a Databricks + workspace. + + + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. + :param sort_order: :class:`ListSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: Iterator over :class:`ServicePrincipal` + parameters: + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: Attribute to sort the results. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListServicePrincipalResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalsV2API + /api/2.0/preview/scim/v2/ServicePrincipals/{id}: + delete: + operationId: service_principals_v2_delete + summary: Delete a single service principal in the Databricks workspace. + tags: + - iam + - service_principals_v2 + description: |- + Delete a single service principal in the Databricks workspace. + + :param id: str + Unique ID for a service principal in the Databricks workspace. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a service principal in the Databricks workspace. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalsV2API + get: + operationId: service_principals_v2_get + summary: >- + Gets the details for a single service principal define in the Databricks + workspace. + tags: + - iam + - service_principals_v2 + description: >- + Gets the details for a single service principal define in the Databricks + workspace. + + + :param id: str + Unique ID for a service principal in the Databricks workspace. + + :returns: :class:`ServicePrincipal` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a service principal in the Databricks workspace. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServicePrincipal' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalsV2API + patch: + operationId: service_principals_v2_patch + summary: >- + Partially updates the details of a single service principal in the + Databricks workspace. + tags: + - iam + - service_principals_v2 + description: >- + Partially updates the details of a single service principal in the + Databricks workspace. + + + :param id: str + Unique ID in the Databricks workspace. + :param operations: List[:class:`Patch`] (optional) + + :param schemas: List[:class:`PatchSchema`] (optional) + The schema of the patch request. Must be ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID in the Databricks workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + operations: + type: string + description: >- + :param schemas: List[:class:`PatchSchema`] (optional) The + schema of the patch request. Must be + ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + schemas: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalsV2API + put: + operationId: service_principals_v2_update + summary: Updates the details of a single service principal. + tags: + - iam + - service_principals_v2 + description: |- + Updates the details of a single service principal. + + This action replaces the existing service principal with the same name. + + :param id: str + Databricks service principal ID. + :param active: bool (optional) + If this user is active + :param application_id: str (optional) + UUID relating to the service principal + :param display_name: str (optional) + String that represents a concatenation of given and family names. + :param entitlements: List[:class:`ComplexValue`] (optional) + Entitlements assigned to the service principal. See [assigning entitlements] for a full list of + supported values. + + [assigning entitlements]: https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + :param external_id: str (optional) + :param groups: List[:class:`ComplexValue`] (optional) + :param roles: List[:class:`ComplexValue`] (optional) + Corresponds to AWS instance profile/arn role. + :param schemas: List[:class:`ServicePrincipalSchema`] (optional) + The schema of the List response. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Databricks service principal ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + application_id: + type: string + description: UUID relating to the service principal + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. + entitlements: + type: string + description: >- + Entitlements assigned to the service principal. See + [assigning entitlements] for a full list of supported + values. [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + external_id: + type: string + description: ':param groups: List[:class:`ComplexValue`] (optional)' + groups: + type: string + roles: + type: string + description: Corresponds to AWS instance profile/arn role. + schemas: + type: string + description: The schema of the List response. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalsV2API + /api/2.0/preview/scim/v2/Users: + post: + operationId: users_v2_create + summary: >- + Creates a new user in the Databricks workspace. This new user will also + be added to the Databricks + tags: + - iam + - users_v2 + description: >- + Creates a new user in the Databricks workspace. This new user will also + be added to the Databricks + + account. + + + :param active: bool (optional) + If this user is active + :param display_name: str (optional) + String that represents a concatenation of given and family names. For example `John Smith`. This + field cannot be updated through the Workspace SCIM APIs when [identity federation is enabled]. Use + Account SCIM APIs to update `displayName`. + + [identity federation is enabled]: https://docs.databricks.com/administration-guide/users-groups/best-practices.html#enable-identity-federation + :param emails: List[:class:`ComplexValue`] (optional) + All the emails associated with the Databricks user. + :param entitlements: List[:class:`ComplexValue`] (optional) + Entitlements assigned to the user. See [assigning entitlements] for a full list of supported values. + + [assigning entitlements]: https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + :param external_id: str (optional) + External ID is not currently supported. It is reserved for future use. + :param groups: List[:class:`ComplexValue`] (optional) + + :param id: str (optional) + Databricks user ID. + :param name: :class:`Name` (optional) + + :param roles: List[:class:`ComplexValue`] (optional) + Corresponds to AWS instance profile/arn role. + :param schemas: List[:class:`UserSchema`] (optional) + The schema of the user. + :param user_name: str (optional) + Email address of the Databricks user. + + :returns: :class:`User` + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. For example `John Smith`. This field cannot be + updated through the Workspace SCIM APIs when [identity + federation is enabled]. Use Account SCIM APIs to update + `displayName`. [identity federation is enabled]: + https://docs.databricks.com/administration-guide/users-groups/best-practices.html#enable-identity-federation + emails: + type: string + description: All the emails associated with the Databricks user. + entitlements: + type: string + description: >- + Entitlements assigned to the user. See [assigning + entitlements] for a full list of supported values. + [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + external_id: + type: string + description: >- + External ID is not currently supported. It is reserved for + future use. + groups: + type: string + description: ':param id: str (optional) Databricks user ID.' + id: + type: string + name: + type: string + description: >- + :param roles: List[:class:`ComplexValue`] (optional) + Corresponds to AWS instance profile/arn role. + roles: + type: string + schemas: + type: string + description: The schema of the user. + user_name: + type: string + description: Email address of the Databricks user. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/User' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + get: + operationId: users_v2_list + summary: Gets details for all the users associated with a Databricks workspace. + tags: + - iam + - users_v2 + description: |- + Gets details for all the users associated with a Databricks workspace. + + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. Multi-part paths are supported. For example, `userName`, + `name.givenName`, and `emails`. + :param sort_order: :class:`ListSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: Iterator over :class:`User` + parameters: + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: >- + Attribute to sort the results. Multi-part paths are supported. For + example, `userName`, `name.givenName`, and `emails`. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListUsersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + /api/2.0/preview/scim/v2/Users/{id}: + delete: + operationId: users_v2_delete + summary: >- + Deletes a user. Deleting a user from a Databricks workspace also removes + objects associated with the + tags: + - iam + - users_v2 + description: >- + Deletes a user. Deleting a user from a Databricks workspace also removes + objects associated with the + + user. + + + :param id: str + Unique ID for a user in the Databricks workspace. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a user in the Databricks workspace. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + get: + operationId: users_v2_get + summary: Gets information for a specific user in Databricks workspace. + tags: + - iam + - users_v2 + description: |- + Gets information for a specific user in Databricks workspace. + + :param id: str + Unique ID for a user in the Databricks workspace. + :param attributes: str (optional) + Comma-separated list of attributes to return in response. + :param count: int (optional) + Desired number of results per page. + :param excluded_attributes: str (optional) + Comma-separated list of attributes to exclude in response. + :param filter: str (optional) + Query by which the results have to be filtered. Supported operators are equals(`eq`), + contains(`co`), starts with(`sw`) and not equals(`ne`). Additionally, simple expressions can be + formed using logical operators - `and` and `or`. The [SCIM RFC] has more details but we currently + only support simple expressions. + + [SCIM RFC]: https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + :param sort_by: str (optional) + Attribute to sort the results. Multi-part paths are supported. For example, `userName`, + `name.givenName`, and `emails`. + :param sort_order: :class:`GetSortOrder` (optional) + The order to sort the results. + :param start_index: int (optional) + Specifies the index of the first result. First item is number 1. + + :returns: :class:`User` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID for a user in the Databricks workspace. + - name: attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to return in response. + - name: count + in: query + required: false + schema: + type: string + description: Desired number of results per page. + - name: excluded_attributes + in: query + required: false + schema: + type: string + description: Comma-separated list of attributes to exclude in response. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Query by which the results have to be filtered. Supported operators + are equals(`eq`), contains(`co`), starts with(`sw`) and not + equals(`ne`). Additionally, simple expressions can be formed using + logical operators - `and` and `or`. The [SCIM RFC] has more details + but we currently only support simple expressions. [SCIM RFC]: + https://tools.ietf.org/html/rfc7644#section-3.4.2.2 + - name: sort_by + in: query + required: false + schema: + type: string + description: >- + Attribute to sort the results. Multi-part paths are supported. For + example, `userName`, `name.givenName`, and `emails`. + - name: sort_order + in: query + required: false + schema: + type: string + description: The order to sort the results. + - name: start_index + in: query + required: false + schema: + type: string + description: Specifies the index of the first result. First item is number 1. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/User' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + patch: + operationId: users_v2_patch + summary: >- + Partially updates a user resource by applying the supplied operations on + specific user attributes. + tags: + - iam + - users_v2 + description: >- + Partially updates a user resource by applying the supplied operations on + specific user attributes. + + + :param id: str + Unique ID in the Databricks workspace. + :param operations: List[:class:`Patch`] (optional) + + :param schemas: List[:class:`PatchSchema`] (optional) + The schema of the patch request. Must be ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Unique ID in the Databricks workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + operations: + type: string + description: >- + :param schemas: List[:class:`PatchSchema`] (optional) The + schema of the patch request. Must be + ["urn:ietf:params:scim:api:messages:2.0:PatchOp"]. + schemas: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + put: + operationId: users_v2_update + summary: Replaces a user's information with the data supplied in request. + tags: + - iam + - users_v2 + description: |- + Replaces a user's information with the data supplied in request. + + :param id: str + Databricks user ID. + :param active: bool (optional) + If this user is active + :param display_name: str (optional) + String that represents a concatenation of given and family names. For example `John Smith`. This + field cannot be updated through the Workspace SCIM APIs when [identity federation is enabled]. Use + Account SCIM APIs to update `displayName`. + + [identity federation is enabled]: https://docs.databricks.com/administration-guide/users-groups/best-practices.html#enable-identity-federation + :param emails: List[:class:`ComplexValue`] (optional) + All the emails associated with the Databricks user. + :param entitlements: List[:class:`ComplexValue`] (optional) + Entitlements assigned to the user. See [assigning entitlements] for a full list of supported values. + + [assigning entitlements]: https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + :param external_id: str (optional) + External ID is not currently supported. It is reserved for future use. + :param groups: List[:class:`ComplexValue`] (optional) + :param name: :class:`Name` (optional) + :param roles: List[:class:`ComplexValue`] (optional) + Corresponds to AWS instance profile/arn role. + :param schemas: List[:class:`UserSchema`] (optional) + The schema of the user. + :param user_name: str (optional) + Email address of the Databricks user. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Databricks user ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + active: + type: string + description: If this user is active + display_name: + type: string + description: >- + String that represents a concatenation of given and family + names. For example `John Smith`. This field cannot be + updated through the Workspace SCIM APIs when [identity + federation is enabled]. Use Account SCIM APIs to update + `displayName`. [identity federation is enabled]: + https://docs.databricks.com/administration-guide/users-groups/best-practices.html#enable-identity-federation + emails: + type: string + description: All the emails associated with the Databricks user. + entitlements: + type: string + description: >- + Entitlements assigned to the user. See [assigning + entitlements] for a full list of supported values. + [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + external_id: + type: string + description: >- + External ID is not currently supported. It is reserved for + future use. + groups: + type: string + description: ':param name: :class:`Name` (optional)' + name: + type: string + roles: + type: string + description: Corresponds to AWS instance profile/arn role. + schemas: + type: string + description: The schema of the user. + user_name: + type: string + description: Email address of the Databricks user. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + /api/2.0/permissions/authorization/passwords/permissionLevels: + get: + operationId: users_v2_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - iam + - users_v2 + description: |- + Gets the permission levels that a user can have on an object. + + + :returns: :class:`GetPasswordPermissionLevelsResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPasswordPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + /api/2.0/permissions/authorization/passwords: + get: + operationId: users_v2_get_permissions + summary: >- + Gets the permissions of all passwords. Passwords can inherit permissions + from their root object. + tags: + - iam + - users_v2 + description: >- + Gets the permissions of all passwords. Passwords can inherit permissions + from their root object. + + + + :returns: :class:`PasswordPermissions` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PasswordPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + put: + operationId: users_v2_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - iam + - users_v2 + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param access_control_list: List[:class:`PasswordAccessControlRequest`] + (optional) + + + :returns: :class:`PasswordPermissions` + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`PasswordPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PasswordPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API + patch: + operationId: users_v2_update_permissions + summary: >- + Updates the permissions on all passwords. Passwords can inherit + permissions from their root object. + tags: + - iam + - users_v2 + description: >- + Updates the permissions on all passwords. Passwords can inherit + permissions from their root object. + + + :param access_control_list: List[:class:`PasswordAccessControlRequest`] + (optional) + + + :returns: :class:`PasswordPermissions` + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`PasswordPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PasswordPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: UsersV2API +components: + schemas: + AccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/PermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + AccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/Permission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + AccountGroup: + type: object + properties: + account_id: + type: string + displayName: + type: string + description: String that represents a human-readable group name + externalId: + type: string + description: external_id should be unique for identifying groups + id: + type: string + description: Databricks group ID + members: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + meta: + $ref: '#/components/schemas/ResourceMeta' + description: Container for the group identifier. Workspace local versus account. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Indicates if the group has the admin role. + AccountServicePrincipal: + type: object + properties: + account_id: + type: string + active: + type: boolean + description: If this user is active + applicationId: + type: string + description: UUID relating to the service principal + displayName: + type: string + description: String that represents a concatenation of given and family names. + externalId: + type: string + id: + type: string + description: Databricks service principal ID. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Indicates if the group has the admin role. + AccountUser: + type: object + properties: + account_id: + type: string + active: + type: boolean + description: If this user is active + displayName: + type: string + description: >- + String that represents a concatenation of given and family names. + For example `John Smith`. + emails: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: All the emails associated with the Databricks user. + externalId: + type: string + description: >- + External ID is not currently supported. It is reserved for future + use. + id: + type: string + description: Databricks user ID. + name: + $ref: '#/components/schemas/Name' + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Indicates if the group has the admin role. + userName: + type: string + description: Email address of the Databricks user. + Actor: + type: object + properties: + actor_id: + type: integer + description: >- + represents an identity trying to access a resource - user or a service + principal group can be a + principal of a permission set assignment but an actor is always a user or a service principal + CheckPolicyResponse: + type: object + properties: + consistency_token: + $ref: '#/components/schemas/ConsistencyToken' + is_permitted: + type: boolean + required: + - consistency_token + ComplexValue: + type: object + properties: + display: + type: string + primary: + type: boolean + $ref: + type: string + type: + type: string + value: + type: string + ConsistencyToken: + type: object + properties: + value: + type: string + required: + - value + DeleteWorkspacePermissionAssignmentResponse: + type: object + properties: {} + GetAssignableRolesForResourceResponse: + type: object + properties: + roles: + type: array + items: + $ref: '#/components/schemas/Role' + GetPasswordPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/PasswordPermissionsDescription' + GetPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/PermissionsDescription' + GrantRule: + type: object + properties: + role: + type: string + principals: + type: array + items: + type: string + description: >- + Principals this grant rule applies to. A principal can be a user + (for end users), a service principal (for applications and compute + workloads), or an account group. Each principal has its own + identifier format: * users/ * groups/ * + servicePrincipals/ + required: + - role + Group: + type: object + properties: + displayName: + type: string + entitlements: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: >- + Entitlements assigned to the group. See [assigning entitlements] for + a full list of supported values. [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + externalId: + type: string + description: external_id should be unique for identifying groups + groups: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + id: + type: string + description: Databricks group ID + members: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + meta: + $ref: '#/components/schemas/ResourceMeta' + description: Container for the group identifier. Workspace local versus account. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Corresponds to AWS instance profile/arn role. + schemas: + type: array + items: + $ref: '#/components/schemas/GroupSchema' + description: The schema of the group. + ListAccountGroupsResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/AccountGroup' + description: User objects returned in the response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListAccountServicePrincipalsResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/AccountServicePrincipal' + description: User objects returned in the response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListAccountUsersResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/AccountUser' + description: User objects returned in the response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListGroupsResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/Group' + description: User objects returned in the response. + schemas: + type: array + items: + $ref: '#/components/schemas/ListResponseSchema' + description: The schema of the service principal. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListServicePrincipalResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/ServicePrincipal' + description: User objects returned in the response. + schemas: + type: array + items: + $ref: '#/components/schemas/ListResponseSchema' + description: The schema of the List response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + ListUsersResponse: + type: object + properties: + itemsPerPage: + type: integer + Resources: + type: array + items: + $ref: '#/components/schemas/User' + description: User objects returned in the response. + schemas: + type: array + items: + $ref: '#/components/schemas/ListResponseSchema' + description: The schema of the List response. + startIndex: + type: integer + description: >- + Starting index of all the results that matched the request filters. + First item is number 1. + totalResults: + type: integer + description: Total results that match the request filters. + MigratePermissionsResponse: + type: object + properties: + permissions_migrated: + type: integer + Name: + type: object + properties: + familyName: + type: string + givenName: + type: string + description: Given name of the Databricks user. + ObjectPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/AccessControlResponse' + object_id: + type: string + object_type: + type: string + PasswordAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/PasswordPermissionLevel' + description: Permission level + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + PasswordAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/PasswordPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + PasswordPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/PasswordPermissionLevel' + description: Permission level + PasswordPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/PasswordAccessControlResponse' + object_id: + type: string + object_type: + type: string + PasswordPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/PasswordPermissionLevel' + description: Permission level + Patch: + type: object + properties: + op: + $ref: '#/components/schemas/PatchOp' + path: + type: string + description: Selection of patch operation + value: + type: object + description: Value to modify + Permission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/PermissionLevel' + PermissionAssignment: + type: object + properties: + error: + type: string + description: >- + Error response associated with a workspace permission assignment, if + any. + permissions: + type: array + items: + $ref: '#/components/schemas/WorkspacePermission' + description: The permissions level of the principal. + principal: + $ref: '#/components/schemas/PrincipalOutput' + description: Information about the principal assigned to the workspace. + description: >- + The output format for existing workspace PermissionAssignment records, + which contains some info + for user consumption. + PermissionAssignments: + type: object + properties: + permission_assignments: + type: array + items: + $ref: '#/components/schemas/PermissionAssignment' + PermissionOutput: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/WorkspacePermission' + PermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/PermissionLevel' + PrincipalOutput: + type: object + properties: + display_name: + type: string + description: The display name of the principal. + group_name: + type: string + description: >- + The group name of the group. Present only if the principal is a + group. + principal_id: + type: integer + description: The unique, opaque id of the principal. + service_principal_name: + type: string + description: >- + The name of the service principal. Present only if the principal is + a service principal. + user_name: + type: string + description: The username of the user. Present only if the principal is a user. + description: Information about the principal assigned to the workspace. + ResourceInfo: + type: object + properties: + id: + type: string + legacy_acl_path: + type: string + description: The legacy acl path of the current resource. + parent_resource_info: + $ref: '#/components/schemas/ResourceInfo' + description: >- + Parent resource info for the current resource. The parent may have + another parent. + required: + - id + ResourceMeta: + type: object + properties: + resourceType: + type: string + Role: + type: object + properties: + name: + type: string + required: + - name + RuleSetResponse: + type: object + properties: + name: + type: string + etag: + type: string + description: >- + Identifies the version of the rule set returned. Etag used for + versioning. The response is at least as fresh as the eTag provided. + Etag is used for optimistic concurrency control as a way to help + prevent simultaneous updates of a rule set from overwriting each + other. It is strongly suggested that systems make use of the etag in + the read -> modify -> write pattern to perform rule set updates in + order to avoid race conditions that is get an etag from a GET rule + set request, and pass it with the PUT update request to identify the + rule set version you are updating. + grant_rules: + type: array + items: + $ref: '#/components/schemas/GrantRule' + required: + - name + - etag + RuleSetUpdateRequest: + type: object + properties: + name: + type: string + etag: + type: string + description: >- + Identifies the version of the rule set returned. Etag used for + versioning. The response is at least as fresh as the eTag provided. + Etag is used for optimistic concurrency control as a way to help + prevent simultaneous updates of a rule set from overwriting each + other. It is strongly suggested that systems make use of the etag in + the read -> modify -> write pattern to perform rule set updates in + order to avoid race conditions that is get an etag from a GET rule + set request, and pass it with the PUT update request to identify the + rule set version you are updating. + grant_rules: + type: array + items: + $ref: '#/components/schemas/GrantRule' + required: + - name + - etag + ServicePrincipal: + type: object + properties: + active: + type: boolean + applicationId: + type: string + description: UUID relating to the service principal + displayName: + type: string + description: String that represents a concatenation of given and family names. + entitlements: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: >- + Entitlements assigned to the service principal. See [assigning + entitlements] for a full list of supported values. [assigning + entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + externalId: + type: string + groups: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + id: + type: string + description: Databricks service principal ID. + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Corresponds to AWS instance profile/arn role. + schemas: + type: array + items: + $ref: '#/components/schemas/ServicePrincipalSchema' + description: The schema of the List response. + User: + type: object + properties: + active: + type: boolean + displayName: + type: string + description: >- + String that represents a concatenation of given and family names. + For example `John Smith`. This field cannot be updated through the + Workspace SCIM APIs when [identity federation is enabled]. Use + Account SCIM APIs to update `displayName`. [identity federation is + enabled]: + https://docs.databricks.com/administration-guide/users-groups/best-practices.html#enable-identity-federation + emails: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: All the emails associated with the Databricks user. + entitlements: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: >- + Entitlements assigned to the user. See [assigning entitlements] for + a full list of supported values. [assigning entitlements]: + https://docs.databricks.com/administration-guide/users-groups/index.html#assigning-entitlements + externalId: + type: string + description: >- + External ID is not currently supported. It is reserved for future + use. + groups: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + id: + type: string + description: Databricks user ID. + name: + $ref: '#/components/schemas/Name' + roles: + type: array + items: + $ref: '#/components/schemas/ComplexValue' + description: Corresponds to AWS instance profile/arn role. + schemas: + type: array + items: + $ref: '#/components/schemas/UserSchema' + description: The schema of the user. + userName: + type: string + description: Email address of the Databricks user. + WorkspacePermissions: + type: object + properties: + permissions: + type: array + items: + $ref: '#/components/schemas/PermissionOutput' + GetSortOrder: + type: string + enum: + - ascending + - descending + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + GroupSchema: + type: string + enum: + - urn:ietf:params:scim:schemas:core:2.0:Group + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListResponseSchema: + type: string + enum: + - urn:ietf:params:scim:api:messages:2.0:ListResponse + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListSortOrder: + type: string + enum: + - ascending + - descending + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PasswordPermissionLevel: + type: string + enum: + - CAN_USE + description: Permission level + PatchOp: + type: string + enum: + - add + - remove + - replace + description: Type of patch operation. + PatchSchema: + type: string + enum: + - urn:ietf:params:scim:api:messages:2.0:PatchOp + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PermissionLevel: + type: string + enum: + - CAN_ATTACH_TO + - CAN_BIND + - CAN_CREATE + - CAN_EDIT + - CAN_EDIT_METADATA + - CAN_MANAGE + - CAN_MANAGE_PRODUCTION_VERSIONS + - CAN_MANAGE_RUN + - CAN_MANAGE_STAGING_VERSIONS + - CAN_MONITOR + - CAN_MONITOR_ONLY + - CAN_QUERY + - CAN_READ + - CAN_RESTART + - CAN_RUN + - CAN_USE + - CAN_VIEW + - CAN_VIEW_METADATA + - IS_OWNER + description: Permission level + RequestAuthzIdentity: + type: string + enum: + - REQUEST_AUTHZ_IDENTITY_SERVICE_IDENTITY + - REQUEST_AUTHZ_IDENTITY_USER_CONTEXT + description: >- + Defines the identity to be used for authZ of the request on the server + side. See one pager for + + for more information: http://go/acl/service-identity + ServicePrincipalSchema: + type: string + enum: + - urn:ietf:params:scim:schemas:core:2.0:ServicePrincipal + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + UserSchema: + type: string + enum: + - urn:ietf:params:scim:schemas:core:2.0:User + - urn:ietf:params:scim:schemas:extension:workspace:2.0:User + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + WorkspacePermission: + type: string + enum: + - ADMIN + - UNKNOWN + - USER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + access_control: + id: databricks_workspace.iam.access_control + name: access_control + title: Access Control + methods: + check: + operation: + $ref: '#/paths/~1api~12.0~1access-control~1check-policy-v2/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/access_control/methods/check' + insert: [] + update: [] + delete: [] + replace: [] + assignable_roles: + id: databricks_workspace.iam.assignable_roles + name: assignable_roles + title: Assignable Roles + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1accounts~1access-control~1assignable-roles/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/assignable_roles/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + rule_sets: + id: databricks_workspace.iam.rule_sets + name: rule_sets + title: Rule Sets + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1accounts~1access-control~1rule-sets/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1accounts~1access-control~1rule-sets/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/rule_sets/methods/get' + insert: [] + update: [] + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/rule_sets/methods/update' + current_user: + id: databricks_workspace.iam.current_user + name: current_user + title: Current User + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Me/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/current_user/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + groups_v2: + id: databricks_workspace.iam.groups_v2 + name: groups_v2 + title: Groups V2 + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.Resources + delete: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + patch: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Groups~1{id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/groups_v2/methods/get' + - $ref: '#/components/x-stackQL-resources/groups_v2/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/groups_v2/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/groups_v2/methods/patch' + delete: + - $ref: '#/components/x-stackQL-resources/groups_v2/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/groups_v2/methods/replace' + permission_migration: + id: databricks_workspace.iam.permission_migration + name: permission_migration + title: Permission Migration + methods: + migrate: + operation: + $ref: '#/paths/~1api~12.0~1permissionmigration/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: >- + #/components/x-stackQL-resources/permission_migration/methods/migrate + update: [] + delete: [] + replace: [] + permissions: + id: databricks_workspace.iam.permissions + name: permissions + title: Permissions + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/permissions/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/permissions/methods/update' + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/permissions/methods/set' + permission_levels: + id: databricks_workspace.iam.permission_levels + name: permission_levels + title: Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{request_object_type}~1{request_object_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/permission_levels/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + service_principals_v2: + id: databricks_workspace.iam.service_principals_v2 + name: service_principals_v2 + title: Service Principals V2 + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.Resources + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + patch: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1ServicePrincipals~1{id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/service_principals_v2/methods/get' + - $ref: >- + #/components/x-stackQL-resources/service_principals_v2/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/service_principals_v2/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/service_principals_v2/methods/patch + delete: + - $ref: >- + #/components/x-stackQL-resources/service_principals_v2/methods/delete + replace: + - $ref: >- + #/components/x-stackQL-resources/service_principals_v2/methods/replace + users_v2: + id: databricks_workspace.iam.users_v2 + name: users_v2 + title: Users V2 + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.Resources + delete: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + patch: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1preview~1scim~1v2~1Users~1{id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/users_v2/methods/get' + - $ref: '#/components/x-stackQL-resources/users_v2/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/users_v2/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/users_v2/methods/patch' + delete: + - $ref: '#/components/x-stackQL-resources/users_v2/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/users_v2/methods/replace' + password_permission_levels: + id: databricks_workspace.iam.password_permission_levels + name: password_permission_levels + title: Password Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1authorization~1passwords~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/password_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + password_permissions: + id: databricks_workspace.iam.password_permissions + name: password_permissions + title: Password Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1authorization~1passwords/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1authorization~1passwords/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1authorization~1passwords/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/password_permissions/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/password_permissions/methods/update + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/password_permissions/methods/set' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/iamv2.yaml b/providers/src/databricks_workspace/v00.00.00000/services/iamv2.yaml new file mode 100644 index 00000000..5706eeda --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/iamv2.yaml @@ -0,0 +1,435 @@ +openapi: 3.0.0 +info: + title: Databricks Iamv2 API (workspace) + description: >- + OpenAPI specification for the Databricks iamv2 service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.iamv2 +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/identity/workspaceAccessDetails/{principal_id}: + get: + operationId: workspace_iam_v2_get_workspace_access_detail_local + summary: >- + Returns the access details for a principal in the current workspace. + Allows for checking access + tags: + - iamv2 + - workspace_iam_v2 + description: >- + Returns the access details for a principal in the current workspace. + Allows for checking access + + details for any provisioned principal (user, service principal, or + group) in the current workspace. * + + Provisioned principal here refers to one that has been synced into + Databricks from the customer's IdP + + or added explicitly to Databricks via SCIM/UI. Allows for passing in a + "view" parameter to control + + what fields are returned (BASIC by default or FULL). + + + :param principal_id: int + Required. The internal ID of the principal (user/sp/group) for which the access details are being + requested. + :param view: :class:`WorkspaceAccessDetailView` (optional) + Controls what fields are returned. + + :returns: :class:`WorkspaceAccessDetail` + parameters: + - name: principal_id + in: path + required: true + schema: + type: integer + description: >- + Required. The internal ID of the principal (user/sp/group) for which + the access details are being requested. + - name: view + in: query + required: false + schema: + type: string + description: Controls what fields are returned. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspaceAccessDetail' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceIamV2API + /api/2.0/identity/groups/resolveByExternalId: + post: + operationId: workspace_iam_v2_resolve_group_proxy + summary: >- + Resolves a group with the given external ID from the customer's IdP. If + the group does not exist, it + tags: + - iamv2 + - workspace_iam_v2 + description: >- + Resolves a group with the given external ID from the customer's IdP. If + the group does not exist, it + + will be created in the account. If the customer is not onboarded onto + Automatic Identity Management + + (AIM), this will return an error. + + + :param external_id: str + Required. The external ID of the group in the customer's IdP. + + :returns: :class:`ResolveGroupResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + external_id: + type: string + description: >- + Required. The external ID of the group in the customer's + IdP. + required: + - external_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ResolveGroupResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceIamV2API + /api/2.0/identity/servicePrincipals/resolveByExternalId: + post: + operationId: workspace_iam_v2_resolve_service_principal_proxy + summary: >- + Resolves an SP with the given external ID from the customer's IdP. If + the SP does not exist, it will + tags: + - iamv2 + - workspace_iam_v2 + description: >- + Resolves an SP with the given external ID from the customer's IdP. If + the SP does not exist, it will + + be created. If the customer is not onboarded onto Automatic Identity + Management (AIM), this will + + return an error. + + + :param external_id: str + Required. The external ID of the service principal in the customer's IdP. + + :returns: :class:`ResolveServicePrincipalResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + external_id: + type: string + description: >- + Required. The external ID of the service principal in the + customer's IdP. + required: + - external_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ResolveServicePrincipalResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceIamV2API + /api/2.0/identity/users/resolveByExternalId: + post: + operationId: workspace_iam_v2_resolve_user_proxy + summary: >- + Resolves a user with the given external ID from the customer's IdP. If + the user does not exist, it + tags: + - iamv2 + - workspace_iam_v2 + description: >- + Resolves a user with the given external ID from the customer's IdP. If + the user does not exist, it + + will be created. If the customer is not onboarded onto Automatic + Identity Management (AIM), this will + + return an error. + + + :param external_id: str + Required. The external ID of the user in the customer's IdP. + + :returns: :class:`ResolveUserResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + external_id: + type: string + description: Required. The external ID of the user in the customer's IdP. + required: + - external_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ResolveUserResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceIamV2API +components: + schemas: + Group: + type: object + properties: + account_id: + type: string + description: The parent account ID for group in Databricks. + external_id: + type: string + description: ExternalId of the group in the customer's IdP. + group_name: + type: string + description: Display name of the group. + internal_id: + type: integer + description: Internal group ID of the group in Databricks. + description: The details of a Group resource. + ResolveGroupResponse: + type: object + properties: + group: + $ref: '#/components/schemas/Group' + ResolveServicePrincipalResponse: + type: object + properties: + service_principal: + $ref: '#/components/schemas/ServicePrincipal' + ResolveUserResponse: + type: object + properties: + user: + $ref: '#/components/schemas/User' + ServicePrincipal: + type: object + properties: + account_id: + type: string + description: The parent account ID for the service principal in Databricks. + account_sp_status: + $ref: '#/components/schemas/State' + description: The activity status of a service principal in a Databricks account. + application_id: + type: string + description: Application ID of the service principal. + display_name: + type: string + description: Display name of the service principal. + external_id: + type: string + description: ExternalId of the service principal in the customer's IdP. + internal_id: + type: integer + description: >- + Internal service principal ID of the service principal in + Databricks. + description: The details of a ServicePrincipal resource. + User: + type: object + properties: + account_id: + type: string + description: The accountId parent of the user in Databricks. + account_user_status: + $ref: '#/components/schemas/State' + description: The activity status of a user in a Databricks account. + external_id: + type: string + description: ExternalId of the user in the customer's IdP. + internal_id: + type: integer + description: Internal userId of the user in Databricks. + name: + $ref: '#/components/schemas/UserName' + username: + type: string + description: Username/email of the user. + description: The details of a User resource. + UserName: + type: object + properties: + family_name: + type: string + given_name: + type: string + WorkspaceAccessDetail: + type: object + properties: + access_type: + $ref: '#/components/schemas/WorkspaceAccessDetailAccessType' + account_id: + type: string + description: >- + The account ID parent of the workspace where the principal has + access. + permissions: + type: array + items: + $ref: '#/components/schemas/WorkspacePermission' + description: The permissions granted to the principal in the workspace. + principal_id: + type: integer + description: The internal ID of the principal (user/sp/group) in Databricks. + principal_type: + $ref: '#/components/schemas/PrincipalType' + status: + $ref: '#/components/schemas/State' + description: >- + The activity status of the principal in the workspace. Not + applicable for groups at the moment. + workspace_id: + type: integer + description: The workspace ID where the principal has access. + description: The details of a principal's access to a workspace. + PrincipalType: + type: string + enum: + - GROUP + - SERVICE_PRINCIPAL + - USER + description: The type of the principal (user/sp/group). + State: + type: string + enum: + - ACTIVE + - INACTIVE + description: >- + The activity status of a user or service principal in a Databricks + account or workspace. + WorkspaceAccessDetailAccessType: + type: string + enum: + - DIRECT + - INDIRECT + description: The type of access the principal has to the workspace. + WorkspaceAccessDetailView: + type: string + enum: + - BASIC + - FULL + description: >- + Controls what fields are returned in the GetWorkspaceAccessDetail + response. + WorkspacePermission: + type: string + enum: + - ADMIN_PERMISSION + - USER_PERMISSION + description: The type of permission a principal has to a workspace (admin/user). + x-stackQL-resources: + workspace_iam_v2: + id: databricks_workspace.iamv2.workspace_iam_v2 + name: workspace_iam_v2 + title: Workspace Iam V2 + methods: + get_workspace_access_detail_local: + operation: + $ref: >- + #/paths/~1api~12.0~1identity~1workspaceAccessDetails~1{principal_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + resolve_group_proxy: + operation: + $ref: '#/paths/~1api~12.0~1identity~1groups~1resolveByExternalId/post' + response: + mediaType: application/json + openAPIDocKey: '200' + resolve_service_principal_proxy: + operation: + $ref: >- + #/paths/~1api~12.0~1identity~1servicePrincipals~1resolveByExternalId/post + response: + mediaType: application/json + openAPIDocKey: '200' + resolve_user_proxy: + operation: + $ref: '#/paths/~1api~12.0~1identity~1users~1resolveByExternalId/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/workspace_iam_v2/methods/get_workspace_access_detail_local + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/jobs.yaml b/providers/src/databricks_workspace/v00.00.00000/services/jobs.yaml new file mode 100644 index 00000000..56bb6531 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/jobs.yaml @@ -0,0 +1,7024 @@ +openapi: 3.0.0 +info: + title: Databricks Jobs API (workspace) + description: >- + OpenAPI specification for the Databricks jobs service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.jobs +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.2/jobs/runs/cancel-all: + post: + operationId: jobs_cancel_all_runs + summary: >- + Cancels all active runs of a job. The runs are canceled asynchronously, + so it doesn't prevent new runs + tags: + - jobs + description: >- + Cancels all active runs of a job. The runs are canceled asynchronously, + so it doesn't prevent new runs + + from being started. + + + :param all_queued_runs: bool (optional) + Optional boolean parameter to cancel all queued runs. If no job_id is provided, all queued runs in + the workspace are canceled. + :param job_id: int (optional) + The canonical identifier of the job to cancel all runs of. + requestBody: + content: + application/json: + schema: + type: object + properties: + all_queued_runs: + type: string + description: >- + Optional boolean parameter to cancel all queued runs. If no + job_id is provided, all queued runs in the workspace are + canceled. + job_id: + type: string + description: The canonical identifier of the job to cancel all runs of. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/cancel: + post: + operationId: jobs_cancel_run + summary: >- + Cancels a job run or a task run. The run is canceled asynchronously, so + it may still be running when + tags: + - jobs + description: >- + Cancels a job run or a task run. The run is canceled asynchronously, so + it may still be running when + + this request completes. + + + :param run_id: int + This field is required. + + :returns: + Long-running operation waiter for :class:`Run`. + See :method:wait_get_run_job_terminated_or_skipped for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: integer + description: This field is required. + required: + - run_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Run' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/create: + post: + operationId: jobs_create + summary: Create a new job. + tags: + - jobs + description: >- + Create a new job. + + + :param access_control_list: List[:class:`JobAccessControlRequest`] + (optional) + List of permissions to set on the job. + :param budget_policy_id: str (optional) + The id of the user specified budget policy to use for this job. If not specified, a default budget + policy may be applied when creating or modifying the job. See `effective_budget_policy_id` for the + budget policy used by this workload. + :param continuous: :class:`Continuous` (optional) + An optional continuous property for this job. The continuous property will ensure that there is + always one run executing. Only one of `schedule` and `continuous` can be used. + :param deployment: :class:`JobDeployment` (optional) + Deployment information for jobs managed by external sources. + :param description: str (optional) + An optional description for the job. The maximum length is 27700 characters in UTF-8 encoding. + :param edit_mode: :class:`JobEditMode` (optional) + Edit mode of the job. + + * `UI_LOCKED`: The job is in a locked UI state and cannot be modified. * `EDITABLE`: The job is in + an editable state and can be modified. + :param email_notifications: :class:`JobEmailNotifications` (optional) + An optional set of email addresses that is notified when runs of this job begin or complete as well + as when this job is deleted. + :param environments: List[:class:`JobEnvironment`] (optional) + A list of task execution environment specifications that can be referenced by serverless tasks of + this job. For serverless notebook tasks, if the environment_key is not specified, the notebook + environment will be used if present. If a jobs environment is specified, it will override the + notebook environment. For other serverless tasks, the task environment is required to be specified + using environment_key in the task settings. + :param format: :class:`Format` (optional) + Used to tell what is the format of the job. This field is ignored in Create/Update/Reset calls. When + using the Jobs API 2.1 this value is always set to `"MULTI_TASK"`. + :param git_source: :class:`GitSource` (optional) + An optional specification for a remote Git repository containing the source code used by tasks. + Version-controlled source code is supported by notebook, dbt, Python script, and SQL File tasks. + + If `git_source` is set, these tasks retrieve the file from the remote repository by default. + However, this behavior can be overridden by setting `source` to `WORKSPACE` on the task. + + Note: dbt and SQL File tasks support only version-controlled sources. If dbt or SQL File tasks are + used, `git_source` must be defined on the job. + :param health: :class:`JobsHealthRules` (optional) + + :param job_clusters: List[:class:`JobCluster`] (optional) + A list of job cluster specifications that can be shared and reused by tasks of this job. Libraries + cannot be declared in a shared job cluster. You must declare dependent libraries in task settings. + :param max_concurrent_runs: int (optional) + An optional maximum allowed number of concurrent runs of the job. Set this value if you want to be + able to execute multiple runs of the same job concurrently. This is useful for example if you + trigger your job on a frequent schedule and want to allow consecutive runs to overlap with each + other, or if you want to trigger multiple runs which differ by their input parameters. This setting + affects only new runs. For example, suppose the job’s concurrency is 4 and there are 4 concurrent + active runs. Then setting the concurrency to 3 won’t kill any of the active runs. However, from + then on, new runs are skipped unless there are fewer than 3 active runs. This value cannot exceed + 1000. Setting this value to `0` causes all new runs to be skipped. + :param name: str (optional) + An optional name for the job. The maximum length is 4096 bytes in UTF-8 encoding. + :param notification_settings: :class:`JobNotificationSettings` + (optional) + Optional notification settings that are used when sending notifications to each of the + `email_notifications` and `webhook_notifications` for this job. + :param parameters: List[:class:`JobParameterDefinition`] (optional) + Job-level parameter definitions + :param performance_target: :class:`PerformanceTarget` (optional) + The performance mode on a serverless job. This field determines the level of compute performance or + cost-efficiency for the run. The performance target does not apply to tasks that run on Serverless + GPU compute. + + * `STANDARD`: Enables cost-efficient execution of serverless workloads. * `PERFORMANCE_OPTIMIZED`: + Prioritizes fast startup and execution times through rapid scaling and optimized cluster + performance. + :param queue: :class:`QueueSettings` (optional) + The queue settings of the job. + :param run_as: :class:`JobRunAs` (optional) + The user or service principal that the job runs as, if specified in the request. This field + indicates the explicit configuration of `run_as` for the job. To find the value in all cases, + explicit or implicit, use `run_as_user_name`. + :param schedule: :class:`CronSchedule` (optional) + An optional periodic schedule for this job. The default behavior is that the job only runs when + triggered by clicking “Run Now” in the Jobs UI or sending an API request to `runNow`. + :param tags: Dict[str,str] (optional) + A map of tags associated with the job. These are forwarded to the cluster as cluster tags for jobs + clusters, and are subject to the same limitations as cluster tags. A maximum of 25 tags can be added + to the job. + :param tasks: List[:class:`Task`] (optional) + A list of task specifications to be executed by this job. It supports up to 1000 elements in write + endpoints (:method:jobs/create, :method:jobs/reset, :method:jobs/update, :method:jobs/submit). Read + endpoints return only 100 tasks. If more than 100 tasks are available, you can paginate through them + using :method:jobs/get. Use the `next_page_token` field at the object root to determine if more + results are available. + :param timeout_seconds: int (optional) + An optional timeout applied to each run of this job. A value of `0` means no timeout. + :param trigger: :class:`TriggerSettings` (optional) + A configuration to trigger a run when certain conditions are met. The default behavior is that the + job runs only when triggered by clicking “Run Now” in the Jobs UI or sending an API request to + `runNow`. + :param usage_policy_id: str (optional) + The id of the user specified usage policy to use for this job. If not specified, a default usage + policy may be applied when creating or modifying the job. See `effective_usage_policy_id` for the + usage policy used by this workload. + :param webhook_notifications: :class:`WebhookNotifications` (optional) + A collection of system notification IDs to notify when runs of this job begin or complete. + + :returns: :class:`CreateResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: List of permissions to set on the job. + budget_policy_id: + type: string + description: >- + The id of the user specified budget policy to use for this + job. If not specified, a default budget policy may be + applied when creating or modifying the job. See + `effective_budget_policy_id` for the budget policy used by + this workload. + continuous: + type: string + description: >- + An optional continuous property for this job. The continuous + property will ensure that there is always one run executing. + Only one of `schedule` and `continuous` can be used. + deployment: + type: string + description: Deployment information for jobs managed by external sources. + description: + type: string + description: >- + An optional description for the job. The maximum length is + 27700 characters in UTF-8 encoding. + edit_mode: + type: string + description: >- + Edit mode of the job. * `UI_LOCKED`: The job is in a locked + UI state and cannot be modified. * `EDITABLE`: The job is in + an editable state and can be modified. + email_notifications: + type: string + description: >- + An optional set of email addresses that is notified when + runs of this job begin or complete as well as when this job + is deleted. + environments: + type: string + description: >- + A list of task execution environment specifications that can + be referenced by serverless tasks of this job. For + serverless notebook tasks, if the environment_key is not + specified, the notebook environment will be used if present. + If a jobs environment is specified, it will override the + notebook environment. For other serverless tasks, the task + environment is required to be specified using + environment_key in the task settings. + format: + type: string + description: >- + Used to tell what is the format of the job. This field is + ignored in Create/Update/Reset calls. When using the Jobs + API 2.1 this value is always set to `"MULTI_TASK"`. + git_source: + type: string + description: >- + An optional specification for a remote Git repository + containing the source code used by tasks. Version-controlled + source code is supported by notebook, dbt, Python script, + and SQL File tasks. If `git_source` is set, these tasks + retrieve the file from the remote repository by default. + However, this behavior can be overridden by setting `source` + to `WORKSPACE` on the task. Note: dbt and SQL File tasks + support only version-controlled sources. If dbt or SQL File + tasks are used, `git_source` must be defined on the job. + health: + type: string + description: >- + :param job_clusters: List[:class:`JobCluster`] (optional) A + list of job cluster specifications that can be shared and + reused by tasks of this job. Libraries cannot be declared in + a shared job cluster. You must declare dependent libraries + in task settings. + job_clusters: + type: string + max_concurrent_runs: + type: string + description: >- + An optional maximum allowed number of concurrent runs of the + job. Set this value if you want to be able to execute + multiple runs of the same job concurrently. This is useful + for example if you trigger your job on a frequent schedule + and want to allow consecutive runs to overlap with each + other, or if you want to trigger multiple runs which differ + by their input parameters. This setting affects only new + runs. For example, suppose the job’s concurrency is 4 and + there are 4 concurrent active runs. Then setting the + concurrency to 3 won’t kill any of the active runs. However, + from then on, new runs are skipped unless there are fewer + than 3 active runs. This value cannot exceed 1000. Setting + this value to `0` causes all new runs to be skipped. + name: + type: string + description: >- + An optional name for the job. The maximum length is 4096 + bytes in UTF-8 encoding. + notification_settings: + type: string + description: >- + Optional notification settings that are used when sending + notifications to each of the `email_notifications` and + `webhook_notifications` for this job. + parameters: + type: string + description: Job-level parameter definitions + performance_target: + type: string + description: >- + The performance mode on a serverless job. This field + determines the level of compute performance or + cost-efficiency for the run. The performance target does not + apply to tasks that run on Serverless GPU compute. * + `STANDARD`: Enables cost-efficient execution of serverless + workloads. * `PERFORMANCE_OPTIMIZED`: Prioritizes fast + startup and execution times through rapid scaling and + optimized cluster performance. + queue: + type: string + description: The queue settings of the job. + run_as: + type: string + description: >- + The user or service principal that the job runs as, if + specified in the request. This field indicates the explicit + configuration of `run_as` for the job. To find the value in + all cases, explicit or implicit, use `run_as_user_name`. + schedule: + type: string + description: >- + An optional periodic schedule for this job. The default + behavior is that the job only runs when triggered by + clicking “Run Now” in the Jobs UI or sending an API request + to `runNow`. + tags: + type: string + description: >- + A map of tags associated with the job. These are forwarded + to the cluster as cluster tags for jobs clusters, and are + subject to the same limitations as cluster tags. A maximum + of 25 tags can be added to the job. + tasks: + type: string + description: >- + A list of task specifications to be executed by this job. It + supports up to 1000 elements in write endpoints + (:method:jobs/create, :method:jobs/reset, + :method:jobs/update, :method:jobs/submit). Read endpoints + return only 100 tasks. If more than 100 tasks are available, + you can paginate through them using :method:jobs/get. Use + the `next_page_token` field at the object root to determine + if more results are available. + timeout_seconds: + type: string + description: >- + An optional timeout applied to each run of this job. A value + of `0` means no timeout. + trigger: + type: string + description: >- + A configuration to trigger a run when certain conditions are + met. The default behavior is that the job runs only when + triggered by clicking “Run Now” in the Jobs UI or sending an + API request to `runNow`. + usage_policy_id: + type: string + description: >- + The id of the user specified usage policy to use for this + job. If not specified, a default usage policy may be applied + when creating or modifying the job. See + `effective_usage_policy_id` for the usage policy used by + this workload. + webhook_notifications: + type: string + description: >- + A collection of system notification IDs to notify when runs + of this job begin or complete. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/delete: + post: + operationId: jobs_delete + summary: Deletes a job. + tags: + - jobs + description: |- + Deletes a job. + + :param job_id: int + The canonical identifier of the job to delete. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + job_id: + type: integer + description: >- + The canonical identifier of the job to delete. This field is + required. + required: + - job_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/delete: + post: + operationId: jobs_delete_run + summary: Deletes a non-active run. Returns an error if the run is active. + tags: + - jobs + description: |- + Deletes a non-active run. Returns an error if the run is active. + + :param run_id: int + ID of the run to delete. + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: integer + description: ID of the run to delete. + required: + - run_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/export: + get: + operationId: jobs_export_run + summary: Export and retrieve the job run task. + tags: + - jobs + description: |- + Export and retrieve the job run task. + + :param run_id: int + The canonical identifier for the run. This field is required. + :param views_to_export: :class:`ViewsToExport` (optional) + Which views to export (CODE, DASHBOARDS, or ALL). Defaults to CODE. + + :returns: :class:`ExportRunOutput` + parameters: + - name: run_id + in: query + required: true + schema: + type: integer + description: The canonical identifier for the run. This field is required. + - name: views_to_export + in: query + required: false + schema: + type: string + description: Which views to export (CODE, DASHBOARDS, or ALL). Defaults to CODE. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExportRunOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/get: + get: + operationId: jobs_get + summary: Retrieves the details for a single job. + tags: + - jobs + description: >- + Retrieves the details for a single job. + + + Large arrays in the results will be paginated when they exceed 100 + elements. A request for a single + + job will return all properties for that job, and the first 100 elements + of array properties (`tasks`, + + `job_clusters`, `environments` and `parameters`). Use the + `next_page_token` field to check for more + + results and pass its value as the `page_token` in subsequent requests. + If any array properties have + + more than 100 elements, additional results will be returned on + subsequent requests. Arrays without + + additional results will be empty on later pages. + + + :param job_id: int + The canonical identifier of the job to retrieve information about. This field is required. + :param page_token: str (optional) + Use `next_page_token` returned from the previous GetJob response to request the next page of the + job's array properties. + + :returns: :class:`Job` + parameters: + - name: job_id + in: query + required: true + schema: + type: integer + description: >- + The canonical identifier of the job to retrieve information about. + This field is required. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Use `next_page_token` returned from the previous GetJob response to + request the next page of the job's array properties. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Job' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.0/permissions/jobs/{job_id}/permissionLevels: + get: + operationId: jobs_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - jobs + description: |- + Gets the permission levels that a user can have on an object. + + :param job_id: str + The job for which to get or manage permissions. + + :returns: :class:`GetJobPermissionLevelsResponse` + parameters: + - name: job_id + in: path + required: true + schema: + type: string + description: The job for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetJobPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.0/permissions/jobs/{job_id}: + get: + operationId: jobs_get_permissions + summary: >- + Gets the permissions of a job. Jobs can inherit permissions from their + root object. + tags: + - jobs + description: >- + Gets the permissions of a job. Jobs can inherit permissions from their + root object. + + + :param job_id: str + The job for which to get or manage permissions. + + :returns: :class:`JobPermissions` + parameters: + - name: job_id + in: path + required: true + schema: + type: string + description: The job for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/JobPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + put: + operationId: jobs_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - jobs + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param job_id: str + The job for which to get or manage permissions. + :param access_control_list: List[:class:`JobAccessControlRequest`] + (optional) + + + :returns: :class:`JobPermissions` + parameters: + - name: job_id + in: path + required: true + schema: + type: string + description: The job for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`JobPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/JobPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + patch: + operationId: jobs_update_permissions + summary: >- + Updates the permissions on a job. Jobs can inherit permissions from + their root object. + tags: + - jobs + description: >- + Updates the permissions on a job. Jobs can inherit permissions from + their root object. + + + :param job_id: str + The job for which to get or manage permissions. + :param access_control_list: List[:class:`JobAccessControlRequest`] + (optional) + + + :returns: :class:`JobPermissions` + parameters: + - name: job_id + in: path + required: true + schema: + type: string + description: The job for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`JobPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/JobPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/get: + get: + operationId: jobs_get_run + summary: Retrieves the metadata of a run. + tags: + - jobs + description: >- + Retrieves the metadata of a run. + + + Large arrays in the results will be paginated when they exceed 100 + elements. A request for a single + + run will return all properties for that run, and the first 100 elements + of array properties (`tasks`, + + `job_clusters`, `job_parameters` and `repair_history`). Use the + next_page_token field to check for + + more results and pass its value as the page_token in subsequent + requests. If any array properties have + + more than 100 elements, additional results will be returned on + subsequent requests. Arrays without + + additional results will be empty on later pages. + + + :param run_id: int + The canonical identifier of the run for which to retrieve the metadata. This field is required. + :param include_history: bool (optional) + Whether to include the repair history in the response. + :param include_resolved_values: bool (optional) + Whether to include resolved parameter values in the response. + :param page_token: str (optional) + Use `next_page_token` returned from the previous GetRun response to request the next page of the + run's array properties. + + :returns: :class:`Run` + parameters: + - name: run_id + in: query + required: true + schema: + type: integer + description: >- + The canonical identifier of the run for which to retrieve the + metadata. This field is required. + - name: include_history + in: query + required: false + schema: + type: string + description: Whether to include the repair history in the response. + - name: include_resolved_values + in: query + required: false + schema: + type: string + description: Whether to include resolved parameter values in the response. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Use `next_page_token` returned from the previous GetRun response to + request the next page of the run's array properties. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Run' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/get-output: + get: + operationId: jobs_get_run_output + summary: >- + Retrieve the output and metadata of a single task run. When a notebook + task returns a value through + tags: + - jobs + description: >- + Retrieve the output and metadata of a single task run. When a notebook + task returns a value through + + the `dbutils.notebook.exit()` call, you can use this endpoint to + retrieve that value. Databricks + + restricts this API to returning the first 5 MB of the output. To return + a larger result, you can store + + job results in a cloud storage service. + + + This endpoint validates that the __run_id__ parameter is valid and + returns an HTTP status code 400 if + + the __run_id__ parameter is invalid. Runs are automatically removed + after 60 days. If you to want to + + reference them beyond 60 days, you must save old run results before they + expire. + + + :param run_id: int + The canonical identifier for the run. + + :returns: :class:`RunOutput` + parameters: + - name: run_id + in: query + required: true + schema: + type: integer + description: The canonical identifier for the run. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RunOutput' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/list: + get: + operationId: jobs_list + summary: Retrieves a list of jobs. + tags: + - jobs + description: |- + Retrieves a list of jobs. + + :param expand_tasks: bool (optional) + Whether to include task and cluster details in the response. Note that only the first 100 elements + will be shown. Use :method:jobs/get to paginate through all tasks and clusters. + :param limit: int (optional) + The number of jobs to return. This value must be greater than 0 and less or equal to 100. The + default value is 20. + :param name: str (optional) + A filter on the list based on the exact (case insensitive) job name. + :param offset: int (optional) + The offset of the first job to return, relative to the most recently created job. Deprecated since + June 2023. Use `page_token` to iterate through the pages instead. + :param page_token: str (optional) + Use `next_page_token` or `prev_page_token` returned from the previous request to list the next or + previous page of jobs respectively. + + :returns: Iterator over :class:`BaseJob` + parameters: + - name: expand_tasks + in: query + required: false + schema: + type: string + description: >- + Whether to include task and cluster details in the response. Note + that only the first 100 elements will be shown. Use :method:jobs/get + to paginate through all tasks and clusters. + - name: limit + in: query + required: false + schema: + type: string + description: >- + The number of jobs to return. This value must be greater than 0 and + less or equal to 100. The default value is 20. + - name: name + in: query + required: false + schema: + type: string + description: A filter on the list based on the exact (case insensitive) job name. + - name: offset + in: query + required: false + schema: + type: string + description: >- + The offset of the first job to return, relative to the most recently + created job. Deprecated since June 2023. Use `page_token` to iterate + through the pages instead. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Use `next_page_token` or `prev_page_token` returned from the + previous request to list the next or previous page of jobs + respectively. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListJobsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/list: + get: + operationId: jobs_list_runs + summary: List runs in descending order by start time. + tags: + - jobs + description: |- + List runs in descending order by start time. + + :param active_only: bool (optional) + If active_only is `true`, only active runs are included in the results; otherwise, lists both active + and completed runs. An active run is a run in the `QUEUED`, `PENDING`, `RUNNING`, or `TERMINATING`. + This field cannot be `true` when completed_only is `true`. + :param completed_only: bool (optional) + If completed_only is `true`, only completed runs are included in the results; otherwise, lists both + active and completed runs. This field cannot be `true` when active_only is `true`. + :param expand_tasks: bool (optional) + Whether to include task and cluster details in the response. Note that only the first 100 elements + will be shown. Use :method:jobs/getrun to paginate through all tasks and clusters. + :param job_id: int (optional) + The job for which to list runs. If omitted, the Jobs service lists runs from all jobs. + :param limit: int (optional) + The number of runs to return. This value must be greater than 0 and less than 25. The default value + is 20. If a request specifies a limit of 0, the service instead uses the maximum limit. + :param offset: int (optional) + The offset of the first run to return, relative to the most recent run. Deprecated since June 2023. + Use `page_token` to iterate through the pages instead. + :param page_token: str (optional) + Use `next_page_token` or `prev_page_token` returned from the previous request to list the next or + previous page of runs respectively. + :param run_type: :class:`RunType` (optional) + The type of runs to return. For a description of run types, see :method:jobs/getRun. + :param start_time_from: int (optional) + Show runs that started _at or after_ this value. The value must be a UTC timestamp in milliseconds. + Can be combined with _start_time_to_ to filter by a time range. + :param start_time_to: int (optional) + Show runs that started _at or before_ this value. The value must be a UTC timestamp in milliseconds. + Can be combined with _start_time_from_ to filter by a time range. + + :returns: Iterator over :class:`BaseRun` + parameters: + - name: active_only + in: query + required: false + schema: + type: string + description: >- + If active_only is `true`, only active runs are included in the + results; otherwise, lists both active and completed runs. An active + run is a run in the `QUEUED`, `PENDING`, `RUNNING`, or + `TERMINATING`. This field cannot be `true` when completed_only is + `true`. + - name: completed_only + in: query + required: false + schema: + type: string + description: >- + If completed_only is `true`, only completed runs are included in the + results; otherwise, lists both active and completed runs. This field + cannot be `true` when active_only is `true`. + - name: expand_tasks + in: query + required: false + schema: + type: string + description: >- + Whether to include task and cluster details in the response. Note + that only the first 100 elements will be shown. Use + :method:jobs/getrun to paginate through all tasks and clusters. + - name: job_id + in: query + required: false + schema: + type: string + description: >- + The job for which to list runs. If omitted, the Jobs service lists + runs from all jobs. + - name: limit + in: query + required: false + schema: + type: string + description: >- + The number of runs to return. This value must be greater than 0 and + less than 25. The default value is 20. If a request specifies a + limit of 0, the service instead uses the maximum limit. + - name: offset + in: query + required: false + schema: + type: string + description: >- + The offset of the first run to return, relative to the most recent + run. Deprecated since June 2023. Use `page_token` to iterate through + the pages instead. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Use `next_page_token` or `prev_page_token` returned from the + previous request to list the next or previous page of runs + respectively. + - name: run_type + in: query + required: false + schema: + type: string + description: >- + The type of runs to return. For a description of run types, see + :method:jobs/getRun. + - name: start_time_from + in: query + required: false + schema: + type: string + description: >- + Show runs that started _at or after_ this value. The value must be a + UTC timestamp in milliseconds. Can be combined with _start_time_to_ + to filter by a time range. + - name: start_time_to + in: query + required: false + schema: + type: string + description: >- + Show runs that started _at or before_ this value. The value must be + a UTC timestamp in milliseconds. Can be combined with + _start_time_from_ to filter by a time range. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListRunsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/repair: + post: + operationId: jobs_repair_run + summary: >- + Re-run one or more tasks. Tasks are re-run as part of the original job + run. They use the current job + tags: + - jobs + description: >- + Re-run one or more tasks. Tasks are re-run as part of the original job + run. They use the current job + + and task settings, and can be viewed in the history for the original job + run. + + + :param run_id: int + The job run ID of the run to repair. The run must not be in progress. + :param dbt_commands: List[str] (optional) + An array of commands to execute for jobs with the dbt task, for example `"dbt_commands": ["dbt + deps", "dbt seed", "dbt deps", "dbt seed", "dbt run"]` + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param jar_params: List[str] (optional) + A list of parameters for jobs with Spark JAR tasks, for example `"jar_params": ["john doe", "35"]`. + The parameters are used to invoke the main function of the main class specified in the Spark JAR + task. If not specified upon `run-now`, it defaults to an empty list. jar_params cannot be specified + in conjunction with notebook_params. The JSON representation of this field (for example + `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 bytes. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param job_parameters: Dict[str,str] (optional) + Job-level parameters used in the run. for example `"param": "overriding_val"` + :param latest_repair_id: int (optional) + The ID of the latest repair. This parameter is not required when repairing a run for the first time, + but must be provided on subsequent requests to repair the same run. + :param notebook_params: Dict[str,str] (optional) + A map from keys to values for jobs with notebook task, for example `"notebook_params": {"name": + "john doe", "age": "35"}`. The map is passed to the notebook and is accessible through the + [dbutils.widgets.get] function. + + If not specified upon `run-now`, the triggered run uses the job’s base parameters. + + notebook_params cannot be specified in conjunction with jar_params. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + The JSON representation of this field (for example `{"notebook_params":{"name":"john + doe","age":"35"}}`) cannot exceed 10,000 bytes. + + [dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param performance_target: :class:`PerformanceTarget` (optional) + The performance mode on a serverless job. The performance target determines the level of compute + performance or cost-efficiency for the run. This field overrides the performance target defined on + the job level. + + * `STANDARD`: Enables cost-efficient execution of serverless workloads. * `PERFORMANCE_OPTIMIZED`: + Prioritizes fast startup and execution times through rapid scaling and optimized cluster + performance. + :param pipeline_params: :class:`PipelineParams` (optional) + Controls whether the pipeline should perform a full refresh + :param python_named_params: Dict[str,str] (optional) + + :param python_params: List[str] (optional) + A list of parameters for jobs with Python tasks, for example `"python_params": ["john doe", "35"]`. + The parameters are passed to Python file as command-line parameters. If specified upon `run-now`, it + would overwrite the parameters specified in job setting. The JSON representation of this field (for + example `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + Important + + These parameters accept only Latin characters (ASCII character set). Using non-ASCII characters + returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese kanjis, and + emojis. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param rerun_all_failed_tasks: bool (optional) + If true, repair all failed tasks. Only one of `rerun_tasks` or `rerun_all_failed_tasks` can be used. + :param rerun_dependent_tasks: bool (optional) + If true, repair all tasks that depend on the tasks in `rerun_tasks`, even if they were previously + successful. Can be also used in combination with `rerun_all_failed_tasks`. + :param rerun_tasks: List[str] (optional) + The task keys of the task runs to repair. + :param spark_submit_params: List[str] (optional) + A list of parameters for jobs with spark submit task, for example `"spark_submit_params": + ["--class", "org.apache.spark.examples.SparkPi"]`. The parameters are passed to spark-submit script + as command-line parameters. If specified upon `run-now`, it would overwrite the parameters specified + in job setting. The JSON representation of this field (for example `{"python_params":["john + doe","35"]}`) cannot exceed 10,000 bytes. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + Important + + These parameters accept only Latin characters (ASCII character set). Using non-ASCII characters + returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese kanjis, and + emojis. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param sql_params: Dict[str,str] (optional) + A map from keys to values for jobs with SQL task, for example `"sql_params": {"name": "john doe", + "age": "35"}`. The SQL alert task does not support custom parameters. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + + :returns: + Long-running operation waiter for :class:`Run`. + See :method:wait_get_run_job_terminated_or_skipped for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: integer + description: >- + The job run ID of the run to repair. The run must not be in + progress. + dbt_commands: + type: string + description: >- + An array of commands to execute for jobs with the dbt task, + for example `"dbt_commands": ["dbt deps", "dbt seed", "dbt + deps", "dbt seed", "dbt run"]` ⚠ **Deprecation note** Use + [job parameters] to pass information down to tasks. [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + jar_params: + type: string + description: >- + A list of parameters for jobs with Spark JAR tasks, for + example `"jar_params": ["john doe", "35"]`. The parameters + are used to invoke the main function of the main class + specified in the Spark JAR task. If not specified upon + `run-now`, it defaults to an empty list. jar_params cannot + be specified in conjunction with notebook_params. The JSON + representation of this field (for example + `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 + bytes. ⚠ **Deprecation note** Use [job parameters] to pass + information down to tasks. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + job_parameters: + type: string + description: >- + Job-level parameters used in the run. for example `"param": + "overriding_val"` + latest_repair_id: + type: string + description: >- + The ID of the latest repair. This parameter is not required + when repairing a run for the first time, but must be + provided on subsequent requests to repair the same run. + notebook_params: + type: string + description: >- + A map from keys to values for jobs with notebook task, for + example `"notebook_params": {"name": "john doe", "age": + "35"}`. The map is passed to the notebook and is accessible + through the [dbutils.widgets.get] function. If not specified + upon `run-now`, the triggered run uses the job’s base + parameters. notebook_params cannot be specified in + conjunction with jar_params. ⚠ **Deprecation note** Use [job + parameters] to pass information down to tasks. The JSON + representation of this field (for example + `{"notebook_params":{"name":"john doe","age":"35"}}`) cannot + exceed 10,000 bytes. [dbutils.widgets.get]: + https://docs.databricks.com/dev-tools/databricks-utils.html + [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + performance_target: + type: string + description: >- + The performance mode on a serverless job. The performance + target determines the level of compute performance or + cost-efficiency for the run. This field overrides the + performance target defined on the job level. * `STANDARD`: + Enables cost-efficient execution of serverless workloads. * + `PERFORMANCE_OPTIMIZED`: Prioritizes fast startup and + execution times through rapid scaling and optimized cluster + performance. + pipeline_params: + type: string + description: Controls whether the pipeline should perform a full refresh + python_named_params: + type: string + description: >- + :param python_params: List[str] (optional) A list of + parameters for jobs with Python tasks, for example + `"python_params": ["john doe", "35"]`. The parameters are + passed to Python file as command-line parameters. If + specified upon `run-now`, it would overwrite the parameters + specified in job setting. The JSON representation of this + field (for example `{"python_params":["john doe","35"]}`) + cannot exceed 10,000 bytes. ⚠ **Deprecation note** Use [job + parameters] to pass information down to tasks. Important + These parameters accept only Latin characters (ASCII + character set). Using non-ASCII characters returns an error. + Examples of invalid, non-ASCII characters are Chinese, + Japanese kanjis, and emojis. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + python_params: + type: string + rerun_all_failed_tasks: + type: string + description: >- + If true, repair all failed tasks. Only one of `rerun_tasks` + or `rerun_all_failed_tasks` can be used. + rerun_dependent_tasks: + type: string + description: >- + If true, repair all tasks that depend on the tasks in + `rerun_tasks`, even if they were previously successful. Can + be also used in combination with `rerun_all_failed_tasks`. + rerun_tasks: + type: string + description: The task keys of the task runs to repair. + spark_submit_params: + type: string + description: >- + A list of parameters for jobs with spark submit task, for + example `"spark_submit_params": ["--class", + "org.apache.spark.examples.SparkPi"]`. The parameters are + passed to spark-submit script as command-line parameters. If + specified upon `run-now`, it would overwrite the parameters + specified in job setting. The JSON representation of this + field (for example `{"python_params":["john doe","35"]}`) + cannot exceed 10,000 bytes. ⚠ **Deprecation note** Use [job + parameters] to pass information down to tasks. Important + These parameters accept only Latin characters (ASCII + character set). Using non-ASCII characters returns an error. + Examples of invalid, non-ASCII characters are Chinese, + Japanese kanjis, and emojis. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + sql_params: + type: string + description: >- + A map from keys to values for jobs with SQL task, for + example `"sql_params": {"name": "john doe", "age": "35"}`. + The SQL alert task does not support custom parameters. ⚠ + **Deprecation note** Use [job parameters] to pass + information down to tasks. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + required: + - run_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Run' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/reset: + post: + operationId: jobs_reset + summary: >- + Overwrite all settings for the given job. Use the [_Update_ + endpoint](:method:jobs/update) to update + tags: + - jobs + description: >- + Overwrite all settings for the given job. Use the [_Update_ + endpoint](:method:jobs/update) to update + + job settings partially. + + + :param job_id: int + The canonical identifier of the job to reset. This field is required. + :param new_settings: :class:`JobSettings` + The new settings of the job. These settings completely replace the old settings. + + Changes to the field `JobBaseSettings.timeout_seconds` are applied to active runs. Changes to other + fields are applied to future runs only. + requestBody: + content: + application/json: + schema: + type: object + properties: + job_id: + type: integer + description: >- + The canonical identifier of the job to reset. This field is + required. + new_settings: + type: string + description: >- + The new settings of the job. These settings completely + replace the old settings. Changes to the field + `JobBaseSettings.timeout_seconds` are applied to active + runs. Changes to other fields are applied to future runs + only. + required: + - job_id + - new_settings + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/run-now: + post: + operationId: jobs_run_now + summary: Run a job and return the `run_id` of the triggered run. + tags: + - jobs + description: |- + Run a job and return the `run_id` of the triggered run. + + :param job_id: int + The ID of the job to be executed + :param dbt_commands: List[str] (optional) + An array of commands to execute for jobs with the dbt task, for example `"dbt_commands": ["dbt + deps", "dbt seed", "dbt deps", "dbt seed", "dbt run"]` + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param idempotency_token: str (optional) + An optional token to guarantee the idempotency of job run requests. If a run with the provided token + already exists, the request does not create a new run but returns the ID of the existing run + instead. If a run with the provided token is deleted, an error is returned. + + If you specify the idempotency token, upon failure you can retry until the request succeeds. + Databricks guarantees that exactly one run is launched with that idempotency token. + + This token must have at most 64 characters. + + For more information, see [How to ensure idempotency for jobs]. + + [How to ensure idempotency for jobs]: https://kb.databricks.com/jobs/jobs-idempotency.html + :param jar_params: List[str] (optional) + A list of parameters for jobs with Spark JAR tasks, for example `"jar_params": ["john doe", "35"]`. + The parameters are used to invoke the main function of the main class specified in the Spark JAR + task. If not specified upon `run-now`, it defaults to an empty list. jar_params cannot be specified + in conjunction with notebook_params. The JSON representation of this field (for example + `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 bytes. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param job_parameters: Dict[str,str] (optional) + Job-level parameters used in the run. for example `"param": "overriding_val"` + :param notebook_params: Dict[str,str] (optional) + A map from keys to values for jobs with notebook task, for example `"notebook_params": {"name": + "john doe", "age": "35"}`. The map is passed to the notebook and is accessible through the + [dbutils.widgets.get] function. + + If not specified upon `run-now`, the triggered run uses the job’s base parameters. + + notebook_params cannot be specified in conjunction with jar_params. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + The JSON representation of this field (for example `{"notebook_params":{"name":"john + doe","age":"35"}}`) cannot exceed 10,000 bytes. + + [dbutils.widgets.get]: https://docs.databricks.com/dev-tools/databricks-utils.html + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param only: List[str] (optional) + A list of task keys to run inside of the job. If this field is not provided, all tasks in the job + will be run. + :param performance_target: :class:`PerformanceTarget` (optional) + The performance mode on a serverless job. The performance target determines the level of compute + performance or cost-efficiency for the run. This field overrides the performance target defined on + the job level. + + * `STANDARD`: Enables cost-efficient execution of serverless workloads. * `PERFORMANCE_OPTIMIZED`: + Prioritizes fast startup and execution times through rapid scaling and optimized cluster + performance. + :param pipeline_params: :class:`PipelineParams` (optional) + Controls whether the pipeline should perform a full refresh + :param python_named_params: Dict[str,str] (optional) + :param python_params: List[str] (optional) + A list of parameters for jobs with Python tasks, for example `"python_params": ["john doe", "35"]`. + The parameters are passed to Python file as command-line parameters. If specified upon `run-now`, it + would overwrite the parameters specified in job setting. The JSON representation of this field (for + example `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + Important + + These parameters accept only Latin characters (ASCII character set). Using non-ASCII characters + returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese kanjis, and + emojis. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param queue: :class:`QueueSettings` (optional) + The queue settings of the run. + :param spark_submit_params: List[str] (optional) + A list of parameters for jobs with spark submit task, for example `"spark_submit_params": + ["--class", "org.apache.spark.examples.SparkPi"]`. The parameters are passed to spark-submit script + as command-line parameters. If specified upon `run-now`, it would overwrite the parameters specified + in job setting. The JSON representation of this field (for example `{"python_params":["john + doe","35"]}`) cannot exceed 10,000 bytes. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + Important + + These parameters accept only Latin characters (ASCII character set). Using non-ASCII characters + returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese kanjis, and + emojis. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + :param sql_params: Dict[str,str] (optional) + A map from keys to values for jobs with SQL task, for example `"sql_params": {"name": "john doe", + "age": "35"}`. The SQL alert task does not support custom parameters. + + ⚠ **Deprecation note** Use [job parameters] to pass information down to tasks. + + [job parameters]: https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + + :returns: + Long-running operation waiter for :class:`Run`. + See :method:wait_get_run_job_terminated_or_skipped for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + job_id: + type: integer + description: The ID of the job to be executed + dbt_commands: + type: string + description: >- + An array of commands to execute for jobs with the dbt task, + for example `"dbt_commands": ["dbt deps", "dbt seed", "dbt + deps", "dbt seed", "dbt run"]` ⚠ **Deprecation note** Use + [job parameters] to pass information down to tasks. [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + idempotency_token: + type: string + description: >- + An optional token to guarantee the idempotency of job run + requests. If a run with the provided token already exists, + the request does not create a new run but returns the ID of + the existing run instead. If a run with the provided token + is deleted, an error is returned. If you specify the + idempotency token, upon failure you can retry until the + request succeeds. Databricks guarantees that exactly one run + is launched with that idempotency token. This token must + have at most 64 characters. For more information, see [How + to ensure idempotency for jobs]. [How to ensure idempotency + for jobs]: + https://kb.databricks.com/jobs/jobs-idempotency.html + jar_params: + type: string + description: >- + A list of parameters for jobs with Spark JAR tasks, for + example `"jar_params": ["john doe", "35"]`. The parameters + are used to invoke the main function of the main class + specified in the Spark JAR task. If not specified upon + `run-now`, it defaults to an empty list. jar_params cannot + be specified in conjunction with notebook_params. The JSON + representation of this field (for example + `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 + bytes. ⚠ **Deprecation note** Use [job parameters] to pass + information down to tasks. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + job_parameters: + type: string + description: >- + Job-level parameters used in the run. for example `"param": + "overriding_val"` + notebook_params: + type: string + description: >- + A map from keys to values for jobs with notebook task, for + example `"notebook_params": {"name": "john doe", "age": + "35"}`. The map is passed to the notebook and is accessible + through the [dbutils.widgets.get] function. If not specified + upon `run-now`, the triggered run uses the job’s base + parameters. notebook_params cannot be specified in + conjunction with jar_params. ⚠ **Deprecation note** Use [job + parameters] to pass information down to tasks. The JSON + representation of this field (for example + `{"notebook_params":{"name":"john doe","age":"35"}}`) cannot + exceed 10,000 bytes. [dbutils.widgets.get]: + https://docs.databricks.com/dev-tools/databricks-utils.html + [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + only: + type: string + description: >- + A list of task keys to run inside of the job. If this field + is not provided, all tasks in the job will be run. + performance_target: + type: string + description: >- + The performance mode on a serverless job. The performance + target determines the level of compute performance or + cost-efficiency for the run. This field overrides the + performance target defined on the job level. * `STANDARD`: + Enables cost-efficient execution of serverless workloads. * + `PERFORMANCE_OPTIMIZED`: Prioritizes fast startup and + execution times through rapid scaling and optimized cluster + performance. + pipeline_params: + type: string + description: Controls whether the pipeline should perform a full refresh + python_named_params: + type: string + description: >- + :param python_params: List[str] (optional) A list of + parameters for jobs with Python tasks, for example + `"python_params": ["john doe", "35"]`. The parameters are + passed to Python file as command-line parameters. If + specified upon `run-now`, it would overwrite the parameters + specified in job setting. The JSON representation of this + field (for example `{"python_params":["john doe","35"]}`) + cannot exceed 10,000 bytes. ⚠ **Deprecation note** Use [job + parameters] to pass information down to tasks. Important + These parameters accept only Latin characters (ASCII + character set). Using non-ASCII characters returns an error. + Examples of invalid, non-ASCII characters are Chinese, + Japanese kanjis, and emojis. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + python_params: + type: string + queue: + type: string + description: The queue settings of the run. + spark_submit_params: + type: string + description: >- + A list of parameters for jobs with spark submit task, for + example `"spark_submit_params": ["--class", + "org.apache.spark.examples.SparkPi"]`. The parameters are + passed to spark-submit script as command-line parameters. If + specified upon `run-now`, it would overwrite the parameters + specified in job setting. The JSON representation of this + field (for example `{"python_params":["john doe","35"]}`) + cannot exceed 10,000 bytes. ⚠ **Deprecation note** Use [job + parameters] to pass information down to tasks. Important + These parameters accept only Latin characters (ASCII + character set). Using non-ASCII characters returns an error. + Examples of invalid, non-ASCII characters are Chinese, + Japanese kanjis, and emojis. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + sql_params: + type: string + description: >- + A map from keys to values for jobs with SQL task, for + example `"sql_params": {"name": "john doe", "age": "35"}`. + The SQL alert task does not support custom parameters. ⚠ + **Deprecation note** Use [job parameters] to pass + information down to tasks. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + required: + - job_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Run' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/runs/submit: + post: + operationId: jobs_submit + summary: >- + Submit a one-time run. This endpoint allows you to submit a workload + directly without creating a job. + tags: + - jobs + description: >- + Submit a one-time run. This endpoint allows you to submit a workload + directly without creating a job. + + Runs submitted using this endpoint don’t display in the UI. Use the + `jobs/runs/get` API to check the + + run state after the job is submitted. + + + **Important:** Jobs submitted using this endpoint are not saved as a + job. They do not show up in the + + Jobs UI, and do not retry when they fail. Because they are not saved, + Databricks cannot auto-optimize + + serverless compute in case of failure. If your job fails, you may want + to use classic compute to + + specify the compute needs for the job. Alternatively, use the `POST + /jobs/create` and `POST + + /jobs/run-now` endpoints to create and run a saved job. + + + :param access_control_list: List[:class:`JobAccessControlRequest`] + (optional) + List of permissions to set on the job. + :param budget_policy_id: str (optional) + The user specified id of the budget policy to use for this one-time run. If not specified, the run + will be not be attributed to any budget policy. + :param email_notifications: :class:`JobEmailNotifications` (optional) + An optional set of email addresses notified when the run begins or completes. + :param environments: List[:class:`JobEnvironment`] (optional) + A list of task execution environment specifications that can be referenced by tasks of this run. + :param git_source: :class:`GitSource` (optional) + An optional specification for a remote Git repository containing the source code used by tasks. + Version-controlled source code is supported by notebook, dbt, Python script, and SQL File tasks. + + If `git_source` is set, these tasks retrieve the file from the remote repository by default. + However, this behavior can be overridden by setting `source` to `WORKSPACE` on the task. + + Note: dbt and SQL File tasks support only version-controlled sources. If dbt or SQL File tasks are + used, `git_source` must be defined on the job. + :param health: :class:`JobsHealthRules` (optional) + + :param idempotency_token: str (optional) + An optional token that can be used to guarantee the idempotency of job run requests. If a run with + the provided token already exists, the request does not create a new run but returns the ID of the + existing run instead. If a run with the provided token is deleted, an error is returned. + + If you specify the idempotency token, upon failure you can retry until the request succeeds. + Databricks guarantees that exactly one run is launched with that idempotency token. + + This token must have at most 64 characters. + + For more information, see [How to ensure idempotency for jobs]. + + [How to ensure idempotency for jobs]: https://kb.databricks.com/jobs/jobs-idempotency.html + :param notification_settings: :class:`JobNotificationSettings` + (optional) + Optional notification settings that are used when sending notifications to each of the + `email_notifications` and `webhook_notifications` for this run. + :param queue: :class:`QueueSettings` (optional) + The queue settings of the one-time run. + :param run_as: :class:`JobRunAs` (optional) + Specifies the user or service principal that the job runs as. If not specified, the job runs as the + user who submits the request. + :param run_name: str (optional) + An optional name for the run. The default value is `Untitled`. + :param tasks: List[:class:`SubmitTask`] (optional) + + :param timeout_seconds: int (optional) + An optional timeout applied to each run of this job. A value of `0` means no timeout. + :param usage_policy_id: str (optional) + The user specified id of the usage policy to use for this one-time run. If not specified, a default + usage policy may be applied when creating or modifying the job. + :param webhook_notifications: :class:`WebhookNotifications` (optional) + A collection of system notification IDs to notify when the run begins or completes. + + :returns: + Long-running operation waiter for :class:`Run`. + See :method:wait_get_run_job_terminated_or_skipped for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: List of permissions to set on the job. + budget_policy_id: + type: string + description: >- + The user specified id of the budget policy to use for this + one-time run. If not specified, the run will be not be + attributed to any budget policy. + email_notifications: + type: string + description: >- + An optional set of email addresses notified when the run + begins or completes. + environments: + type: string + description: >- + A list of task execution environment specifications that can + be referenced by tasks of this run. + git_source: + type: string + description: >- + An optional specification for a remote Git repository + containing the source code used by tasks. Version-controlled + source code is supported by notebook, dbt, Python script, + and SQL File tasks. If `git_source` is set, these tasks + retrieve the file from the remote repository by default. + However, this behavior can be overridden by setting `source` + to `WORKSPACE` on the task. Note: dbt and SQL File tasks + support only version-controlled sources. If dbt or SQL File + tasks are used, `git_source` must be defined on the job. + health: + type: string + description: >- + :param idempotency_token: str (optional) An optional token + that can be used to guarantee the idempotency of job run + requests. If a run with the provided token already exists, + the request does not create a new run but returns the ID of + the existing run instead. If a run with the provided token + is deleted, an error is returned. If you specify the + idempotency token, upon failure you can retry until the + request succeeds. Databricks guarantees that exactly one run + is launched with that idempotency token. This token must + have at most 64 characters. For more information, see [How + to ensure idempotency for jobs]. [How to ensure idempotency + for jobs]: + https://kb.databricks.com/jobs/jobs-idempotency.html + idempotency_token: + type: string + notification_settings: + type: string + description: >- + Optional notification settings that are used when sending + notifications to each of the `email_notifications` and + `webhook_notifications` for this run. + queue: + type: string + description: The queue settings of the one-time run. + run_as: + type: string + description: >- + Specifies the user or service principal that the job runs + as. If not specified, the job runs as the user who submits + the request. + run_name: + type: string + description: >- + An optional name for the run. The default value is + `Untitled`. + tasks: + type: string + description: >- + :param timeout_seconds: int (optional) An optional timeout + applied to each run of this job. A value of `0` means no + timeout. + timeout_seconds: + type: string + usage_policy_id: + type: string + description: >- + The user specified id of the usage policy to use for this + one-time run. If not specified, a default usage policy may + be applied when creating or modifying the job. + webhook_notifications: + type: string + description: >- + A collection of system notification IDs to notify when the + run begins or completes. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Run' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.2/jobs/update: + post: + operationId: jobs_update + summary: >- + Add, update, or remove specific settings of an existing job. Use the + [_Reset_ + tags: + - jobs + description: >- + Add, update, or remove specific settings of an existing job. Use the + [_Reset_ + + endpoint](:method:jobs/reset) to overwrite all job settings. + + + :param job_id: int + The canonical identifier of the job to update. This field is required. + :param fields_to_remove: List[str] (optional) + Remove top-level fields in the job settings. Removing nested fields is not supported, except for + tasks and job clusters (`tasks/task_1`). This field is optional. + :param new_settings: :class:`JobSettings` (optional) + The new settings for the job. + + Top-level fields specified in `new_settings` are completely replaced, except for arrays which are + merged. That is, new and existing entries are completely replaced based on the respective key + fields, i.e. `task_key` or `job_cluster_key`, while previous entries are kept. + + Partially updating nested fields is not supported. + + Changes to the field `JobSettings.timeout_seconds` are applied to active runs. Changes to other + fields are applied to future runs only. + requestBody: + content: + application/json: + schema: + type: object + properties: + job_id: + type: integer + description: >- + The canonical identifier of the job to update. This field is + required. + fields_to_remove: + type: string + description: >- + Remove top-level fields in the job settings. Removing nested + fields is not supported, except for tasks and job clusters + (`tasks/task_1`). This field is optional. + new_settings: + type: string + description: >- + The new settings for the job. Top-level fields specified in + `new_settings` are completely replaced, except for arrays + which are merged. That is, new and existing entries are + completely replaced based on the respective key fields, i.e. + `task_key` or `job_cluster_key`, while previous entries are + kept. Partially updating nested fields is not supported. + Changes to the field `JobSettings.timeout_seconds` are + applied to active runs. Changes to other fields are applied + to future runs only. + required: + - job_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: JobsAPI + /api/2.0/policies/jobs/enforce-compliance: + post: + operationId: policy_compliance_for_jobs_enforce_compliance + summary: >- + Updates a job so the job clusters that are created when running the job + (specified in `new_cluster`) + tags: + - jobs + - policy_compliance_for_jobs + description: >- + Updates a job so the job clusters that are created when running the job + (specified in `new_cluster`) + + are compliant with the current versions of their respective cluster + policies. All-purpose clusters + + used in the job will not be updated. + + + :param job_id: int + The ID of the job you want to enforce policy compliance on. + :param validate_only: bool (optional) + If set, previews changes made to the job to comply with its policy, but does not update the job. + + :returns: :class:`EnforcePolicyComplianceResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + job_id: + type: integer + description: The ID of the job you want to enforce policy compliance on. + validate_only: + type: string + description: >- + If set, previews changes made to the job to comply with its + policy, but does not update the job. + required: + - job_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnforcePolicyComplianceResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyComplianceForJobsAPI + /api/2.0/policies/jobs/get-compliance: + get: + operationId: policy_compliance_for_jobs_get_compliance + summary: >- + Returns the policy compliance status of a job. Jobs could be out of + compliance if a cluster policy + tags: + - jobs + - policy_compliance_for_jobs + description: >- + Returns the policy compliance status of a job. Jobs could be out of + compliance if a cluster policy + + they use was updated after the job was last edited and some of its job + clusters no longer comply with + + their updated policies. + + + :param job_id: int + The ID of the job whose compliance status you are requesting. + + :returns: :class:`GetPolicyComplianceResponse` + parameters: + - name: job_id + in: query + required: true + schema: + type: integer + description: The ID of the job whose compliance status you are requesting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPolicyComplianceResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyComplianceForJobsAPI + /api/2.0/policies/jobs/list-compliance: + get: + operationId: policy_compliance_for_jobs_list_compliance + summary: >- + Returns the policy compliance status of all jobs that use a given + policy. Jobs could be out of + tags: + - jobs + - policy_compliance_for_jobs + description: >- + Returns the policy compliance status of all jobs that use a given + policy. Jobs could be out of + + compliance if a cluster policy they use was updated after the job was + last edited and its job clusters + + no longer comply with the updated policy. + + + :param policy_id: str + Canonical unique identifier for the cluster policy. + :param page_size: int (optional) + Use this field to specify the maximum number of results to be returned by the server. The server may + further constrain the maximum number of results returned in a single page. + :param page_token: str (optional) + A page token that can be used to navigate to the next page or previous page as returned by + `next_page_token` or `prev_page_token`. + + :returns: Iterator over :class:`JobCompliance` + parameters: + - name: policy_id + in: query + required: true + schema: + type: string + description: Canonical unique identifier for the cluster policy. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + Use this field to specify the maximum number of results to be + returned by the server. The server may further constrain the maximum + number of results returned in a single page. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token that can be used to navigate to the next page or + previous page as returned by `next_page_token` or `prev_page_token`. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListJobComplianceForPolicyResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PolicyComplianceForJobsAPI +components: + schemas: + BaseJob: + type: object + properties: + created_time: + type: integer + creator_user_name: + type: string + description: >- + The creator user name. This field won’t be included in the response + if the user has already been deleted. + effective_budget_policy_id: + type: string + description: >- + The id of the budget policy used by this job for cost attribution + purposes. This may be set through (in order of precedence): 1. + Budget admins through the account or workspace console 2. Jobs UI in + the job details page and Jobs API using `budget_policy_id` 3. + Inferred default based on accessible budget policies of the run_as + identity on job creation or modification. + effective_usage_policy_id: + type: string + description: >- + The id of the usage policy used by this job for cost attribution + purposes. + has_more: + type: boolean + description: >- + Indicates if the job has more array properties (`tasks`, + `job_clusters`) that are not shown. They can be accessed via + :method:jobs/get endpoint. It is only relevant for API 2.2 + :method:jobs/list requests with `expand_tasks=true`. + job_id: + type: integer + description: The canonical identifier for this job. + settings: + $ref: '#/components/schemas/JobSettings' + description: >- + Settings for this job and all of its runs. These settings can be + updated using the `resetJob` method. + trigger_state: + $ref: '#/components/schemas/TriggerStateProto' + description: State of the trigger associated with the job. + BaseRun: + type: object + properties: + attempt_number: + type: integer + cleanup_duration: + type: integer + description: >- + The time in milliseconds it took to terminate the cluster and clean + up any associated artifacts. The duration of a task run is the sum + of the `setup_duration`, `execution_duration`, and the + `cleanup_duration`. The `cleanup_duration` field is set to 0 for + multitask job runs. The total duration of a multitask job run is the + value of the `run_duration` field. + cluster_instance: + $ref: '#/components/schemas/ClusterInstance' + description: >- + The cluster used for this run. If the run is specified to use a new + cluster, this field is set once the Jobs service has requested a + cluster for the run. + cluster_spec: + $ref: '#/components/schemas/ClusterSpec' + description: >- + A snapshot of the job’s cluster specification when this run was + created. + creator_user_name: + type: string + description: >- + The creator user name. This field won’t be included in the response + if the user has already been deleted. + description: + type: string + description: Description of the run + effective_performance_target: + $ref: '#/components/schemas/PerformanceTarget' + description: >- + The actual performance target used by the serverless run during + execution. This can differ from the client-set performance target on + the request depending on whether the performance mode is supported + by the job type. * `STANDARD`: Enables cost-efficient execution of + serverless workloads. * `PERFORMANCE_OPTIMIZED`: Prioritizes fast + startup and execution times through rapid scaling and optimized + cluster performance. + effective_usage_policy_id: + type: string + description: >- + The id of the usage policy used by this run for cost attribution + purposes. + end_time: + type: integer + description: >- + The time at which this run ended in epoch milliseconds (milliseconds + since 1/1/1970 UTC). This field is set to 0 if the job is still + running. + execution_duration: + type: integer + description: >- + The time in milliseconds it took to execute the commands in the JAR + or notebook until they completed, failed, timed out, were cancelled, + or encountered an unexpected error. The duration of a task run is + the sum of the `setup_duration`, `execution_duration`, and the + `cleanup_duration`. The `execution_duration` field is set to 0 for + multitask job runs. The total duration of a multitask job run is the + value of the `run_duration` field. + git_source: + $ref: '#/components/schemas/GitSource' + description: >- + An optional specification for a remote Git repository containing the + source code used by tasks. Version-controlled source code is + supported by notebook, dbt, Python script, and SQL File tasks. If + `git_source` is set, these tasks retrieve the file from the remote + repository by default. However, this behavior can be overridden by + setting `source` to `WORKSPACE` on the task. Note: dbt and SQL File + tasks support only version-controlled sources. If dbt or SQL File + tasks are used, `git_source` must be defined on the job. + has_more: + type: boolean + description: >- + Indicates if the run has more array properties (`tasks`, + `job_clusters`) that are not shown. They can be accessed via + :method:jobs/getrun endpoint. It is only relevant for API 2.2 + :method:jobs/listruns requests with `expand_tasks=true`. + job_clusters: + type: array + items: + $ref: '#/components/schemas/JobCluster' + description: >- + A list of job cluster specifications that can be shared and reused + by tasks of this job. Libraries cannot be declared in a shared job + cluster. You must declare dependent libraries in task settings. If + more than 100 job clusters are available, you can paginate through + them using :method:jobs/getrun. + job_id: + type: integer + description: The canonical identifier of the job that contains this run. + job_parameters: + type: array + items: + $ref: '#/components/schemas/JobParameter' + description: Job-level parameters used in the run + job_run_id: + type: integer + description: >- + ID of the job run that this run belongs to. For legacy and + single-task job runs the field is populated with the job run ID. For + task runs, the field is populated with the ID of the job run that + the task run belongs to. + number_in_job: + type: integer + description: >- + A unique identifier for this job run. This is set to the same value + as `run_id`. + original_attempt_run_id: + type: integer + description: >- + If this run is a retry of a prior run attempt, this field contains + the run_id of the original attempt; otherwise, it is the same as the + run_id. + overriding_parameters: + $ref: '#/components/schemas/RunParameters' + description: The parameters used for this run. + queue_duration: + type: integer + description: The time in milliseconds that the run has spent in the queue. + repair_history: + type: array + items: + $ref: '#/components/schemas/RepairHistoryItem' + description: The repair history of the run. + run_duration: + type: integer + description: >- + The time in milliseconds it took the job run and all of its repairs + to finish. + run_id: + type: integer + description: >- + The canonical identifier of the run. This ID is unique across all + runs of all jobs. + run_name: + type: string + description: >- + An optional name for the run. The maximum length is 4096 bytes in + UTF-8 encoding. + run_page_url: + type: string + description: The URL to the detail page of the run. + run_type: + $ref: '#/components/schemas/RunType' + schedule: + $ref: '#/components/schemas/CronSchedule' + description: >- + The cron schedule that triggered this run if it was triggered by the + periodic scheduler. + setup_duration: + type: integer + description: >- + The time in milliseconds it took to set up the cluster. For runs + that run on new clusters this is the cluster creation time, for runs + that run on existing clusters this time should be very short. The + duration of a task run is the sum of the `setup_duration`, + `execution_duration`, and the `cleanup_duration`. The + `setup_duration` field is set to 0 for multitask job runs. The total + duration of a multitask job run is the value of the `run_duration` + field. + start_time: + type: integer + description: >- + The time at which this run was started in epoch milliseconds + (milliseconds since 1/1/1970 UTC). This may not be the time when the + job task starts executing, for example, if the job is scheduled to + run on a new cluster, this is the time the cluster creation call is + issued. + state: + $ref: '#/components/schemas/RunState' + description: Deprecated. Please use the `status` field instead. + status: + $ref: '#/components/schemas/RunStatus' + tasks: + type: array + items: + $ref: '#/components/schemas/RunTask' + description: >- + The list of tasks performed by the run. Each task has its own + `run_id` which you can use to call `JobsGetOutput` to retrieve the + run resutls. If more than 100 tasks are available, you can paginate + through them using :method:jobs/getrun. Use the `next_page_token` + field at the object root to determine if more results are available. + trigger: + $ref: '#/components/schemas/TriggerType' + trigger_info: + $ref: '#/components/schemas/TriggerInfo' + CleanRoomTaskRunState: + type: object + properties: + life_cycle_state: + $ref: '#/components/schemas/CleanRoomTaskRunLifeCycleState' + description: >- + A value indicating the run's current lifecycle state. This field is + always available in the response. Note: Additional states might be + introduced in future releases. + result_state: + $ref: '#/components/schemas/CleanRoomTaskRunResultState' + description: >- + A value indicating the run's result. This field is only available + for terminal lifecycle states. Note: Additional states might be + introduced in future releases. + description: Stores the run state of the clean rooms notebook task. + CleanRoomsNotebookTask: + type: object + properties: + clean_room_name: + type: string + description: The clean room that the notebook belongs to. + notebook_name: + type: string + description: Name of the notebook being run. + etag: + type: string + description: >- + Checksum to validate the freshness of the notebook resource (i.e. + the notebook being run is the latest version). It can be fetched by + calling the :method:cleanroomassets/get API. + notebook_base_parameters: + type: object + description: Base parameters to be used for the clean room notebook job. + required: + - clean_room_name + - notebook_name + description: >- + Clean Rooms notebook task for V1 Clean Room service (GA). Replaces the + deprecated + CleanRoomNotebookTask (defined above) which was for V0 service. + CleanRoomsNotebookTaskCleanRoomsNotebookTaskOutput: + type: object + properties: + clean_room_job_run_state: + $ref: '#/components/schemas/CleanRoomTaskRunState' + notebook_output: + $ref: '#/components/schemas/NotebookOutput' + description: The notebook output for the clean room run + output_schema_info: + $ref: '#/components/schemas/OutputSchemaInfo' + description: >- + Information on how to access the output schema for the clean room + run + ClusterInstance: + type: object + properties: + cluster_id: + type: string + spark_context_id: + type: string + description: >- + The canonical identifier for the Spark context used by a run. This + field is filled in once the run begins execution. This value can be + used to view the Spark UI by browsing to + `/#setting/sparkui/$cluster_id/$spark_context_id`. The Spark UI + continues to be available after the run has completed. The response + won’t include this field if the identifier is not available yet. + ClusterSpec: + type: object + properties: + existing_cluster_id: + type: string + job_cluster_key: + type: string + description: >- + If job_cluster_key, this task is executed reusing the cluster + specified in `job.settings.job_clusters`. + libraries: + type: string + description: >- + An optional list of libraries to be installed on the cluster. The + default value is an empty list. + new_cluster: + type: string + description: >- + If new_cluster, a description of a new cluster that is created for + each run. + Compute: + type: object + properties: + hardware_accelerator: + type: string + ComputeConfig: + type: object + properties: + num_gpus: + type: integer + gpu_node_pool_id: + type: string + description: IDof the GPU pool to use. + gpu_type: + type: string + description: GPU type. + required: + - num_gpus + ConditionTask: + type: object + properties: + op: + $ref: '#/components/schemas/ConditionTaskOp' + left: + type: string + description: >- + The left operand of the condition task. Can be either a string value + or a job state or parameter reference. + right: + type: string + description: >- + The right operand of the condition task. Can be either a string + value or a job state or parameter reference. + required: + - op + - left + - right + Continuous: + type: object + properties: + pause_status: + $ref: '#/components/schemas/PauseStatus' + task_retry_mode: + $ref: '#/components/schemas/TaskRetryMode' + description: >- + Indicate whether the continuous job is applying task level retries + or not. Defaults to NEVER. + CreateResponse: + type: object + properties: + job_id: + type: integer + description: The canonical identifier for the newly created job. + description: Job was created successfully + CronSchedule: + type: object + properties: + quartz_cron_expression: + type: string + timezone_id: + type: string + description: >- + A Java timezone ID. The schedule for a job is resolved with respect + to this timezone. See [Java TimeZone] for details. This field is + required. [Java TimeZone]: + https://docs.oracle.com/javase/7/docs/api/java/util/TimeZone.html + pause_status: + $ref: '#/components/schemas/PauseStatus' + description: Indicate whether this schedule is paused or not. + required: + - quartz_cron_expression + - timezone_id + DashboardPageSnapshot: + type: object + properties: + page_display_name: + type: string + widget_error_details: + type: array + items: + $ref: '#/components/schemas/WidgetErrorDetail' + DashboardTask: + type: object + properties: + dashboard_id: + type: string + description: The identifier of the dashboard to refresh. + filters: + type: object + description: >- + Dashboard task parameters. Used to apply dashboard filter values + during dashboard task execution. Parameter values get applied to any + dashboard filters that have a matching URL identifier as the + parameter key. The parameter value format is dependent on the filter + type: - For text and single-select filters, provide a single value + (e.g. `"value"`) - For date and datetime filters, provide the value + in ISO 8601 format (e.g. `"2000-01-01T00:00:00"`) - For multi-select + filters, provide a JSON array of values (e.g. + `"[\"value1\",\"value2\"]"`) - For range and date range filters, + provide a JSON object with `start` and `end` (e.g. + `"{\"start\":\"1\",\"end\":\"10\"}"`) + subscription: + $ref: '#/components/schemas/Subscription' + description: >- + Optional: subscription configuration for sending the dashboard + snapshot. + warehouse_id: + type: string + description: >- + Optional: The warehouse id to execute the dashboard with for the + schedule. If not specified, the default warehouse of the dashboard + will be used. + description: Configures the Lakeview Dashboard job task type. + DashboardTaskOutput: + type: object + properties: + page_snapshots: + type: array + items: + $ref: '#/components/schemas/DashboardPageSnapshot' + DbtCloudJobRunStep: + type: object + properties: + index: + type: integer + description: Orders the steps in the job + logs: + type: string + description: Output of the step + name: + type: string + description: Name of the step in the job + status: + $ref: '#/components/schemas/DbtPlatformRunStatus' + description: State of the step + description: >- + Format of response retrieved from dbt Cloud, for inclusion in output + Deprecated in favor of + DbtPlatformJobRunStep + DbtCloudTask: + type: object + properties: + connection_resource_name: + type: string + description: >- + The resource name of the UC connection that authenticates the dbt + Cloud for this task + dbt_cloud_job_id: + type: integer + description: Id of the dbt Cloud job to be triggered + description: Deprecated in favor of DbtPlatformTask + DbtCloudTaskOutput: + type: object + properties: + dbt_cloud_job_run_id: + type: integer + description: Id of the job run in dbt Cloud + dbt_cloud_job_run_output: + type: array + items: + $ref: '#/components/schemas/DbtCloudJobRunStep' + description: Steps of the job run as received from dbt Cloud + dbt_cloud_job_run_url: + type: string + description: Url where full run details can be viewed + description: Deprecated in favor of DbtPlatformTaskOutput + DbtOutput: + type: object + properties: + artifacts_headers: + type: object + artifacts_link: + type: string + description: >- + A pre-signed URL to download the (compressed) dbt artifacts. This + link is valid for a limited time (30 minutes). This information is + only available after the run has finished. + DbtPlatformJobRunStep: + type: object + properties: + index: + type: integer + description: Orders the steps in the job + logs: + type: string + description: Output of the step + logs_truncated: + type: boolean + description: >- + Whether the logs of this step have been truncated. If true, the logs + has been truncated to 10000 characters. + name: + type: string + description: Name of the step in the job + name_truncated: + type: boolean + description: >- + Whether the name of the job has been truncated. If true, the name + has been truncated to 100 characters. + status: + $ref: '#/components/schemas/DbtPlatformRunStatus' + description: State of the step + description: Format of response retrieved from dbt platform, for inclusion in output + DbtPlatformTask: + type: object + properties: + connection_resource_name: + type: string + dbt_platform_job_id: + type: string + description: >- + Id of the dbt platform job to be triggered. Specified as a string + for maximum compatibility with clients. + DbtPlatformTaskOutput: + type: object + properties: + dbt_platform_job_run_id: + type: string + dbt_platform_job_run_output: + type: array + items: + $ref: '#/components/schemas/DbtPlatformJobRunStep' + description: Steps of the job run as received from dbt platform + dbt_platform_job_run_url: + type: string + description: Url where full run details can be viewed + steps_truncated: + type: boolean + description: >- + Whether the number of steps in the output has been truncated. If + true, the output will contain the first 20 steps of the output. + DbtTask: + type: object + properties: + commands: + type: array + items: + type: string + catalog: + type: string + description: >- + Optional name of the catalog to use. The value is the top level in + the 3-level namespace of Unity Catalog (catalog / schema / + relation). The catalog value can only be specified if a warehouse_id + is specified. Requires dbt-databricks >= 1.1.1. + profiles_directory: + type: string + description: >- + Optional (relative) path to the profiles directory. Can only be + specified if no warehouse_id is specified. If no warehouse_id is + specified and this folder is unset, the root directory is used. + project_directory: + type: string + description: >- + Path to the project directory. Optional for Git sourced tasks, in + which case if no value is provided, the root of the Git repository + is used. + schema: + type: string + description: >- + Optional schema to write to. This parameter is only used when a + warehouse_id is also provided. If not provided, the `default` schema + is used. + source: + $ref: '#/components/schemas/Source' + description: >- + Optional location type of the project directory. When set to + `WORKSPACE`, the project will be retrieved from the local Databricks + workspace. When set to `GIT`, the project will be retrieved from a + Git repository defined in `git_source`. If the value is empty, the + task will use `GIT` if `git_source` is defined and `WORKSPACE` + otherwise. * `WORKSPACE`: Project is located in Databricks + workspace. * `GIT`: Project is located in cloud Git provider. + warehouse_id: + type: string + description: >- + ID of the SQL warehouse to connect to. If provided, we automatically + generate and provide the profile and connection details to dbt. It + can be overridden on a per-command basis by using the + `--profiles-dir` command line argument. + required: + - commands + EnforcePolicyComplianceForJobResponseJobClusterSettingsChange: + type: object + properties: + field: + type: string + description: >- + The field where this change would be made, prepended with the job + cluster key. + new_value: + type: string + description: >- + The new value of this field after enforcing policy compliance + (either a number, a boolean, or a string) converted to a string. + This is intended to be read by a human. The typed new value of this + field can be retrieved by reading the settings field in the API + response. + previous_value: + type: string + description: >- + The previous value of this field before enforcing policy compliance + (either a number, a boolean, or a string) converted to a string. + This is intended to be read by a human. The type of the field can be + retrieved by reading the settings field in the API response. + description: >- + Represents a change to the job cluster's settings that would be required + for the job clusters to + become compliant with their policies. + EnforcePolicyComplianceResponse: + type: object + properties: + has_changes: + type: boolean + job_cluster_changes: + type: array + items: + $ref: >- + #/components/schemas/EnforcePolicyComplianceForJobResponseJobClusterSettingsChange + description: >- + A list of job cluster changes that have been made to the job’s + cluster settings in order for all job clusters to become compliant + with their policies. + settings: + $ref: '#/components/schemas/JobSettings' + description: >- + Updated job settings after policy enforcement. Policy enforcement + only applies to job clusters that are created when running the job + (which are specified in new_cluster) and does not apply to existing + all-purpose clusters. Updated job settings are derived by applying + policy default values to the existing job clusters in order to + satisfy policy requirements. + ExportRunOutput: + type: object + properties: + views: + type: array + items: + $ref: '#/components/schemas/ViewItem' + description: >- + The exported content in HTML format (one for every view item). To + extract the HTML notebook from the JSON response, download and run + this [Python script](/_static/examples/extract.py). + description: Run was exported successfully. + FileArrivalTriggerConfiguration: + type: object + properties: + url: + type: string + min_time_between_triggers_seconds: + type: integer + description: >- + If set, the trigger starts a run only after the specified amount of + time passed since the last time the trigger fired. The minimum + allowed value is 60 seconds + wait_after_last_change_seconds: + type: integer + description: >- + If set, the trigger starts a run only after no file activity has + occurred for the specified amount of time. This makes it possible to + wait for a batch of incoming files to arrive before triggering a + run. The minimum allowed value is 60 seconds. + required: + - url + FileArrivalTriggerState: + type: object + properties: + using_file_events: + type: boolean + ForEachStats: + type: object + properties: + error_message_stats: + type: array + items: + $ref: '#/components/schemas/ForEachTaskErrorMessageStats' + task_run_stats: + $ref: '#/components/schemas/ForEachTaskTaskRunStats' + description: >- + Describes stats of the iteration. Only latest retries are + considered. + ForEachTask: + type: object + properties: + inputs: + type: string + task: + $ref: '#/components/schemas/Task' + description: >- + Configuration for the task that will be run for each element in the + array + concurrency: + type: integer + description: >- + An optional maximum allowed number of concurrent runs of the task. + Set this value if you want to be able to execute multiple runs of + the task concurrently. + required: + - inputs + - task + ForEachTaskErrorMessageStats: + type: object + properties: + count: + type: integer + error_message: + type: string + description: Describes the error message occured during the iterations. + termination_category: + type: string + description: Describes the termination reason for the error message. + ForEachTaskTaskRunStats: + type: object + properties: + active_iterations: + type: integer + completed_iterations: + type: integer + description: Describes the number of failed and succeeded iteration runs. + failed_iterations: + type: integer + description: Describes the number of failed iteration runs. + scheduled_iterations: + type: integer + description: Describes the number of iteration runs that have been scheduled. + succeeded_iterations: + type: integer + description: Describes the number of succeeded iteration runs. + total_iterations: + type: integer + description: Describes the length of the list of items to iterate over. + GenAiComputeTask: + type: object + properties: + dl_runtime_image: + type: string + command: + type: string + description: Command launcher to run the actual script, e.g. bash, python etc. + compute: + $ref: '#/components/schemas/ComputeConfig' + mlflow_experiment_name: + type: string + description: >- + Optional string containing the name of the MLflow experiment to log + the run to. If name is not found, backend will create the mlflow + experiment using the name. + source: + $ref: '#/components/schemas/Source' + description: >- + Optional location type of the training script. When set to + `WORKSPACE`, the script will be retrieved from the local Databricks + workspace. When set to `GIT`, the script will be retrieved from a + Git repository defined in `git_source`. If the value is empty, the + task will use `GIT` if `git_source` is defined and `WORKSPACE` + otherwise. * `WORKSPACE`: Script is located in Databricks workspace. + * `GIT`: Script is located in cloud Git provider. + training_script_path: + type: string + description: >- + The training script file path to be executed. Cloud file URIs (such + as dbfs:/, s3:/, adls:/, gcs:/) and workspace paths are supported. + For python files stored in the Databricks workspace, the path must + be absolute and begin with `/`. For files stored in a remote + repository, the path must be relative. This field is required. + yaml_parameters: + type: string + description: >- + Optional string containing model parameters passed to the training + script in yaml format. If present, then the content in + yaml_parameters_file_path will be ignored. + yaml_parameters_file_path: + type: string + description: >- + Optional path to a YAML file containing model parameters passed to + the training script. + required: + - dl_runtime_image + GetJobPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/JobPermissionsDescription' + GetPolicyComplianceResponse: + type: object + properties: + is_compliant: + type: boolean + violations: + type: object + description: >- + An object containing key-value mappings representing the first 200 + policy validation errors. The keys indicate the path where the + policy validation error is occurring. An identifier for the job + cluster is prepended to the path. The values indicate an error + message describing the policy validation error. + GitSnapshot: + type: object + properties: + used_commit: + type: string + description: >- + Commit that was used to execute the run. If git_branch was + specified, this points to the HEAD of the branch at the time of the + run; if git_tag was specified, this points to the commit the tag + points to. + description: >- + Read-only state of the remote repository at the time the job was run. + This field is only + included on job runs. + GitSource: + type: object + properties: + git_url: + type: string + description: URL of the repository to be cloned by this job. + git_provider: + $ref: '#/components/schemas/GitProvider' + description: >- + Unique identifier of the service used to host the Git repository. + The value is case insensitive. + git_branch: + type: string + description: >- + Name of the branch to be checked out and used by this job. This + field cannot be specified in conjunction with git_tag or git_commit. + git_commit: + type: string + description: >- + Commit to be checked out and used by this job. This field cannot be + specified in conjunction with git_branch or git_tag. + git_snapshot: + $ref: '#/components/schemas/GitSnapshot' + git_tag: + type: string + description: >- + Name of the tag to be checked out and used by this job. This field + cannot be specified in conjunction with git_branch or git_commit. + job_source: + $ref: '#/components/schemas/JobSource' + description: >- + The source of the job specification in the remote repository when + the job is source controlled. + required: + - git_url + - git_provider + description: >- + An optional specification for a remote Git repository containing the + source code used by tasks. + Version-controlled source code is supported by notebook, dbt, Python script, and SQL File tasks. + + If `git_source` is set, these tasks retrieve the file from the remote repository by default. + However, this behavior can be overridden by setting `source` to `WORKSPACE` on the task. + + Note: dbt and SQL File tasks support only version-controlled sources. If dbt or SQL File tasks + are used, `git_source` must be defined on the job. + Job: + type: object + properties: + created_time: + type: integer + description: >- + The time at which this job was created in epoch milliseconds + (milliseconds since 1/1/1970 UTC). + creator_user_name: + type: string + description: >- + The creator user name. This field won’t be included in the response + if the user has already been deleted. + effective_budget_policy_id: + type: string + description: >- + The id of the budget policy used by this job for cost attribution + purposes. This may be set through (in order of precedence): 1. + Budget admins through the account or workspace console 2. Jobs UI in + the job details page and Jobs API using `budget_policy_id` 3. + Inferred default based on accessible budget policies of the run_as + identity on job creation or modification. + effective_usage_policy_id: + type: string + description: >- + The id of the usage policy used by this job for cost attribution + purposes. + has_more: + type: boolean + description: >- + Indicates if the job has more array properties (`tasks`, + `job_clusters`) that are not shown. They can be accessed via + :method:jobs/get endpoint. It is only relevant for API 2.2 + :method:jobs/list requests with `expand_tasks=true`. + job_id: + type: integer + description: The canonical identifier for this job. + next_page_token: + type: string + description: A token that can be used to list the next page of array properties. + run_as_user_name: + type: string + description: >- + The email of an active workspace user or the application ID of a + service principal that the job runs as. This value can be changed by + setting the `run_as` field when creating or updating a job. By + default, `run_as_user_name` is based on the current job settings and + is set to the creator of the job if job access control is disabled + or to the user with the `is_owner` permission if job access control + is enabled. + settings: + $ref: '#/components/schemas/JobSettings' + description: >- + Settings for this job and all of its runs. These settings can be + updated using the `resetJob` method. + trigger_state: + $ref: '#/components/schemas/TriggerStateProto' + description: State of the trigger associated with the job. + description: Job was retrieved successfully. + JobAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/JobPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + JobAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/JobPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + JobCluster: + type: object + properties: + job_cluster_key: + type: string + new_cluster: + type: string + description: >- + If new_cluster, a description of a cluster that is created for each + task. + required: + - job_cluster_key + - new_cluster + JobCompliance: + type: object + properties: + job_id: + type: integer + is_compliant: + type: boolean + description: >- + Whether this job is in compliance with the latest version of its + policy. + violations: + type: object + description: >- + An object containing key-value mappings representing the first 200 + policy validation errors. The keys indicate the path where the + policy validation error is occurring. An identifier for the job + cluster is prepended to the path. The values indicate an error + message describing the policy validation error. + required: + - job_id + JobDeployment: + type: object + properties: + kind: + $ref: '#/components/schemas/JobDeploymentKind' + metadata_file_path: + type: string + description: Path of the file that contains deployment metadata. + required: + - kind + JobEmailNotifications: + type: object + properties: + no_alert_for_skipped_runs: + type: boolean + on_duration_warning_threshold_exceeded: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when the duration of a run + exceeds the threshold specified for the `RUN_DURATION_SECONDS` + metric in the `health` field. If no rule for the + `RUN_DURATION_SECONDS` metric is specified in the `health` field for + the job, notifications are not sent. + on_failure: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when a run unsuccessfully + completes. A run is considered to have completed unsuccessfully if + it ends with an `INTERNAL_ERROR` `life_cycle_state` or a `FAILED`, + or `TIMED_OUT` result_state. If this is not specified on job + creation, reset, or update the list is empty, and notifications are + not sent. + on_start: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when a run begins. If not + specified on job creation, reset, or update, the list is empty, and + notifications are not sent. + on_streaming_backlog_exceeded: + type: array + items: + type: string + description: >- + A list of email addresses to notify when any streaming backlog + thresholds are exceeded for any stream. Streaming backlog thresholds + can be set in the `health` field using the following metrics: + `STREAMING_BACKLOG_BYTES`, `STREAMING_BACKLOG_RECORDS`, + `STREAMING_BACKLOG_SECONDS`, or `STREAMING_BACKLOG_FILES`. Alerting + is based on the 10-minute average of these metrics. If the issue + persists, notifications are resent every 30 minutes. + on_success: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when a run successfully + completes. A run is considered to have completed successfully if it + ends with a `TERMINATED` `life_cycle_state` and a `SUCCESS` + result_state. If not specified on job creation, reset, or update, + the list is empty, and notifications are not sent. + JobEnvironment: + type: object + properties: + environment_key: + type: string + spec: + type: string + required: + - environment_key + JobNotificationSettings: + type: object + properties: + no_alert_for_canceled_runs: + type: boolean + no_alert_for_skipped_runs: + type: boolean + description: >- + If true, do not send notifications to recipients specified in + `on_failure` if the run is skipped. + JobParameter: + type: object + properties: + default: + type: string + name: + type: string + description: The name of the parameter + value: + type: string + description: The value used in the run + JobParameterDefinition: + type: object + properties: + name: + type: string + default: + type: string + description: Default value of the parameter. + required: + - name + - default + JobPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/JobPermissionLevel' + JobPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/JobAccessControlResponse' + object_id: + type: string + object_type: + type: string + JobPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/JobPermissionLevel' + JobRunAs: + type: object + properties: + group_name: + type: string + description: >- + Group name of an account group assigned to the workspace. Setting + this field requires being a member of the group. + service_principal_name: + type: string + description: >- + Application ID of an active service principal. Setting this field + requires the `servicePrincipal/user` role. + user_name: + type: string + description: >- + The email of an active workspace user. Non-admin users can only set + this field to their own email. + description: >- + Write-only setting. Specifies the user or service principal that the job + runs as. If not + specified, the job runs as the user who created the job. + + Either `user_name` or `service_principal_name` should be specified. If not, an error is thrown. + JobSettings: + type: object + properties: + budget_policy_id: + type: string + continuous: + $ref: '#/components/schemas/Continuous' + description: >- + An optional continuous property for this job. The continuous + property will ensure that there is always one run executing. Only + one of `schedule` and `continuous` can be used. + deployment: + $ref: '#/components/schemas/JobDeployment' + description: Deployment information for jobs managed by external sources. + description: + type: string + description: >- + An optional description for the job. The maximum length is 27700 + characters in UTF-8 encoding. + edit_mode: + $ref: '#/components/schemas/JobEditMode' + description: >- + Edit mode of the job. * `UI_LOCKED`: The job is in a locked UI state + and cannot be modified. * `EDITABLE`: The job is in an editable + state and can be modified. + email_notifications: + $ref: '#/components/schemas/JobEmailNotifications' + description: >- + An optional set of email addresses that is notified when runs of + this job begin or complete as well as when this job is deleted. + environments: + type: array + items: + $ref: '#/components/schemas/JobEnvironment' + description: >- + A list of task execution environment specifications that can be + referenced by serverless tasks of this job. For serverless notebook + tasks, if the environment_key is not specified, the notebook + environment will be used if present. If a jobs environment is + specified, it will override the notebook environment. For other + serverless tasks, the task environment is required to be specified + using environment_key in the task settings. + format: + $ref: '#/components/schemas/Format' + description: >- + Used to tell what is the format of the job. This field is ignored in + Create/Update/Reset calls. When using the Jobs API 2.1 this value is + always set to `"MULTI_TASK"`. + git_source: + $ref: '#/components/schemas/GitSource' + description: >- + An optional specification for a remote Git repository containing the + source code used by tasks. Version-controlled source code is + supported by notebook, dbt, Python script, and SQL File tasks. If + `git_source` is set, these tasks retrieve the file from the remote + repository by default. However, this behavior can be overridden by + setting `source` to `WORKSPACE` on the task. Note: dbt and SQL File + tasks support only version-controlled sources. If dbt or SQL File + tasks are used, `git_source` must be defined on the job. + health: + $ref: '#/components/schemas/JobsHealthRules' + job_clusters: + type: array + items: + $ref: '#/components/schemas/JobCluster' + description: >- + A list of job cluster specifications that can be shared and reused + by tasks of this job. Libraries cannot be declared in a shared job + cluster. You must declare dependent libraries in task settings. + max_concurrent_runs: + type: integer + description: >- + An optional maximum allowed number of concurrent runs of the job. + Set this value if you want to be able to execute multiple runs of + the same job concurrently. This is useful for example if you trigger + your job on a frequent schedule and want to allow consecutive runs + to overlap with each other, or if you want to trigger multiple runs + which differ by their input parameters. This setting affects only + new runs. For example, suppose the job’s concurrency is 4 and there + are 4 concurrent active runs. Then setting the concurrency to 3 + won’t kill any of the active runs. However, from then on, new runs + are skipped unless there are fewer than 3 active runs. This value + cannot exceed 1000. Setting this value to `0` causes all new runs to + be skipped. + name: + type: string + description: >- + An optional name for the job. The maximum length is 4096 bytes in + UTF-8 encoding. + notification_settings: + $ref: '#/components/schemas/JobNotificationSettings' + description: >- + Optional notification settings that are used when sending + notifications to each of the `email_notifications` and + `webhook_notifications` for this job. + parameters: + type: array + items: + $ref: '#/components/schemas/JobParameterDefinition' + description: Job-level parameter definitions + performance_target: + $ref: '#/components/schemas/PerformanceTarget' + description: >- + The performance mode on a serverless job. This field determines the + level of compute performance or cost-efficiency for the run. The + performance target does not apply to tasks that run on Serverless + GPU compute. * `STANDARD`: Enables cost-efficient execution of + serverless workloads. * `PERFORMANCE_OPTIMIZED`: Prioritizes fast + startup and execution times through rapid scaling and optimized + cluster performance. + queue: + $ref: '#/components/schemas/QueueSettings' + description: The queue settings of the job. + run_as: + $ref: '#/components/schemas/JobRunAs' + description: >- + The user or service principal that the job runs as, if specified in + the request. This field indicates the explicit configuration of + `run_as` for the job. To find the value in all cases, explicit or + implicit, use `run_as_user_name`. + schedule: + $ref: '#/components/schemas/CronSchedule' + description: >- + An optional periodic schedule for this job. The default behavior is + that the job only runs when triggered by clicking “Run Now” in the + Jobs UI or sending an API request to `runNow`. + tags: + type: object + description: >- + A map of tags associated with the job. These are forwarded to the + cluster as cluster tags for jobs clusters, and are subject to the + same limitations as cluster tags. A maximum of 25 tags can be added + to the job. + tasks: + type: array + items: + $ref: '#/components/schemas/Task' + description: >- + A list of task specifications to be executed by this job. It + supports up to 1000 elements in write endpoints + (:method:jobs/create, :method:jobs/reset, :method:jobs/update, + :method:jobs/submit). Read endpoints return only 100 tasks. If more + than 100 tasks are available, you can paginate through them using + :method:jobs/get. Use the `next_page_token` field at the object root + to determine if more results are available. + timeout_seconds: + type: integer + description: >- + An optional timeout applied to each run of this job. A value of `0` + means no timeout. + trigger: + $ref: '#/components/schemas/TriggerSettings' + description: >- + A configuration to trigger a run when certain conditions are met. + The default behavior is that the job runs only when triggered by + clicking “Run Now” in the Jobs UI or sending an API request to + `runNow`. + usage_policy_id: + type: string + description: >- + The id of the user specified usage policy to use for this job. If + not specified, a default usage policy may be applied when creating + or modifying the job. See `effective_usage_policy_id` for the usage + policy used by this workload. + webhook_notifications: + $ref: '#/components/schemas/WebhookNotifications' + description: >- + A collection of system notification IDs to notify when runs of this + job begin or complete. + JobSource: + type: object + properties: + job_config_path: + type: string + description: Path of the job YAML file that contains the job specification. + import_from_git_branch: + type: string + description: Name of the branch which the job is imported from. + dirty_state: + $ref: '#/components/schemas/JobSourceDirtyState' + description: >- + Dirty state indicates the job is not fully synced with the job + specification in the remote repository. Possible values are: * + `NOT_SYNCED`: The job is not yet synced with the remote job + specification. Import the remote job specification from UI to make + the job fully synced. * `DISCONNECTED`: The job is temporary + disconnected from the remote job specification and is allowed for + live edit. Import the remote job specification again from UI to make + the job fully synced. + required: + - job_config_path + - import_from_git_branch + description: >- + The source of the job specification in the remote repository when the + job is source controlled. + JobsHealthRule: + type: object + properties: + metric: + $ref: '#/components/schemas/JobsHealthMetric' + op: + $ref: '#/components/schemas/JobsHealthOperator' + value: + type: integer + description: >- + Specifies the threshold value that the health metric should obey to + satisfy the health rule. + required: + - metric + - op + - value + JobsHealthRules: + type: object + properties: + rules: + type: array + items: + $ref: '#/components/schemas/JobsHealthRule' + description: An optional set of health rules that can be defined for this job. + ListJobComplianceForPolicyResponse: + type: object + properties: + jobs: + type: array + items: + $ref: '#/components/schemas/JobCompliance' + next_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the next page + of results. If this field is not in the response, it means no + further results for the request. + prev_page_token: + type: string + description: >- + This field represents the pagination token to retrieve the previous + page of results. If this field is not in the response, it means no + further results for the request. + ListJobsResponse: + type: object + properties: + has_more: + type: boolean + description: >- + If true, additional jobs matching the provided filter are available + for listing. + jobs: + type: array + items: + $ref: '#/components/schemas/BaseJob' + description: >- + The list of jobs. Only included in the response if there are jobs to + list. + next_page_token: + type: string + description: >- + A token that can be used to list the next page of jobs (if + applicable). + prev_page_token: + type: string + description: >- + A token that can be used to list the previous page of jobs (if + applicable). + description: List of jobs was retrieved successfully. + ListRunsResponse: + type: object + properties: + has_more: + type: boolean + description: >- + If true, additional runs matching the provided filter are available + for listing. + next_page_token: + type: string + description: >- + A token that can be used to list the next page of runs (if + applicable). + prev_page_token: + type: string + description: >- + A token that can be used to list the previous page of runs (if + applicable). + runs: + type: array + items: + $ref: '#/components/schemas/BaseRun' + description: >- + A list of runs, from most recently started to least. Only included + in the response if there are runs to list. + description: List of runs was retrieved successfully. + ModelTriggerConfiguration: + type: object + properties: + condition: + $ref: '#/components/schemas/ModelTriggerConfigurationCondition' + aliases: + type: array + items: + type: string + description: >- + Aliases of the model versions to monitor. Can only be used in + conjunction with condition MODEL_ALIAS_SET. + min_time_between_triggers_seconds: + type: integer + description: >- + If set, the trigger starts a run only after the specified amount of + time has passed since the last time the trigger fired. The minimum + allowed value is 60 seconds. + securable_name: + type: string + description: >- + Name of the securable to monitor ("mycatalog.myschema.mymodel" in + the case of model-level triggers, "mycatalog.myschema" in the case + of schema-level triggers) or empty in the case of metastore-level + triggers. + wait_after_last_change_seconds: + type: integer + description: >- + If set, the trigger starts a run only after no model updates have + occurred for the specified time and can be used to wait for a series + of model updates before triggering a run. The minimum allowed value + is 60 seconds. + required: + - condition + NotebookOutput: + type: object + properties: + result: + type: string + truncated: + type: boolean + description: Whether or not the result was truncated. + NotebookTask: + type: object + properties: + notebook_path: + type: string + base_parameters: + type: object + description: >- + Base parameters to be used for each run of this job. If the run is + initiated by a call to :method:jobs/run Now with parameters + specified, the two parameters maps are merged. If the same key is + specified in `base_parameters` and in `run-now`, the value from + `run-now` is used. Use [Task parameter variables] to set parameters + containing information about job runs. If the notebook takes a + parameter that is not specified in the job’s `base_parameters` or + the `run-now` override parameters, the default value from the + notebook is used. Retrieve these parameters in a notebook using + [dbutils.widgets.get]. The JSON representation of this field cannot + exceed 1MB. [Task parameter variables]: + https://docs.databricks.com/jobs.html#parameter-variables + [dbutils.widgets.get]: + https://docs.databricks.com/dev-tools/databricks-utils.html#dbutils-widgets + source: + $ref: '#/components/schemas/Source' + description: >- + Optional location type of the notebook. When set to `WORKSPACE`, the + notebook will be retrieved from the local Databricks workspace. When + set to `GIT`, the notebook will be retrieved from a Git repository + defined in `git_source`. If the value is empty, the task will use + `GIT` if `git_source` is defined and `WORKSPACE` otherwise. * + `WORKSPACE`: Notebook is located in Databricks workspace. * `GIT`: + Notebook is located in cloud Git provider. + warehouse_id: + type: string + description: >- + Optional `warehouse_id` to run the notebook on a SQL warehouse. + Classic SQL warehouses are NOT supported, please use serverless or + pro SQL warehouses. Note that SQL warehouses only support SQL cells; + if the notebook contains non-SQL cells, the run will fail. + required: + - notebook_path + OutputSchemaInfo: + type: object + properties: + catalog_name: + type: string + expiration_time: + type: integer + description: >- + The expiration time for the output schema as a Unix timestamp in + milliseconds. + schema_name: + type: string + description: >- + Stores the catalog name, schema name, and the output schema expiration + time for the clean room + run. + PeriodicTriggerConfiguration: + type: object + properties: + interval: + type: integer + unit: + $ref: '#/components/schemas/PeriodicTriggerConfigurationTimeUnit' + description: The unit of time for the interval. + required: + - interval + - unit + PipelineParams: + type: object + properties: + full_refresh: + type: boolean + PipelineTask: + type: object + properties: + pipeline_id: + type: string + full_refresh: + type: boolean + description: If true, triggers a full refresh on the delta live table. + required: + - pipeline_id + PowerBiModel: + type: object + properties: + authentication_method: + $ref: '#/components/schemas/AuthenticationMethod' + model_name: + type: string + description: The name of the Power BI model + overwrite_existing: + type: boolean + description: Whether to overwrite existing Power BI models + storage_mode: + $ref: '#/components/schemas/StorageMode' + description: The default storage mode of the Power BI model + workspace_name: + type: string + description: The name of the Power BI workspace of the model + PowerBiTable: + type: object + properties: + catalog: + type: string + name: + type: string + description: The table name in Databricks + schema: + type: string + description: The schema name in Databricks + storage_mode: + $ref: '#/components/schemas/StorageMode' + description: The Power BI storage mode of the table + PowerBiTask: + type: object + properties: + connection_resource_name: + type: string + power_bi_model: + $ref: '#/components/schemas/PowerBiModel' + description: The semantic model to update + refresh_after_update: + type: boolean + description: Whether the model should be refreshed after the update + tables: + type: array + items: + $ref: '#/components/schemas/PowerBiTable' + description: The tables to be exported to Power BI + warehouse_id: + type: string + description: The SQL warehouse ID to use as the Power BI data source + PythonWheelTask: + type: object + properties: + package_name: + type: string + entry_point: + type: string + description: >- + Named entry point to use, if it does not exist in the metadata of + the package it executes the function from the package directly using + `$packageName.$entryPoint()` + named_parameters: + type: object + description: >- + Command-line parameters passed to Python wheel task in the form of + `["--name=task", "--data=dbfs:/path/to/data.json"]`. Leave it empty + if `parameters` is not null. + parameters: + type: array + items: + type: string + description: >- + Command-line parameters passed to Python wheel task. Leave it empty + if `named_parameters` is not null. + required: + - package_name + - entry_point + QueueDetails: + type: object + properties: + code: + $ref: '#/components/schemas/QueueDetailsCodeCode' + message: + type: string + description: >- + A descriptive message with the queuing details. This field is + unstructured, and its exact format is subject to change. + QueueSettings: + type: object + properties: + enabled: + type: boolean + required: + - enabled + RepairHistoryItem: + type: object + properties: + effective_performance_target: + $ref: '#/components/schemas/PerformanceTarget' + end_time: + type: integer + description: The end time of the (repaired) run. + id: + type: integer + description: >- + The ID of the repair. Only returned for the items that represent a + repair in `repair_history`. + start_time: + type: integer + description: The start time of the (repaired) run. + state: + $ref: '#/components/schemas/RunState' + description: Deprecated. Please use the `status` field instead. + status: + $ref: '#/components/schemas/RunStatus' + task_run_ids: + type: array + items: + type: integer + description: >- + The run IDs of the task runs that ran as part of this repair history + item. + type: + $ref: '#/components/schemas/RepairHistoryItemType' + description: >- + The repair history item type. Indicates whether a run is the + original run or a repair run. + RepairRunResponse: + type: object + properties: + repair_id: + type: integer + description: >- + The ID of the repair. Must be provided in subsequent repairs using + the `latest_repair_id` field to ensure sequential repairs. + description: Run repair was initiated. + ResolvedConditionTaskValues: + type: object + properties: + left: + type: string + right: + type: string + ResolvedDbtTaskValues: + type: object + properties: + commands: + type: array + items: + type: string + ResolvedNotebookTaskValues: + type: object + properties: + base_parameters: + type: object + ResolvedParamPairValues: + type: object + properties: + parameters: + type: object + ResolvedPythonWheelTaskValues: + type: object + properties: + named_parameters: + type: object + parameters: + type: array + items: + type: string + ResolvedRunJobTaskValues: + type: object + properties: + job_parameters: + type: object + parameters: + type: object + ResolvedStringParamsValues: + type: object + properties: + parameters: + type: array + items: + type: string + ResolvedValues: + type: object + properties: + condition_task: + $ref: '#/components/schemas/ResolvedConditionTaskValues' + dbt_task: + $ref: '#/components/schemas/ResolvedDbtTaskValues' + notebook_task: + $ref: '#/components/schemas/ResolvedNotebookTaskValues' + python_wheel_task: + $ref: '#/components/schemas/ResolvedPythonWheelTaskValues' + run_job_task: + $ref: '#/components/schemas/ResolvedRunJobTaskValues' + simulation_task: + $ref: '#/components/schemas/ResolvedParamPairValues' + spark_jar_task: + $ref: '#/components/schemas/ResolvedStringParamsValues' + spark_python_task: + $ref: '#/components/schemas/ResolvedStringParamsValues' + spark_submit_task: + $ref: '#/components/schemas/ResolvedStringParamsValues' + sql_task: + $ref: '#/components/schemas/ResolvedParamPairValues' + Run: + type: object + properties: + attempt_number: + type: integer + description: >- + The sequence number of this run attempt for a triggered job run. The + initial attempt of a run has an attempt_number of 0. If the initial + run attempt fails, and the job has a retry policy (`max_retries` > + 0), subsequent runs are created with an `original_attempt_run_id` of + the original attempt’s ID and an incrementing `attempt_number`. Runs + are retried only until they succeed, and the maximum + `attempt_number` is the same as the `max_retries` value for the job. + cleanup_duration: + type: integer + description: >- + The time in milliseconds it took to terminate the cluster and clean + up any associated artifacts. The duration of a task run is the sum + of the `setup_duration`, `execution_duration`, and the + `cleanup_duration`. The `cleanup_duration` field is set to 0 for + multitask job runs. The total duration of a multitask job run is the + value of the `run_duration` field. + cluster_instance: + $ref: '#/components/schemas/ClusterInstance' + description: >- + The cluster used for this run. If the run is specified to use a new + cluster, this field is set once the Jobs service has requested a + cluster for the run. + cluster_spec: + $ref: '#/components/schemas/ClusterSpec' + description: >- + A snapshot of the job’s cluster specification when this run was + created. + creator_user_name: + type: string + description: >- + The creator user name. This field won’t be included in the response + if the user has already been deleted. + description: + type: string + description: Description of the run + effective_performance_target: + $ref: '#/components/schemas/PerformanceTarget' + description: >- + The actual performance target used by the serverless run during + execution. This can differ from the client-set performance target on + the request depending on whether the performance mode is supported + by the job type. * `STANDARD`: Enables cost-efficient execution of + serverless workloads. * `PERFORMANCE_OPTIMIZED`: Prioritizes fast + startup and execution times through rapid scaling and optimized + cluster performance. + effective_usage_policy_id: + type: string + description: >- + The id of the usage policy used by this run for cost attribution + purposes. + end_time: + type: integer + description: >- + The time at which this run ended in epoch milliseconds (milliseconds + since 1/1/1970 UTC). This field is set to 0 if the job is still + running. + execution_duration: + type: integer + description: >- + The time in milliseconds it took to execute the commands in the JAR + or notebook until they completed, failed, timed out, were cancelled, + or encountered an unexpected error. The duration of a task run is + the sum of the `setup_duration`, `execution_duration`, and the + `cleanup_duration`. The `execution_duration` field is set to 0 for + multitask job runs. The total duration of a multitask job run is the + value of the `run_duration` field. + git_source: + $ref: '#/components/schemas/GitSource' + description: >- + An optional specification for a remote Git repository containing the + source code used by tasks. Version-controlled source code is + supported by notebook, dbt, Python script, and SQL File tasks. If + `git_source` is set, these tasks retrieve the file from the remote + repository by default. However, this behavior can be overridden by + setting `source` to `WORKSPACE` on the task. Note: dbt and SQL File + tasks support only version-controlled sources. If dbt or SQL File + tasks are used, `git_source` must be defined on the job. + has_more: + type: boolean + description: >- + Indicates if the run has more array properties (`tasks`, + `job_clusters`) that are not shown. They can be accessed via + :method:jobs/getrun endpoint. It is only relevant for API 2.2 + :method:jobs/listruns requests with `expand_tasks=true`. + iterations: + type: array + items: + $ref: '#/components/schemas/RunTask' + description: >- + Only populated by for-each iterations. The parent for-each task is + located in tasks array. + job_clusters: + type: array + items: + $ref: '#/components/schemas/JobCluster' + description: >- + A list of job cluster specifications that can be shared and reused + by tasks of this job. Libraries cannot be declared in a shared job + cluster. You must declare dependent libraries in task settings. If + more than 100 job clusters are available, you can paginate through + them using :method:jobs/getrun. + job_id: + type: integer + description: The canonical identifier of the job that contains this run. + job_parameters: + type: array + items: + $ref: '#/components/schemas/JobParameter' + description: Job-level parameters used in the run + job_run_id: + type: integer + description: >- + ID of the job run that this run belongs to. For legacy and + single-task job runs the field is populated with the job run ID. For + task runs, the field is populated with the ID of the job run that + the task run belongs to. + next_page_token: + type: string + description: A token that can be used to list the next page of array properties. + number_in_job: + type: integer + description: >- + A unique identifier for this job run. This is set to the same value + as `run_id`. + original_attempt_run_id: + type: integer + description: >- + If this run is a retry of a prior run attempt, this field contains + the run_id of the original attempt; otherwise, it is the same as the + run_id. + overriding_parameters: + $ref: '#/components/schemas/RunParameters' + description: The parameters used for this run. + queue_duration: + type: integer + description: The time in milliseconds that the run has spent in the queue. + repair_history: + type: array + items: + $ref: '#/components/schemas/RepairHistoryItem' + description: The repair history of the run. + run_duration: + type: integer + description: >- + The time in milliseconds it took the job run and all of its repairs + to finish. + run_id: + type: integer + description: >- + The canonical identifier of the run. This ID is unique across all + runs of all jobs. + run_name: + type: string + description: >- + An optional name for the run. The maximum length is 4096 bytes in + UTF-8 encoding. + run_page_url: + type: string + description: The URL to the detail page of the run. + run_type: + $ref: '#/components/schemas/RunType' + schedule: + $ref: '#/components/schemas/CronSchedule' + description: >- + The cron schedule that triggered this run if it was triggered by the + periodic scheduler. + setup_duration: + type: integer + description: >- + The time in milliseconds it took to set up the cluster. For runs + that run on new clusters this is the cluster creation time, for runs + that run on existing clusters this time should be very short. The + duration of a task run is the sum of the `setup_duration`, + `execution_duration`, and the `cleanup_duration`. The + `setup_duration` field is set to 0 for multitask job runs. The total + duration of a multitask job run is the value of the `run_duration` + field. + start_time: + type: integer + description: >- + The time at which this run was started in epoch milliseconds + (milliseconds since 1/1/1970 UTC). This may not be the time when the + job task starts executing, for example, if the job is scheduled to + run on a new cluster, this is the time the cluster creation call is + issued. + state: + $ref: '#/components/schemas/RunState' + description: Deprecated. Please use the `status` field instead. + status: + $ref: '#/components/schemas/RunStatus' + tasks: + type: array + items: + $ref: '#/components/schemas/RunTask' + description: >- + The list of tasks performed by the run. Each task has its own + `run_id` which you can use to call `JobsGetOutput` to retrieve the + run resutls. If more than 100 tasks are available, you can paginate + through them using :method:jobs/getrun. Use the `next_page_token` + field at the object root to determine if more results are available. + trigger: + $ref: '#/components/schemas/TriggerType' + trigger_info: + $ref: '#/components/schemas/TriggerInfo' + description: Run was retrieved successfully + RunConditionTask: + type: object + properties: + op: + $ref: '#/components/schemas/ConditionTaskOp' + left: + type: string + description: >- + The left operand of the condition task. Can be either a string value + or a job state or parameter reference. + right: + type: string + description: >- + The right operand of the condition task. Can be either a string + value or a job state or parameter reference. + outcome: + type: string + description: >- + The condition expression evaluation result. Filled in if the task + was successfully completed. Can be `"true"` or `"false"` + required: + - op + - left + - right + RunForEachTask: + type: object + properties: + inputs: + type: string + task: + $ref: '#/components/schemas/Task' + description: >- + Configuration for the task that will be run for each element in the + array + concurrency: + type: integer + description: >- + An optional maximum allowed number of concurrent runs of the task. + Set this value if you want to be able to execute multiple runs of + the task concurrently. + stats: + $ref: '#/components/schemas/ForEachStats' + description: >- + Read only field. Populated for GetRun and ListRuns RPC calls and + stores the execution stats of an For each task + required: + - inputs + - task + RunJobOutput: + type: object + properties: + run_id: + type: integer + RunJobTask: + type: object + properties: + job_id: + type: integer + dbt_commands: + type: array + items: + type: string + description: >- + An array of commands to execute for jobs with the dbt task, for + example `"dbt_commands": ["dbt deps", "dbt seed", "dbt deps", "dbt + seed", "dbt run"]` ⚠ **Deprecation note** Use [job parameters] to + pass information down to tasks. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + jar_params: + type: array + items: + type: string + description: >- + A list of parameters for jobs with Spark JAR tasks, for example + `"jar_params": ["john doe", "35"]`. The parameters are used to + invoke the main function of the main class specified in the Spark + JAR task. If not specified upon `run-now`, it defaults to an empty + list. jar_params cannot be specified in conjunction with + notebook_params. The JSON representation of this field (for example + `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 bytes. ⚠ + **Deprecation note** Use [job parameters] to pass information down + to tasks. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + job_parameters: + type: object + description: Job-level parameters used to trigger the job. + notebook_params: + type: object + description: >- + A map from keys to values for jobs with notebook task, for example + `"notebook_params": {"name": "john doe", "age": "35"}`. The map is + passed to the notebook and is accessible through the + [dbutils.widgets.get] function. If not specified upon `run-now`, the + triggered run uses the job’s base parameters. notebook_params cannot + be specified in conjunction with jar_params. ⚠ **Deprecation note** + Use [job parameters] to pass information down to tasks. The JSON + representation of this field (for example + `{"notebook_params":{"name":"john doe","age":"35"}}`) cannot exceed + 10,000 bytes. [dbutils.widgets.get]: + https://docs.databricks.com/dev-tools/databricks-utils.html [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + pipeline_params: + $ref: '#/components/schemas/PipelineParams' + description: Controls whether the pipeline should perform a full refresh + python_named_params: + type: object + python_params: + type: array + items: + type: string + description: >- + A list of parameters for jobs with Python tasks, for example + `"python_params": ["john doe", "35"]`. The parameters are passed to + Python file as command-line parameters. If specified upon `run-now`, + it would overwrite the parameters specified in job setting. The JSON + representation of this field (for example `{"python_params":["john + doe","35"]}`) cannot exceed 10,000 bytes. ⚠ **Deprecation note** Use + [job parameters] to pass information down to tasks. Important These + parameters accept only Latin characters (ASCII character set). Using + non-ASCII characters returns an error. Examples of invalid, + non-ASCII characters are Chinese, Japanese kanjis, and emojis. [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + spark_submit_params: + type: array + items: + type: string + description: >- + A list of parameters for jobs with spark submit task, for example + `"spark_submit_params": ["--class", + "org.apache.spark.examples.SparkPi"]`. The parameters are passed to + spark-submit script as command-line parameters. If specified upon + `run-now`, it would overwrite the parameters specified in job + setting. The JSON representation of this field (for example + `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. ⚠ + **Deprecation note** Use [job parameters] to pass information down + to tasks. Important These parameters accept only Latin characters + (ASCII character set). Using non-ASCII characters returns an error. + Examples of invalid, non-ASCII characters are Chinese, Japanese + kanjis, and emojis. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + sql_params: + type: object + description: >- + A map from keys to values for jobs with SQL task, for example + `"sql_params": {"name": "john doe", "age": "35"}`. The SQL alert + task does not support custom parameters. ⚠ **Deprecation note** Use + [job parameters] to pass information down to tasks. [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + required: + - job_id + RunNowResponse: + type: object + properties: + number_in_job: + type: integer + description: >- + A unique identifier for this job run. This is set to the same value + as `run_id`. + run_id: + type: integer + description: The globally unique ID of the newly triggered run. + description: Run was started successfully. + RunOutput: + type: object + properties: + clean_rooms_notebook_output: + $ref: >- + #/components/schemas/CleanRoomsNotebookTaskCleanRoomsNotebookTaskOutput + description: The output of a clean rooms notebook task, if available + dashboard_output: + $ref: '#/components/schemas/DashboardTaskOutput' + description: The output of a dashboard task, if available + dbt_cloud_output: + $ref: '#/components/schemas/DbtCloudTaskOutput' + description: Deprecated in favor of the new dbt_platform_output + dbt_output: + $ref: '#/components/schemas/DbtOutput' + description: The output of a dbt task, if available. + dbt_platform_output: + $ref: '#/components/schemas/DbtPlatformTaskOutput' + error: + type: string + description: >- + An error message indicating why a task failed or why output is not + available. The message is unstructured, and its exact format is + subject to change. + error_trace: + type: string + description: >- + If there was an error executing the run, this field contains any + available stack traces. + info: + type: string + logs: + type: string + description: >- + The output from tasks that write to standard streams (stdout/stderr) + such as spark_jar_task, spark_python_task, python_wheel_task. It's + not supported for the notebook_task, pipeline_task or + spark_submit_task. Databricks restricts this API to return the last + 5 MB of these logs. + logs_truncated: + type: boolean + description: Whether the logs are truncated. + metadata: + $ref: '#/components/schemas/Run' + description: All details of the run except for its output. + notebook_output: + $ref: '#/components/schemas/NotebookOutput' + description: >- + The output of a notebook task, if available. A notebook task that + terminates (either successfully or with a failure) without calling + `dbutils.notebook.exit()` is considered to have an empty output. + This field is set but its result value is empty. Databricks + restricts this API to return the first 5 MB of the output. To return + a larger result, use the [ClusterLogConf] field to configure log + storage for the job cluster. [ClusterLogConf]: + https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterlogconf + run_job_output: + $ref: '#/components/schemas/RunJobOutput' + description: The output of a run job task, if available + sql_output: + $ref: '#/components/schemas/SqlOutput' + description: The output of a SQL task, if available. + description: Run output was retrieved successfully. + RunParameters: + type: object + properties: + dbt_commands: + type: array + items: + type: string + jar_params: + type: array + items: + type: string + description: >- + A list of parameters for jobs with Spark JAR tasks, for example + `"jar_params": ["john doe", "35"]`. The parameters are used to + invoke the main function of the main class specified in the Spark + JAR task. If not specified upon `run-now`, it defaults to an empty + list. jar_params cannot be specified in conjunction with + notebook_params. The JSON representation of this field (for example + `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 bytes. ⚠ + **Deprecation note** Use [job parameters] to pass information down + to tasks. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + notebook_params: + type: object + description: >- + A map from keys to values for jobs with notebook task, for example + `"notebook_params": {"name": "john doe", "age": "35"}`. The map is + passed to the notebook and is accessible through the + [dbutils.widgets.get] function. If not specified upon `run-now`, the + triggered run uses the job’s base parameters. notebook_params cannot + be specified in conjunction with jar_params. ⚠ **Deprecation note** + Use [job parameters] to pass information down to tasks. The JSON + representation of this field (for example + `{"notebook_params":{"name":"john doe","age":"35"}}`) cannot exceed + 10,000 bytes. [dbutils.widgets.get]: + https://docs.databricks.com/dev-tools/databricks-utils.html [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + pipeline_params: + $ref: '#/components/schemas/PipelineParams' + description: Controls whether the pipeline should perform a full refresh + python_named_params: + type: object + python_params: + type: array + items: + type: string + description: >- + A list of parameters for jobs with Python tasks, for example + `"python_params": ["john doe", "35"]`. The parameters are passed to + Python file as command-line parameters. If specified upon `run-now`, + it would overwrite the parameters specified in job setting. The JSON + representation of this field (for example `{"python_params":["john + doe","35"]}`) cannot exceed 10,000 bytes. ⚠ **Deprecation note** Use + [job parameters] to pass information down to tasks. Important These + parameters accept only Latin characters (ASCII character set). Using + non-ASCII characters returns an error. Examples of invalid, + non-ASCII characters are Chinese, Japanese kanjis, and emojis. [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + spark_submit_params: + type: array + items: + type: string + description: >- + A list of parameters for jobs with spark submit task, for example + `"spark_submit_params": ["--class", + "org.apache.spark.examples.SparkPi"]`. The parameters are passed to + spark-submit script as command-line parameters. If specified upon + `run-now`, it would overwrite the parameters specified in job + setting. The JSON representation of this field (for example + `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. ⚠ + **Deprecation note** Use [job parameters] to pass information down + to tasks. Important These parameters accept only Latin characters + (ASCII character set). Using non-ASCII characters returns an error. + Examples of invalid, non-ASCII characters are Chinese, Japanese + kanjis, and emojis. [job parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + sql_params: + type: object + description: >- + A map from keys to values for jobs with SQL task, for example + `"sql_params": {"name": "john doe", "age": "35"}`. The SQL alert + task does not support custom parameters. ⚠ **Deprecation note** Use + [job parameters] to pass information down to tasks. [job + parameters]: + https://docs.databricks.com/jobs/job-parameters.html#job-parameter-pushdown + RunState: + type: object + properties: + life_cycle_state: + $ref: '#/components/schemas/RunLifeCycleState' + description: >- + A value indicating the run's current lifecycle state. This field is + always available in the response. Note: Additional states might be + introduced in future releases. + queue_reason: + type: string + description: The reason indicating why the run was queued. + result_state: + $ref: '#/components/schemas/RunResultState' + description: >- + A value indicating the run's result. This field is only available + for terminal lifecycle states. Note: Additional states might be + introduced in future releases. + state_message: + type: string + description: >- + A descriptive message for the current state. This field is + unstructured, and its exact format is subject to change. + user_cancelled_or_timedout: + type: boolean + description: >- + A value indicating whether a run was canceled manually by a user or + by the scheduler because the run timed out. + description: The current state of the run. + RunStatus: + type: object + properties: + queue_details: + $ref: '#/components/schemas/QueueDetails' + description: If the run was queued, details about the reason for queuing the run. + state: + $ref: '#/components/schemas/RunLifecycleStateV2State' + termination_details: + $ref: '#/components/schemas/TerminationDetails' + description: >- + If the run is in a TERMINATING or TERMINATED state, details about + the reason for terminating the run. + description: The current status of the run + RunTask: + type: object + properties: + task_key: + type: string + description: >- + A unique name for the task. This field is used to refer to this task + from other tasks. This field is required and must be unique within + its parent job. On Update or Reset, this field is used to reference + the tasks to be updated or reset. + attempt_number: + type: integer + description: >- + The sequence number of this run attempt for a triggered job run. The + initial attempt of a run has an attempt_number of 0. If the initial + run attempt fails, and the job has a retry policy (`max_retries` > + 0), subsequent runs are created with an `original_attempt_run_id` of + the original attempt’s ID and an incrementing `attempt_number`. Runs + are retried only until they succeed, and the maximum + `attempt_number` is the same as the `max_retries` value for the job. + clean_rooms_notebook_task: + $ref: '#/components/schemas/CleanRoomsNotebookTask' + description: >- + The task runs a [clean rooms] notebook when the + `clean_rooms_notebook_task` field is present. [clean rooms]: + https://docs.databricks.com/clean-rooms/index.html + cleanup_duration: + type: integer + description: >- + The time in milliseconds it took to terminate the cluster and clean + up any associated artifacts. The duration of a task run is the sum + of the `setup_duration`, `execution_duration`, and the + `cleanup_duration`. The `cleanup_duration` field is set to 0 for + multitask job runs. The total duration of a multitask job run is the + value of the `run_duration` field. + cluster_instance: + $ref: '#/components/schemas/ClusterInstance' + description: >- + The cluster used for this run. If the run is specified to use a new + cluster, this field is set once the Jobs service has requested a + cluster for the run. + compute: + $ref: '#/components/schemas/Compute' + description: Task level compute configuration. + condition_task: + $ref: '#/components/schemas/RunConditionTask' + description: >- + The task evaluates a condition that can be used to control the + execution of other tasks when the `condition_task` field is present. + The condition task does not require a cluster to execute and does + not support retries or notifications. + dashboard_task: + $ref: '#/components/schemas/DashboardTask' + description: The task refreshes a dashboard and sends a snapshot to subscribers. + dbt_cloud_task: + $ref: '#/components/schemas/DbtCloudTask' + description: >- + Task type for dbt cloud, deprecated in favor of the new name + dbt_platform_task + dbt_platform_task: + $ref: '#/components/schemas/DbtPlatformTask' + dbt_task: + $ref: '#/components/schemas/DbtTask' + description: >- + The task runs one or more dbt commands when the `dbt_task` field is + present. The dbt task requires both Databricks SQL and the ability + to use a serverless or a pro SQL warehouse. + depends_on: + type: array + items: + $ref: '#/components/schemas/TaskDependency' + description: >- + An optional array of objects specifying the dependency graph of the + task. All tasks specified in this field must complete successfully + before executing this task. The key is `task_key`, and the value is + the name assigned to the dependent task. + description: + type: string + description: An optional description for this task. + effective_performance_target: + $ref: '#/components/schemas/PerformanceTarget' + description: >- + The actual performance target used by the serverless run during + execution. This can differ from the client-set performance target on + the request depending on whether the performance mode is supported + by the job type. * `STANDARD`: Enables cost-efficient execution of + serverless workloads. * `PERFORMANCE_OPTIMIZED`: Prioritizes fast + startup and execution times through rapid scaling and optimized + cluster performance. + email_notifications: + $ref: '#/components/schemas/JobEmailNotifications' + description: >- + An optional set of email addresses notified when the task run begins + or completes. The default behavior is to not send any emails. + end_time: + type: integer + description: >- + The time at which this run ended in epoch milliseconds (milliseconds + since 1/1/1970 UTC). This field is set to 0 if the job is still + running. + environment_key: + type: string + description: >- + The key that references an environment spec in a job. This field is + required for Python script, Python wheel and dbt tasks when using + serverless compute. + execution_duration: + type: integer + description: >- + The time in milliseconds it took to execute the commands in the JAR + or notebook until they completed, failed, timed out, were cancelled, + or encountered an unexpected error. The duration of a task run is + the sum of the `setup_duration`, `execution_duration`, and the + `cleanup_duration`. The `execution_duration` field is set to 0 for + multitask job runs. The total duration of a multitask job run is the + value of the `run_duration` field. + existing_cluster_id: + type: string + description: >- + If existing_cluster_id, the ID of an existing cluster that is used + for all runs. When running jobs or tasks on an existing cluster, you + may need to manually restart the cluster if it stops responding. We + suggest running jobs and tasks on new clusters for greater + reliability + for_each_task: + $ref: '#/components/schemas/RunForEachTask' + description: >- + The task executes a nested task for every input provided when the + `for_each_task` field is present. + gen_ai_compute_task: + $ref: '#/components/schemas/GenAiComputeTask' + git_source: + $ref: '#/components/schemas/GitSource' + description: >- + An optional specification for a remote Git repository containing the + source code used by tasks. Version-controlled source code is + supported by notebook, dbt, Python script, and SQL File tasks. If + `git_source` is set, these tasks retrieve the file from the remote + repository by default. However, this behavior can be overridden by + setting `source` to `WORKSPACE` on the task. Note: dbt and SQL File + tasks support only version-controlled sources. If dbt or SQL File + tasks are used, `git_source` must be defined on the job. + job_cluster_key: + type: string + description: >- + If job_cluster_key, this task is executed reusing the cluster + specified in `job.settings.job_clusters`. + libraries: + type: string + description: >- + An optional list of libraries to be installed on the cluster. The + default value is an empty list. + new_cluster: + type: string + description: >- + If new_cluster, a description of a new cluster that is created for + each run. + notebook_task: + $ref: '#/components/schemas/NotebookTask' + description: The task runs a notebook when the `notebook_task` field is present. + notification_settings: + $ref: '#/components/schemas/TaskNotificationSettings' + description: >- + Optional notification settings that are used when sending + notifications to each of the `email_notifications` and + `webhook_notifications` for this task run. + pipeline_task: + $ref: '#/components/schemas/PipelineTask' + description: >- + The task triggers a pipeline update when the `pipeline_task` field + is present. Only pipelines configured to use triggered more are + supported. + power_bi_task: + $ref: '#/components/schemas/PowerBiTask' + description: >- + The task triggers a Power BI semantic model update when the + `power_bi_task` field is present. + python_wheel_task: + $ref: '#/components/schemas/PythonWheelTask' + description: >- + The task runs a Python wheel when the `python_wheel_task` field is + present. + queue_duration: + type: integer + description: The time in milliseconds that the run has spent in the queue. + resolved_values: + $ref: '#/components/schemas/ResolvedValues' + description: Parameter values including resolved references + run_duration: + type: integer + description: >- + The time in milliseconds it took the job run and all of its repairs + to finish. + run_id: + type: integer + description: The ID of the task run. + run_if: + $ref: '#/components/schemas/RunIf' + description: >- + An optional value indicating the condition that determines whether + the task should be run once its dependencies have been completed. + When omitted, defaults to `ALL_SUCCESS`. See :method:jobs/create for + a list of possible values. + run_job_task: + $ref: '#/components/schemas/RunJobTask' + description: >- + The task triggers another job when the `run_job_task` field is + present. + run_page_url: + type: string + setup_duration: + type: integer + description: >- + The time in milliseconds it took to set up the cluster. For runs + that run on new clusters this is the cluster creation time, for runs + that run on existing clusters this time should be very short. The + duration of a task run is the sum of the `setup_duration`, + `execution_duration`, and the `cleanup_duration`. The + `setup_duration` field is set to 0 for multitask job runs. The total + duration of a multitask job run is the value of the `run_duration` + field. + spark_jar_task: + $ref: '#/components/schemas/SparkJarTask' + description: The task runs a JAR when the `spark_jar_task` field is present. + spark_python_task: + $ref: '#/components/schemas/SparkPythonTask' + description: >- + The task runs a Python file when the `spark_python_task` field is + present. + spark_submit_task: + $ref: '#/components/schemas/SparkSubmitTask' + description: >- + (Legacy) The task runs the spark-submit script when the + spark_submit_task field is present. Databricks recommends using the + spark_jar_task instead; see [Spark Submit task for + jobs](/jobs/spark-submit). + sql_task: + $ref: '#/components/schemas/SqlTask' + description: >- + The task runs a SQL query or file, or it refreshes a SQL alert or a + legacy SQL dashboard when the `sql_task` field is present. + start_time: + type: integer + description: >- + The time at which this run was started in epoch milliseconds + (milliseconds since 1/1/1970 UTC). This may not be the time when the + job task starts executing, for example, if the job is scheduled to + run on a new cluster, this is the time the cluster creation call is + issued. + state: + $ref: '#/components/schemas/RunState' + description: Deprecated. Please use the `status` field instead. + status: + $ref: '#/components/schemas/RunStatus' + timeout_seconds: + type: integer + description: >- + An optional timeout applied to each run of this job task. A value of + `0` means no timeout. + webhook_notifications: + $ref: '#/components/schemas/WebhookNotifications' + description: >- + A collection of system notification IDs to notify when the run + begins or completes. The default behavior is to not send any system + notifications. Task webhooks respect the task notification settings. + required: + - task_key + description: Used when outputting a child run, in GetRun or ListRuns. + SparkJarTask: + type: object + properties: + jar_uri: + type: string + main_class_name: + type: string + description: >- + The full name of the class containing the main method to be + executed. This class must be contained in a JAR provided as a + library. The code must use `SparkContext.getOrCreate` to obtain a + Spark context; otherwise, runs of the job fail. + parameters: + type: array + items: + type: string + description: >- + Parameters passed to the main method. Use [Task parameter variables] + to set parameters containing information about job runs. [Task + parameter variables]: + https://docs.databricks.com/jobs.html#parameter-variables + run_as_repl: + type: boolean + description: Deprecated. A value of `false` is no longer supported. + SparkPythonTask: + type: object + properties: + python_file: + type: string + parameters: + type: array + items: + type: string + description: >- + Command line parameters passed to the Python file. Use [Task + parameter variables] to set parameters containing information about + job runs. [Task parameter variables]: + https://docs.databricks.com/jobs.html#parameter-variables + source: + $ref: '#/components/schemas/Source' + description: >- + Optional location type of the Python file. When set to `WORKSPACE` + or not specified, the file will be retrieved from the local + Databricks workspace or cloud location (if the `python_file` has a + URI format). When set to `GIT`, the Python file will be retrieved + from a Git repository defined in `git_source`. * `WORKSPACE`: The + Python file is located in a Databricks workspace or at a cloud + filesystem URI. * `GIT`: The Python file is located in a remote Git + repository. + required: + - python_file + SparkSubmitTask: + type: object + properties: + parameters: + type: array + items: + type: string + SqlAlertOutput: + type: object + properties: + alert_state: + $ref: '#/components/schemas/SqlAlertState' + output_link: + type: string + description: The link to find the output results. + query_text: + type: string + description: >- + The text of the SQL query. Can Run permission of the SQL query + associated with the SQL alert is required to view this field. + sql_statements: + type: array + items: + $ref: '#/components/schemas/SqlStatementOutput' + description: Information about SQL statements executed in the run. + warehouse_id: + type: string + description: The canonical identifier of the SQL warehouse. + SqlDashboardOutput: + type: object + properties: + warehouse_id: + type: string + widgets: + type: array + items: + $ref: '#/components/schemas/SqlDashboardWidgetOutput' + description: >- + Widgets executed in the run. Only SQL query based widgets are + listed. + SqlDashboardWidgetOutput: + type: object + properties: + end_time: + type: integer + error: + $ref: '#/components/schemas/SqlOutputError' + description: The information about the error when execution fails. + output_link: + type: string + description: The link to find the output results. + start_time: + type: integer + description: >- + Time (in epoch milliseconds) when execution of the SQL widget + starts. + status: + $ref: '#/components/schemas/SqlDashboardWidgetOutputStatus' + description: The execution status of the SQL widget. + widget_id: + type: string + description: The canonical identifier of the SQL widget. + widget_title: + type: string + description: The title of the SQL widget. + SqlOutput: + type: object + properties: + alert_output: + $ref: '#/components/schemas/SqlAlertOutput' + dashboard_output: + $ref: '#/components/schemas/SqlDashboardOutput' + description: The output of a SQL dashboard task, if available. + query_output: + $ref: '#/components/schemas/SqlQueryOutput' + description: The output of a SQL query task, if available. + SqlOutputError: + type: object + properties: + message: + type: string + SqlQueryOutput: + type: object + properties: + endpoint_id: + type: string + output_link: + type: string + description: The link to find the output results. + query_text: + type: string + description: >- + The text of the SQL query. Can Run permission of the SQL query is + required to view this field. + sql_statements: + type: array + items: + $ref: '#/components/schemas/SqlStatementOutput' + description: Information about SQL statements executed in the run. + warehouse_id: + type: string + description: The canonical identifier of the SQL warehouse. + SqlStatementOutput: + type: object + properties: + lookup_key: + type: string + SqlTask: + type: object + properties: + warehouse_id: + type: string + alert: + $ref: '#/components/schemas/SqlTaskAlert' + description: If alert, indicates that this job must refresh a SQL alert. + dashboard: + $ref: '#/components/schemas/SqlTaskDashboard' + description: If dashboard, indicates that this job must refresh a SQL dashboard. + file: + $ref: '#/components/schemas/SqlTaskFile' + description: >- + If file, indicates that this job runs a SQL file in a remote Git + repository. + parameters: + type: object + description: >- + Parameters to be used for each run of this job. The SQL alert task + does not support custom parameters. + query: + $ref: '#/components/schemas/SqlTaskQuery' + description: If query, indicates that this job must execute a SQL query. + required: + - warehouse_id + SqlTaskAlert: + type: object + properties: + alert_id: + type: string + pause_subscriptions: + type: boolean + description: If true, the alert notifications are not sent to subscribers. + subscriptions: + type: array + items: + $ref: '#/components/schemas/SqlTaskSubscription' + description: If specified, alert notifications are sent to subscribers. + required: + - alert_id + SqlTaskDashboard: + type: object + properties: + dashboard_id: + type: string + custom_subject: + type: string + description: Subject of the email sent to subscribers of this task. + pause_subscriptions: + type: boolean + description: >- + If true, the dashboard snapshot is not taken, and emails are not + sent to subscribers. + subscriptions: + type: array + items: + $ref: '#/components/schemas/SqlTaskSubscription' + description: If specified, dashboard snapshots are sent to subscriptions. + required: + - dashboard_id + SqlTaskFile: + type: object + properties: + path: + type: string + source: + $ref: '#/components/schemas/Source' + description: >- + Optional location type of the SQL file. When set to `WORKSPACE`, the + SQL file will be retrieved from the local Databricks workspace. When + set to `GIT`, the SQL file will be retrieved from a Git repository + defined in `git_source`. If the value is empty, the task will use + `GIT` if `git_source` is defined and `WORKSPACE` otherwise. * + `WORKSPACE`: SQL file is located in Databricks workspace. * `GIT`: + SQL file is located in cloud Git provider. + required: + - path + SqlTaskQuery: + type: object + properties: + query_id: + type: string + required: + - query_id + SqlTaskSubscription: + type: object + properties: + destination_id: + type: string + user_name: + type: string + description: >- + The user name to receive the subscription email. This parameter is + mutually exclusive with destination_id. You cannot set both + destination_id and user_name for subscription notifications. + SubmitRunResponse: + type: object + properties: + run_id: + type: integer + description: The canonical identifier for the newly submitted run. + description: Run was created and started successfully. + SubmitTask: + type: object + properties: + task_key: + type: string + clean_rooms_notebook_task: + $ref: '#/components/schemas/CleanRoomsNotebookTask' + description: >- + The task runs a [clean rooms] notebook when the + `clean_rooms_notebook_task` field is present. [clean rooms]: + https://docs.databricks.com/clean-rooms/index.html + compute: + $ref: '#/components/schemas/Compute' + description: Task level compute configuration. + condition_task: + $ref: '#/components/schemas/ConditionTask' + description: >- + The task evaluates a condition that can be used to control the + execution of other tasks when the `condition_task` field is present. + The condition task does not require a cluster to execute and does + not support retries or notifications. + dashboard_task: + $ref: '#/components/schemas/DashboardTask' + description: The task refreshes a dashboard and sends a snapshot to subscribers. + dbt_cloud_task: + $ref: '#/components/schemas/DbtCloudTask' + description: >- + Task type for dbt cloud, deprecated in favor of the new name + dbt_platform_task + dbt_platform_task: + $ref: '#/components/schemas/DbtPlatformTask' + dbt_task: + $ref: '#/components/schemas/DbtTask' + description: >- + The task runs one or more dbt commands when the `dbt_task` field is + present. The dbt task requires both Databricks SQL and the ability + to use a serverless or a pro SQL warehouse. + depends_on: + type: array + items: + $ref: '#/components/schemas/TaskDependency' + description: >- + An optional array of objects specifying the dependency graph of the + task. All tasks specified in this field must complete successfully + before executing this task. The key is `task_key`, and the value is + the name assigned to the dependent task. + description: + type: string + description: An optional description for this task. + email_notifications: + $ref: '#/components/schemas/JobEmailNotifications' + description: >- + An optional set of email addresses notified when the task run begins + or completes. The default behavior is to not send any emails. + environment_key: + type: string + description: >- + The key that references an environment spec in a job. This field is + required for Python script, Python wheel and dbt tasks when using + serverless compute. + existing_cluster_id: + type: string + description: >- + If existing_cluster_id, the ID of an existing cluster that is used + for all runs. When running jobs or tasks on an existing cluster, you + may need to manually restart the cluster if it stops responding. We + suggest running jobs and tasks on new clusters for greater + reliability + for_each_task: + $ref: '#/components/schemas/ForEachTask' + description: >- + The task executes a nested task for every input provided when the + `for_each_task` field is present. + gen_ai_compute_task: + $ref: '#/components/schemas/GenAiComputeTask' + health: + $ref: '#/components/schemas/JobsHealthRules' + libraries: + type: string + description: >- + An optional list of libraries to be installed on the cluster. The + default value is an empty list. + new_cluster: + type: string + description: >- + If new_cluster, a description of a new cluster that is created for + each run. + notebook_task: + $ref: '#/components/schemas/NotebookTask' + description: The task runs a notebook when the `notebook_task` field is present. + notification_settings: + $ref: '#/components/schemas/TaskNotificationSettings' + description: >- + Optional notification settings that are used when sending + notifications to each of the `email_notifications` and + `webhook_notifications` for this task run. + pipeline_task: + $ref: '#/components/schemas/PipelineTask' + description: >- + The task triggers a pipeline update when the `pipeline_task` field + is present. Only pipelines configured to use triggered more are + supported. + power_bi_task: + $ref: '#/components/schemas/PowerBiTask' + description: >- + The task triggers a Power BI semantic model update when the + `power_bi_task` field is present. + python_wheel_task: + $ref: '#/components/schemas/PythonWheelTask' + description: >- + The task runs a Python wheel when the `python_wheel_task` field is + present. + run_if: + $ref: '#/components/schemas/RunIf' + description: >- + An optional value indicating the condition that determines whether + the task should be run once its dependencies have been completed. + When omitted, defaults to `ALL_SUCCESS`. See :method:jobs/create for + a list of possible values. + run_job_task: + $ref: '#/components/schemas/RunJobTask' + description: >- + The task triggers another job when the `run_job_task` field is + present. + spark_jar_task: + $ref: '#/components/schemas/SparkJarTask' + description: The task runs a JAR when the `spark_jar_task` field is present. + spark_python_task: + $ref: '#/components/schemas/SparkPythonTask' + description: >- + The task runs a Python file when the `spark_python_task` field is + present. + spark_submit_task: + $ref: '#/components/schemas/SparkSubmitTask' + description: >- + (Legacy) The task runs the spark-submit script when the + spark_submit_task field is present. Databricks recommends using the + spark_jar_task instead; see [Spark Submit task for + jobs](/jobs/spark-submit). + sql_task: + $ref: '#/components/schemas/SqlTask' + description: >- + The task runs a SQL query or file, or it refreshes a SQL alert or a + legacy SQL dashboard when the `sql_task` field is present. + timeout_seconds: + type: integer + description: >- + An optional timeout applied to each run of this job task. A value of + `0` means no timeout. + webhook_notifications: + $ref: '#/components/schemas/WebhookNotifications' + description: >- + A collection of system notification IDs to notify when the run + begins or completes. The default behavior is to not send any system + notifications. Task webhooks respect the task notification settings. + required: + - task_key + Subscription: + type: object + properties: + custom_subject: + type: string + paused: + type: boolean + description: When true, the subscription will not send emails. + subscribers: + type: array + items: + $ref: '#/components/schemas/SubscriptionSubscriber' + description: The list of subscribers to send the snapshot of the dashboard to. + SubscriptionSubscriber: + type: object + properties: + destination_id: + type: string + user_name: + type: string + description: >- + A snapshot of the dashboard will be sent to the user's email when + the `user_name` field is present. + TableState: + type: object + properties: + has_seen_updates: + type: boolean + table_name: + type: string + description: >- + Full table name of the table to monitor, e.g. + `mycatalog.myschema.mytable` + TableTriggerState: + type: object + properties: + last_seen_table_states: + type: array + items: + $ref: '#/components/schemas/TableState' + using_scalable_monitoring: + type: boolean + description: Indicates whether the trigger is using scalable monitoring. + TableUpdateTriggerConfiguration: + type: object + properties: + table_names: + type: array + items: + type: string + condition: + $ref: '#/components/schemas/Condition' + description: The table(s) condition based on which to trigger a job run. + min_time_between_triggers_seconds: + type: integer + description: >- + If set, the trigger starts a run only after the specified amount of + time has passed since the last time the trigger fired. The minimum + allowed value is 60 seconds. + wait_after_last_change_seconds: + type: integer + description: >- + If set, the trigger starts a run only after no table updates have + occurred for the specified time and can be used to wait for a series + of table updates before triggering a run. The minimum allowed value + is 60 seconds. + required: + - table_names + Task: + type: object + properties: + task_key: + type: string + clean_rooms_notebook_task: + $ref: '#/components/schemas/CleanRoomsNotebookTask' + description: >- + The task runs a [clean rooms] notebook when the + `clean_rooms_notebook_task` field is present. [clean rooms]: + https://docs.databricks.com/clean-rooms/index.html + compute: + $ref: '#/components/schemas/Compute' + description: Task level compute configuration. + condition_task: + $ref: '#/components/schemas/ConditionTask' + description: >- + The task evaluates a condition that can be used to control the + execution of other tasks when the `condition_task` field is present. + The condition task does not require a cluster to execute and does + not support retries or notifications. + dashboard_task: + $ref: '#/components/schemas/DashboardTask' + description: The task refreshes a dashboard and sends a snapshot to subscribers. + dbt_cloud_task: + $ref: '#/components/schemas/DbtCloudTask' + description: >- + Task type for dbt cloud, deprecated in favor of the new name + dbt_platform_task + dbt_platform_task: + $ref: '#/components/schemas/DbtPlatformTask' + dbt_task: + $ref: '#/components/schemas/DbtTask' + description: >- + The task runs one or more dbt commands when the `dbt_task` field is + present. The dbt task requires both Databricks SQL and the ability + to use a serverless or a pro SQL warehouse. + depends_on: + type: array + items: + $ref: '#/components/schemas/TaskDependency' + description: >- + An optional array of objects specifying the dependency graph of the + task. All tasks specified in this field must complete before + executing this task. The task will run only if the `run_if` + condition is true. The key is `task_key`, and the value is the name + assigned to the dependent task. + description: + type: string + description: An optional description for this task. + disable_auto_optimization: + type: boolean + description: An option to disable auto optimization in serverless + disabled: + type: boolean + description: >- + An optional flag to disable the task. If set to true, the task will + not run even if it is part of a job. + email_notifications: + $ref: '#/components/schemas/TaskEmailNotifications' + description: >- + An optional set of email addresses that is notified when runs of + this task begin or complete as well as when this task is deleted. + The default behavior is to not send any emails. + environment_key: + type: string + description: >- + The key that references an environment spec in a job. This field is + required for Python script, Python wheel and dbt tasks when using + serverless compute. + existing_cluster_id: + type: string + description: >- + If existing_cluster_id, the ID of an existing cluster that is used + for all runs. When running jobs or tasks on an existing cluster, you + may need to manually restart the cluster if it stops responding. We + suggest running jobs and tasks on new clusters for greater + reliability + for_each_task: + $ref: '#/components/schemas/ForEachTask' + description: >- + The task executes a nested task for every input provided when the + `for_each_task` field is present. + gen_ai_compute_task: + $ref: '#/components/schemas/GenAiComputeTask' + health: + $ref: '#/components/schemas/JobsHealthRules' + job_cluster_key: + type: string + description: >- + If job_cluster_key, this task is executed reusing the cluster + specified in `job.settings.job_clusters`. + libraries: + type: string + description: >- + An optional list of libraries to be installed on the cluster. The + default value is an empty list. + max_retries: + type: integer + description: >- + An optional maximum number of times to retry an unsuccessful run. A + run is considered to be unsuccessful if it completes with the + `FAILED` result_state or `INTERNAL_ERROR` `life_cycle_state`. The + value `-1` means to retry indefinitely and the value `0` means to + never retry. + min_retry_interval_millis: + type: integer + description: >- + An optional minimal interval in milliseconds between the start of + the failed run and the subsequent retry run. The default behavior is + that unsuccessful runs are immediately retried. + new_cluster: + type: string + description: >- + If new_cluster, a description of a new cluster that is created for + each run. + notebook_task: + $ref: '#/components/schemas/NotebookTask' + description: The task runs a notebook when the `notebook_task` field is present. + notification_settings: + $ref: '#/components/schemas/TaskNotificationSettings' + description: >- + Optional notification settings that are used when sending + notifications to each of the `email_notifications` and + `webhook_notifications` for this task. + pipeline_task: + $ref: '#/components/schemas/PipelineTask' + description: >- + The task triggers a pipeline update when the `pipeline_task` field + is present. Only pipelines configured to use triggered more are + supported. + power_bi_task: + $ref: '#/components/schemas/PowerBiTask' + description: >- + The task triggers a Power BI semantic model update when the + `power_bi_task` field is present. + python_wheel_task: + $ref: '#/components/schemas/PythonWheelTask' + description: >- + The task runs a Python wheel when the `python_wheel_task` field is + present. + retry_on_timeout: + type: boolean + description: >- + An optional policy to specify whether to retry a job when it times + out. The default behavior is to not retry on timeout. + run_if: + $ref: '#/components/schemas/RunIf' + description: >- + An optional value specifying the condition determining whether the + task is run once its dependencies have been completed. * + `ALL_SUCCESS`: All dependencies have executed and succeeded * + `AT_LEAST_ONE_SUCCESS`: At least one dependency has succeeded * + `NONE_FAILED`: None of the dependencies have failed and at least one + was executed * `ALL_DONE`: All dependencies have been completed * + `AT_LEAST_ONE_FAILED`: At least one dependency failed * + `ALL_FAILED`: ALl dependencies have failed + run_job_task: + $ref: '#/components/schemas/RunJobTask' + description: >- + The task triggers another job when the `run_job_task` field is + present. + spark_jar_task: + $ref: '#/components/schemas/SparkJarTask' + description: The task runs a JAR when the `spark_jar_task` field is present. + spark_python_task: + $ref: '#/components/schemas/SparkPythonTask' + description: >- + The task runs a Python file when the `spark_python_task` field is + present. + spark_submit_task: + $ref: '#/components/schemas/SparkSubmitTask' + description: >- + (Legacy) The task runs the spark-submit script when the + spark_submit_task field is present. Databricks recommends using the + spark_jar_task instead; see [Spark Submit task for + jobs](/jobs/spark-submit). + sql_task: + $ref: '#/components/schemas/SqlTask' + description: >- + The task runs a SQL query or file, or it refreshes a SQL alert or a + legacy SQL dashboard when the `sql_task` field is present. + timeout_seconds: + type: integer + description: >- + An optional timeout applied to each run of this job task. A value of + `0` means no timeout. + webhook_notifications: + $ref: '#/components/schemas/WebhookNotifications' + description: >- + A collection of system notification IDs to notify when runs of this + task begin or complete. The default behavior is to not send any + system notifications. + required: + - task_key + TaskDependency: + type: object + properties: + task_key: + type: string + outcome: + type: string + description: >- + Can only be specified on condition task dependencies. The outcome of + the dependent task that must be met for this task to run. + required: + - task_key + TaskEmailNotifications: + type: object + properties: + no_alert_for_skipped_runs: + type: boolean + on_duration_warning_threshold_exceeded: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when the duration of a run + exceeds the threshold specified for the `RUN_DURATION_SECONDS` + metric in the `health` field. If no rule for the + `RUN_DURATION_SECONDS` metric is specified in the `health` field for + the job, notifications are not sent. + on_failure: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when a run unsuccessfully + completes. A run is considered to have completed unsuccessfully if + it ends with an `INTERNAL_ERROR` `life_cycle_state` or a `FAILED`, + or `TIMED_OUT` result_state. If this is not specified on job + creation, reset, or update the list is empty, and notifications are + not sent. + on_start: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when a run begins. If not + specified on job creation, reset, or update, the list is empty, and + notifications are not sent. + on_streaming_backlog_exceeded: + type: array + items: + type: string + description: >- + A list of email addresses to notify when any streaming backlog + thresholds are exceeded for any stream. Streaming backlog thresholds + can be set in the `health` field using the following metrics: + `STREAMING_BACKLOG_BYTES`, `STREAMING_BACKLOG_RECORDS`, + `STREAMING_BACKLOG_SECONDS`, or `STREAMING_BACKLOG_FILES`. Alerting + is based on the 10-minute average of these metrics. If the issue + persists, notifications are resent every 30 minutes. + on_success: + type: array + items: + type: string + description: >- + A list of email addresses to be notified when a run successfully + completes. A run is considered to have completed successfully if it + ends with a `TERMINATED` `life_cycle_state` and a `SUCCESS` + result_state. If not specified on job creation, reset, or update, + the list is empty, and notifications are not sent. + TaskNotificationSettings: + type: object + properties: + alert_on_last_attempt: + type: boolean + no_alert_for_canceled_runs: + type: boolean + description: >- + If true, do not send notifications to recipients specified in + `on_failure` if the run is canceled. + no_alert_for_skipped_runs: + type: boolean + description: >- + If true, do not send notifications to recipients specified in + `on_failure` if the run is skipped. + TerminationDetails: + type: object + properties: + code: + $ref: '#/components/schemas/TerminationCodeCode' + message: + type: string + description: >- + A descriptive message with the termination details. This field is + unstructured and the format might change. + type: + $ref: '#/components/schemas/TerminationTypeType' + TriggerInfo: + type: object + properties: + run_id: + type: integer + description: The run id of the Run Job task run + description: Additional details about what triggered the run + TriggerSettings: + type: object + properties: + file_arrival: + $ref: '#/components/schemas/FileArrivalTriggerConfiguration' + model: + $ref: '#/components/schemas/ModelTriggerConfiguration' + pause_status: + $ref: '#/components/schemas/PauseStatus' + description: Whether this trigger is paused or not. + periodic: + $ref: '#/components/schemas/PeriodicTriggerConfiguration' + description: Periodic trigger settings. + table_update: + $ref: '#/components/schemas/TableUpdateTriggerConfiguration' + TriggerStateProto: + type: object + properties: + file_arrival: + $ref: '#/components/schemas/FileArrivalTriggerState' + table: + $ref: '#/components/schemas/TableTriggerState' + ViewItem: + type: object + properties: + content: + type: string + name: + type: string + description: >- + Name of the view item. In the case of code view, it would be the + notebook’s name. In the case of dashboard view, it would be the + dashboard’s name. + type: + $ref: '#/components/schemas/ViewType' + description: Type of the view item. + Webhook: + type: object + properties: + id: + type: string + required: + - id + WebhookNotifications: + type: object + properties: + on_duration_warning_threshold_exceeded: + type: array + items: + $ref: '#/components/schemas/Webhook' + on_failure: + type: array + items: + $ref: '#/components/schemas/Webhook' + description: >- + An optional list of system notification IDs to call when the run + fails. A maximum of 3 destinations can be specified for the + `on_failure` property. + on_start: + type: array + items: + $ref: '#/components/schemas/Webhook' + description: >- + An optional list of system notification IDs to call when the run + starts. A maximum of 3 destinations can be specified for the + `on_start` property. + on_streaming_backlog_exceeded: + type: array + items: + $ref: '#/components/schemas/Webhook' + description: >- + An optional list of system notification IDs to call when any + streaming backlog thresholds are exceeded for any stream. Streaming + backlog thresholds can be set in the `health` field using the + following metrics: `STREAMING_BACKLOG_BYTES`, + `STREAMING_BACKLOG_RECORDS`, `STREAMING_BACKLOG_SECONDS`, or + `STREAMING_BACKLOG_FILES`. Alerting is based on the 10-minute + average of these metrics. If the issue persists, notifications are + resent every 30 minutes. A maximum of 3 destinations can be + specified for the `on_streaming_backlog_exceeded` property. + on_success: + type: array + items: + $ref: '#/components/schemas/Webhook' + description: >- + An optional list of system notification IDs to call when the run + completes successfully. A maximum of 3 destinations can be specified + for the `on_success` property. + WidgetErrorDetail: + type: object + properties: + message: + type: string + AuthenticationMethod: + type: string + enum: + - OAUTH + - PAT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CleanRoomTaskRunLifeCycleState: + type: string + enum: + - BLOCKED + - INTERNAL_ERROR + - PENDING + - QUEUED + - RUNNING + - RUN_LIFE_CYCLE_STATE_UNSPECIFIED + - SKIPPED + - TERMINATED + - TERMINATING + - WAITING_FOR_RETRY + description: >- + Copied from elastic-spark-common/api/messages/runs.proto. Using the + original definition to + + remove coupling with jobs API definition + CleanRoomTaskRunResultState: + type: string + enum: + - CANCELED + - DISABLED + - EVICTED + - EXCLUDED + - FAILED + - MAXIMUM_CONCURRENT_RUNS_REACHED + - RUN_RESULT_STATE_UNSPECIFIED + - SUCCESS + - SUCCESS_WITH_FAILURES + - TIMEDOUT + - UPSTREAM_CANCELED + - UPSTREAM_EVICTED + - UPSTREAM_FAILED + description: >- + Copied from elastic-spark-common/api/messages/runs.proto. Using the + original definition to avoid + + cyclic dependency. + Condition: + type: string + enum: + - ALL_UPDATED + - ANY_UPDATED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ConditionTaskOp: + type: string + enum: + - EQUAL_TO + - GREATER_THAN + - GREATER_THAN_OR_EQUAL + - LESS_THAN + - LESS_THAN_OR_EQUAL + - NOT_EQUAL + description: >- + * `EQUAL_TO`, `NOT_EQUAL` operators perform string comparison of their + operands. This means that + + `“12.0” == “12”` will evaluate to `false`. * `GREATER_THAN`, + `GREATER_THAN_OR_EQUAL`, + + `LESS_THAN`, `LESS_THAN_OR_EQUAL` operators perform numeric comparison + of their operands. + + `“12.0” >= “12”` will evaluate to `true`, `“10.0” >= “12”` will evaluate + to + + `false`. + + + The boolean comparison to task values can be implemented with operators + `EQUAL_TO`, `NOT_EQUAL`. + + If a task value was set to a boolean value, it will be serialized to + `“true”` or + + `“false”` for the comparison. + DbtPlatformRunStatus: + type: string + enum: + - CANCELLED + - ERROR + - QUEUED + - RUNNING + - STARTING + - SUCCESS + description: >- + Response enumeration from calling the dbt platform API, for inclusion in + output + Format: + type: string + enum: + - MULTI_TASK + - SINGLE_TASK + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + GitProvider: + type: string + enum: + - awsCodeCommit + - azureDevOpsServices + - bitbucketCloud + - bitbucketServer + - gitHub + - gitHubEnterprise + - gitLab + - gitLabEnterpriseEdition + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + JobDeploymentKind: + type: string + enum: + - BUNDLE + - SYSTEM_MANAGED + description: >- + * `BUNDLE`: The job is managed by Databricks Asset Bundle. * + `SYSTEM_MANAGED`: The job is + + managed by Databricks and is read-only. + JobEditMode: + type: string + enum: + - EDITABLE + - UI_LOCKED + description: >- + Edit mode of the job. + + + * `UI_LOCKED`: The job is in a locked UI state and cannot be modified. * + `EDITABLE`: The job is + + in an editable state and can be modified. + JobPermissionLevel: + type: string + enum: + - CAN_MANAGE + - CAN_MANAGE_RUN + - CAN_VIEW + - IS_OWNER + description: Permission level + JobSourceDirtyState: + type: string + enum: + - DISCONNECTED + - NOT_SYNCED + description: >- + Dirty state indicates the job is not fully synced with the job + specification in the remote + + repository. + + + Possible values are: * `NOT_SYNCED`: The job is not yet synced with the + remote job + + specification. Import the remote job specification from UI to make the + job fully synced. * + + `DISCONNECTED`: The job is temporary disconnected from the remote job + specification and is + + allowed for live edit. Import the remote job specification again from UI + to make the job fully + + synced. + JobsHealthMetric: + type: string + enum: + - RUN_DURATION_SECONDS + - STREAMING_BACKLOG_BYTES + - STREAMING_BACKLOG_FILES + - STREAMING_BACKLOG_RECORDS + - STREAMING_BACKLOG_SECONDS + description: >- + Specifies the health metric that is being evaluated for a particular + health rule. + + + * `RUN_DURATION_SECONDS`: Expected total time for a run in seconds. * + `STREAMING_BACKLOG_BYTES`: + + An estimate of the maximum bytes of data waiting to be consumed across + all streams. This metric + + is in Public Preview. * `STREAMING_BACKLOG_RECORDS`: An estimate of the + maximum offset lag + + across all streams. This metric is in Public Preview. * + `STREAMING_BACKLOG_SECONDS`: An estimate + + of the maximum consumer delay across all streams. This metric is in + Public Preview. * + + `STREAMING_BACKLOG_FILES`: An estimate of the maximum number of + outstanding files across all + + streams. This metric is in Public Preview. + JobsHealthOperator: + type: string + enum: + - GREATER_THAN + description: >- + Specifies the operator used to compare the health metric value with the + specified threshold. + ModelTriggerConfigurationCondition: + type: string + enum: + - MODEL_ALIAS_SET + - MODEL_CREATED + - MODEL_VERSION_READY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PauseStatus: + type: string + enum: + - PAUSED + - UNPAUSED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PerformanceTarget: + type: string + enum: + - PERFORMANCE_OPTIMIZED + - STANDARD + description: >- + PerformanceTarget defines how performant (lower latency) or cost + efficient the execution of run + + on serverless compute should be. The performance mode on the job or + pipeline should map to a + + performance setting that is passed to Cluster Manager (see + cluster-common PerformanceTarget). + PeriodicTriggerConfigurationTimeUnit: + type: string + enum: + - DAYS + - HOURS + - WEEKS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + QueueDetailsCodeCode: + type: string + enum: + - ACTIVE_RUNS_LIMIT_REACHED + - ACTIVE_RUN_JOB_TASKS_LIMIT_REACHED + - MAX_CONCURRENT_RUNS_REACHED + description: >- + The reason for queuing the run. * `ACTIVE_RUNS_LIMIT_REACHED`: The run + was queued due to + + reaching the workspace limit of active task runs. * + `MAX_CONCURRENT_RUNS_REACHED`: The run was + + queued due to reaching the per-job limit of concurrent job runs. * + + `ACTIVE_RUN_JOB_TASKS_LIMIT_REACHED`: The run was queued due to reaching + the workspace limit of + + active run job tasks. + RepairHistoryItemType: + type: string + enum: + - ORIGINAL + - REPAIR + description: >- + The repair history item type. Indicates whether a run is the original + run or a repair run. + RunIf: + type: string + enum: + - ALL_DONE + - ALL_FAILED + - ALL_SUCCESS + - AT_LEAST_ONE_FAILED + - AT_LEAST_ONE_SUCCESS + - NONE_FAILED + description: >- + An optional value indicating the condition that determines whether the + task should be run once + + its dependencies have been completed. When omitted, defaults to + `ALL_SUCCESS`. + + + Possible values are: * `ALL_SUCCESS`: All dependencies have executed and + succeeded * + + `AT_LEAST_ONE_SUCCESS`: At least one dependency has succeeded * + `NONE_FAILED`: None of the + + dependencies have failed and at least one was executed * `ALL_DONE`: All + dependencies have been + + completed * `AT_LEAST_ONE_FAILED`: At least one dependency failed * + `ALL_FAILED`: ALl + + dependencies have failed + RunLifeCycleState: + type: string + enum: + - BLOCKED + - INTERNAL_ERROR + - PENDING + - QUEUED + - RUNNING + - SKIPPED + - TERMINATED + - TERMINATING + - WAITING_FOR_RETRY + description: >- + A value indicating the run's lifecycle state. The possible values are: * + `QUEUED`: The run is + + queued. * `PENDING`: The run is waiting to be executed while the cluster + and execution context + + are being prepared. * `RUNNING`: The task of this run is being executed. + * `TERMINATING`: The + + task of this run has completed, and the cluster and execution context + are being cleaned up. * + + `TERMINATED`: The task of this run has completed, and the cluster and + execution context have + + been cleaned up. This state is terminal. * `SKIPPED`: This run was + aborted because a previous + + run of the same job was already active. This state is terminal. * + `INTERNAL_ERROR`: An + + exceptional state that indicates a failure in the Jobs service, such as + network failure over a + + long period. If a run on a new cluster ends in the `INTERNAL_ERROR` + state, the Jobs service + + terminates the cluster as soon as possible. This state is terminal. * + `BLOCKED`: The run is + + blocked on an upstream dependency. * `WAITING_FOR_RETRY`: The run is + waiting for a retry. + RunLifecycleStateV2State: + type: string + enum: + - BLOCKED + - PENDING + - QUEUED + - RUNNING + - TERMINATED + - TERMINATING + - WAITING + description: The current state of the run. + RunResultState: + type: string + enum: + - CANCELED + - DISABLED + - EXCLUDED + - FAILED + - MAXIMUM_CONCURRENT_RUNS_REACHED + - SUCCESS + - SUCCESS_WITH_FAILURES + - TIMEDOUT + - UPSTREAM_CANCELED + - UPSTREAM_FAILED + description: >- + A value indicating the run's result. The possible values are: * + `SUCCESS`: The task completed + + successfully. * `FAILED`: The task completed with an error. * + `TIMEDOUT`: The run was stopped + + after reaching the timeout. * `CANCELED`: The run was canceled at user + request. * + + `MAXIMUM_CONCURRENT_RUNS_REACHED`: The run was skipped because the + maximum concurrent runs were + + reached. * `EXCLUDED`: The run was skipped because the necessary + conditions were not met. * + + `SUCCESS_WITH_FAILURES`: The job run completed successfully with some + failures; leaf tasks were + + successful. * `UPSTREAM_FAILED`: The run was skipped because of an + upstream failure. * + + `UPSTREAM_CANCELED`: The run was skipped because an upstream task was + canceled. * `DISABLED`: + + The run was skipped because it was disabled explicitly by the user. + RunType: + type: string + enum: + - JOB_RUN + - SUBMIT_RUN + - WORKFLOW_RUN + description: >- + The type of a run. * `JOB_RUN`: Normal job run. A run created with + :method:jobs/runNow. * + + `WORKFLOW_RUN`: Workflow run. A run created with [dbutils.notebook.run]. + * `SUBMIT_RUN`: Submit + + run. A run created with :method:jobs/submit. + + + [dbutils.notebook.run]: + https://docs.databricks.com/dev-tools/databricks-utils.html#dbutils-workflow + Source: + type: string + enum: + - GIT + - WORKSPACE + description: >- + Optional location type of the SQL file. When set to `WORKSPACE`, the SQL + file will be retrieved from the local Databricks workspace. When set + to `GIT`, the SQL file will be retrieved from a + + Git repository defined in `git_source`. If the value is empty, the task + will use `GIT` if + + `git_source` is defined and `WORKSPACE` otherwise. + + + * `WORKSPACE`: SQL file is located in Databricks workspace. * `GIT`: SQL + file is located in + + cloud Git provider. + SqlAlertState: + type: string + enum: + - OK + - TRIGGERED + - UNKNOWN + description: >- + The state of the SQL alert. + + + * UNKNOWN: alert yet to be evaluated * OK: alert evaluated and did not + fulfill trigger + + conditions * TRIGGERED: alert evaluated and fulfilled trigger conditions + SqlDashboardWidgetOutputStatus: + type: string + enum: + - CANCELLED + - FAILED + - PENDING + - RUNNING + - SUCCESS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + StorageMode: + type: string + enum: + - DIRECT_QUERY + - DUAL + - IMPORT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TaskRetryMode: + type: string + enum: + - NEVER + - ON_FAILURE + description: >- + task retry mode of the continuous job * NEVER: The failed task will not + be retried. * + + ON_FAILURE: Retry a failed task if at least one other task in the job is + still running its first + + attempt. When this condition is no longer met or the retry limit is + reached, the job run is + + cancelled and a new run is started. + TerminationCodeCode: + type: string + enum: + - BUDGET_POLICY_LIMIT_EXCEEDED + - CANCELED + - CLOUD_FAILURE + - CLUSTER_ERROR + - CLUSTER_REQUEST_LIMIT_EXCEEDED + - DISABLED + - DRIVER_ERROR + - FEATURE_DISABLED + - INTERNAL_ERROR + - INVALID_CLUSTER_REQUEST + - INVALID_RUN_CONFIGURATION + - LIBRARY_INSTALLATION_ERROR + - MAX_CONCURRENT_RUNS_EXCEEDED + - MAX_JOB_QUEUE_SIZE_EXCEEDED + - MAX_SPARK_CONTEXTS_EXCEEDED + - REPOSITORY_CHECKOUT_FAILED + - RESOURCE_NOT_FOUND + - RUN_EXECUTION_ERROR + - SKIPPED + - STORAGE_ACCESS_ERROR + - SUCCESS + - SUCCESS_WITH_FAILURES + - UNAUTHORIZED_ERROR + - USER_CANCELED + - WORKSPACE_RUN_LIMIT_EXCEEDED + description: >- + The code indicates why the run was terminated. Additional codes might be + introduced in future + + releases. * `SUCCESS`: The run was completed successfully. * + `SUCCESS_WITH_FAILURES`: The run + + was completed successfully but some child runs failed. * + `USER_CANCELED`: The run was + + successfully canceled during execution by a user. * `CANCELED`: The run + was canceled during + + execution by the Databricks platform; for example, if the maximum run + duration was exceeded. * + + `SKIPPED`: Run was never executed, for example, if the upstream task run + failed, the dependency + + type condition was not met, or there were no material tasks to execute. + * `INTERNAL_ERROR`: The + + run encountered an unexpected error. Refer to the state message for + further details. * + + `DRIVER_ERROR`: The run encountered an error while communicating with + the Spark Driver. * + + `CLUSTER_ERROR`: The run failed due to a cluster error. Refer to the + state message for further + + details. * `REPOSITORY_CHECKOUT_FAILED`: Failed to complete the checkout + due to an error when + + communicating with the third party service. * `INVALID_CLUSTER_REQUEST`: + The run failed because + + it issued an invalid request to start the cluster. * + `WORKSPACE_RUN_LIMIT_EXCEEDED`: The + + workspace has reached the quota for the maximum number of concurrent + active runs. Consider + + scheduling the runs over a larger time frame. * `FEATURE_DISABLED`: The + run failed because it + + tried to access a feature unavailable for the workspace. * + `CLUSTER_REQUEST_LIMIT_EXCEEDED`: The + + number of cluster creation, start, and upsize requests have exceeded the + allotted rate limit. + + Consider spreading the run execution over a larger time frame. * + `STORAGE_ACCESS_ERROR`: The run + + failed due to an error when accessing the customer blob storage. Refer + to the state message for + + further details. * `RUN_EXECUTION_ERROR`: The run was completed with + task failures. For more + + details, refer to the state message or run output. * + `UNAUTHORIZED_ERROR`: The run failed due to + + a permission issue while accessing a resource. Refer to the state + message for further details. * + + `LIBRARY_INSTALLATION_ERROR`: The run failed while installing the + user-requested library. Refer + + to the state message for further details. The causes might include, but + are not limited to: The + + provided library is invalid, there are insufficient permissions to + install the library, and so + + forth. * `MAX_CONCURRENT_RUNS_EXCEEDED`: The scheduled run exceeds the + limit of maximum + + concurrent runs set for the job. * `MAX_SPARK_CONTEXTS_EXCEEDED`: The + run is scheduled on a + + cluster that has already reached the maximum number of contexts it is + configured to create. See: + + [Link]. * `RESOURCE_NOT_FOUND`: A resource necessary for run execution + does not exist. Refer to + + the state message for further details. * `INVALID_RUN_CONFIGURATION`: + The run failed due to an + + invalid configuration. Refer to the state message for further details. * + `CLOUD_FAILURE`: The + + run failed due to a cloud provider issue. Refer to the state message for + further details. * + + `MAX_JOB_QUEUE_SIZE_EXCEEDED`: The run was skipped due to reaching the + job level queue size + + limit. * `DISABLED`: The run was never executed because it was disabled + explicitly by the user. + + * `BREAKING_CHANGE`: Run failed because of an intentional breaking + change in Spark, but it will + + be retried with a mitigation config. + + + [Link]: + https://kb.databricks.com/en_US/notebooks/too-many-execution-contexts-are-open-right-now + TerminationTypeType: + type: string + enum: + - CLIENT_ERROR + - CLOUD_FAILURE + - INTERNAL_ERROR + - SUCCESS + description: >- + * `SUCCESS`: The run terminated without any issues * `INTERNAL_ERROR`: + An error occurred in the + + Databricks platform. Please look at the [status page] or contact support + if the issue persists. + + * `CLIENT_ERROR`: The run was terminated because of an error caused by + user input or the job + + configuration. * `CLOUD_FAILURE`: The run was terminated because of an + issue with your cloud + + provider. + + + [status page]: https://status.databricks.com/ + TriggerType: + type: string + enum: + - CONTINUOUS + - CONTINUOUS_RESTART + - FILE_ARRIVAL + - ONE_TIME + - PERIODIC + - RETRY + - RUN_JOB_TASK + - TABLE + description: >- + The type of trigger that fired this run. + + + * `PERIODIC`: Schedules that periodically trigger runs, such as a cron + scheduler. * `ONE_TIME`: + + One time triggers that fire a single run. This occurs you triggered a + single run on demand + + through the UI or the API. * `RETRY`: Indicates a run that is triggered + as a retry of a + + previously failed run. This occurs when you request to re-run the job in + case of failures. * + + `RUN_JOB_TASK`: Indicates a run that is triggered using a Run Job task. + * `FILE_ARRIVAL`: + + Indicates a run that is triggered by a file arrival. * `CONTINUOUS`: + Indicates a run that is + + triggered by a continuous job. * `TABLE`: Indicates a run that is + triggered by a table update. * + + `CONTINUOUS_RESTART`: Indicates a run created by user to manually + restart a continuous job run. + + * `MODEL`: Indicates a run that is triggered by a model update. + ViewType: + type: string + enum: + - DASHBOARD + - NOTEBOOK + description: '* `NOTEBOOK`: Notebook view item. * `DASHBOARD`: Dashboard view item.' + ViewsToExport: + type: string + enum: + - ALL + - CODE + - DASHBOARDS + description: >- + * `CODE`: Code view of the notebook. * `DASHBOARDS`: All dashboard views + of the notebook. * + + `ALL`: All views of the notebook. + x-stackQL-resources: + job_runs: + id: databricks_workspace.jobs.job_runs + name: job_runs + title: Job Runs + methods: + cancel_all: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1cancel-all/post' + response: + mediaType: application/json + openAPIDocKey: '200' + cancel: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1cancel/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + export: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1export/get' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.runs + repair: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1repair/post' + response: + mediaType: application/json + openAPIDocKey: '200' + submit: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1submit/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/job_runs/methods/get' + - $ref: '#/components/x-stackQL-resources/job_runs/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/job_runs/methods/submit' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/job_runs/methods/delete' + replace: [] + jobs: + id: databricks_workspace.jobs.jobs + name: jobs + title: Jobs + methods: + create: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.jobs + reset: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1reset/post' + response: + mediaType: application/json + openAPIDocKey: '200' + run_now: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1run-now/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1update/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/jobs/methods/get' + - $ref: '#/components/x-stackQL-resources/jobs/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/jobs/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/jobs/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/jobs/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/jobs/methods/reset' + job_permission_levels: + id: databricks_workspace.jobs.job_permission_levels + name: job_permission_levels + title: Job Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1jobs~1{job_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/job_permission_levels/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + job_permissions: + id: databricks_workspace.jobs.job_permissions + name: job_permissions + title: Job Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1jobs~1{job_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1jobs~1{job_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1jobs~1{job_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/job_permissions/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/job_permissions/methods/update' + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/job_permissions/methods/set' + job_run_output: + id: databricks_workspace.jobs.job_run_output + name: job_run_output + title: Job Run Output + methods: + get_output: + operation: + $ref: '#/paths/~1api~12.2~1jobs~1runs~1get-output/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/job_run_output/methods/get_output' + insert: [] + update: [] + delete: [] + replace: [] + policy_compliance_for_jobs: + id: databricks_workspace.jobs.policy_compliance_for_jobs + name: policy_compliance_for_jobs + title: Policy Compliance For Jobs + methods: + enforce: + operation: + $ref: '#/paths/~1api~12.0~1policies~1jobs~1enforce-compliance/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1policies~1jobs~1get-compliance/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1policies~1jobs~1list-compliance/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.jobs + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/policy_compliance_for_jobs/methods/get + - $ref: >- + #/components/x-stackQL-resources/policy_compliance_for_jobs/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/policy_compliance_for_jobs/methods/enforce + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/lakeview.yaml b/providers/src/databricks_workspace/v00.00.00000/services/lakeview.yaml deleted file mode 100644 index 496d9be8..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/lakeview.yaml +++ /dev/null @@ -1,1270 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Lakeview API - description: Lakeview -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/lakeview/dashboards: - post: - operationId: dashboards-create - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/create - x-stackQL-resource: dashboards - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - dashboard_id: - type: string - display_name: - type: string - path: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - etag: - type: string - serialized_dashboard: - type: string - lifecycle_state: - type: string - parent_path: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - display_name: - type: string - warehouse_id: - type: string - etag: - type: string - serialized_dashboard: - type: string - parent_path: - type: string - example: - display_name: Monthly Traffic Report - warehouse_id: 47bb1c472649e711 - etag: '80611980' - serialized_dashboard: '{"pages":[{"name":"b532570b","displayName":"New - Page"}]}' - parent_path: /path/to/dir - description: Create a draft dashboard. - get: - operationId: dashboards-list - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/list - x-stackQL-resource: dashboards - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_size - in: query - - name: page_token - in: query - - name: show_trashed - in: query - - name: view - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - dashboards: - type: array - items: - type: object - properties: - dashboard_id: - type: string - display_name: - type: string - lifecycle_state: - type: string - warehouse_id: - type: string - create_time: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.dashboards - /api/2.0/lakeview/dashboards/{dashboard_id}: - get: - operationId: dashboards-get - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/get - x-stackQL-resource: dashboards - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - dashboard_id: - type: string - display_name: - type: string - path: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - etag: - type: string - serialized_dashboard: - type: string - lifecycle_state: - type: string - parent_path: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get a draft dashboard. - delete: - operationId: dashboards-trash - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/trash - x-stackQL-resource: dashboards - x-stackQL-method: trash - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Trash a dashboard. - patch: - operationId: dashboards-update - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/update - x-stackQL-resource: dashboards - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - dashboard_id: - type: string - display_name: - type: string - path: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - etag: - type: string - serialized_dashboard: - type: string - lifecycle_state: - type: string - parent_path: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - display_name: - type: string - warehouse_id: - type: string - etag: - type: string - serialized_dashboard: - type: string - example: - display_name: Monthly Traffic Report - warehouse_id: 47bb1c472649e711 - etag: '80611980' - serialized_dashboard: '{"pages":[{"name":"b532570b","displayName":"New - Page"}]}' - description: Update a draft dashboard. - /api/2.0/lakeview/dashboards/migrate: - post: - operationId: dashboards-migrate - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/migrate - x-stackQL-resource: dashboards - x-stackQL-method: migrate - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - dashboard_id: - type: string - display_name: - type: string - path: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - etag: - type: string - serialized_dashboard: - type: string - lifecycle_state: - type: string - parent_path: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - source_dashboard_id: - type: required - display_name: - type: string - parent_path: - type: string - update_parameter_syntax: - type: string - example: - source_dashboard_id: 4e443c27-9f61-4f2e-a12d-ea5668460bf1 - display_name: Monthly Traffic Report - parent_path: /path/to/dir - update_parameter_syntax: true - description: Migrates a classic SQL dashboard to Lakeview. - /api/2.0/lakeview/dashboards/{dashboard_id}/published: - post: - operationId: dashboards-publish - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/publish - x-stackQL-resource: dashboards - x-stackQL-method: publish - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - display_name: - type: string - warehouse_id: - type: string - embed_credentials: - type: boolean - revision_create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - embed_credentials: - type: boolean - warehouse_id: - type: string - example: - embed_credentials: true - warehouse_id: string - description: Publish the current draft dashboard. - delete: - operationId: dashboards-unpublish - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/unpublish - x-stackQL-resource: dashboards - x-stackQL-method: unpublish - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Unpublish the dashboard. - get: - operationId: dashboards-published-getpublished - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/getpublished - x-stackQL-resource: dashboards_published - x-stackQL-method: getpublished - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - display_name: - type: string - warehouse_id: - type: string - embed_credentials: - type: boolean - revision_create_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get the current published dashboard. - /api/2.0/lakeview/dashboards/{dashboard_id}/schedules: - post: - operationId: dashboard-schedules-createschedule - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/createschedule - x-stackQL-resource: dashboard_schedules - x-stackQL-method: createschedule - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - schedule_id: - type: string - dashboard_id: - type: string - cron_schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - display_name: - type: string - etag: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - pause_status: - type: string - display_name: - type: string - etag: - type: string - warehouse_id: - type: string - cron_schedule: - type: object - example: - cron_schedule: - quartz_cron_expression: 0 0 8 * * ? - timezone_id: Europe/London - pause_status: UNPAUSED - display_name: Monthly Traffic Report Snapshot - etag: '80611980' - warehouse_id: 47bb1c472649e711 - get: - operationId: dashboard-schedules-listschedules - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/listschedules - x-stackQL-resource: dashboard_schedules - x-stackQL-method: listschedules - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: dashboard_id - required: true - description: string - in: path - - name: page_size - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - schedules: - type: array - items: - type: object - properties: - schedule_id: - type: string - dashboard_id: - type: string - cron_schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - display_name: - type: string - etag: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.schedules - /api/2.0/lakeview/dashboards/{dashboard_id}/schedules/{schedule_id}: - delete: - operationId: dashboard-schedules-deleteschedule - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/deleteschedule - x-stackQL-resource: dashboard_schedules - x-stackQL-method: deleteschedule - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: dashboard_id - required: true - description: string - in: path - - name: schedule_id - required: true - description: string - in: path - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - get: - operationId: dashboard-schedules-getschedule - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/getschedule - x-stackQL-resource: dashboard_schedules - x-stackQL-method: getschedule - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: dashboard_id - required: true - description: string - in: path - - name: schedule_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - schedule_id: - type: string - dashboard_id: - type: string - cron_schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - display_name: - type: string - etag: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - put: - operationId: dashboard-schedules-updateschedule - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/updateschedule - x-stackQL-resource: dashboard_schedules - x-stackQL-method: updateschedule - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - schedule_id: - type: string - dashboard_id: - type: string - cron_schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - display_name: - type: string - etag: - type: string - create_time: - type: string - update_time: - type: string - warehouse_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - pause_status: - type: string - display_name: - type: string - etag: - type: string - warehouse_id: - type: string - cron_schedule: - type: object - example: - cron_schedule: - quartz_cron_expression: 0 0 8 * * ? - timezone_id: Europe/London - pause_status: UNPAUSED - display_name: Monthly Traffic Report Snapshot - etag: '80611980' - warehouse_id: 47bb1c472649e711 - /api/2.0/lakeview/dashboards/{dashboard_id}/schedules/{schedule_id}/subscriptions: - post: - operationId: dashboard-subscriptions-createsubscription - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/createsubscription - x-stackQL-resource: dashboard_subscriptions - x-stackQL-method: createsubscription - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - subscription_id: - type: string - schedule_id: - type: string - dashboard_id: - type: string - subscriber: - type: object - properties: - user_subscriber: - type: object - properties: - user_id: - type: string - created_by_user_id: - type: string - etag: - type: string - create_time: - type: string - update_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - etag: - type: string - subscriber: - type: required - example: - subscriber: - user_subscriber: - user_id: '3294322584244938' - etag: '80611980' - get: - operationId: dashboard-subscriptions-listsubscriptions - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/listsubscriptions - x-stackQL-resource: dashboard_subscriptions - x-stackQL-method: listsubscriptions - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: dashboard_id - required: true - description: string - in: path - - name: schedule_id - required: true - description: string - in: path - - name: page_size - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - subscriptions: - type: array - items: - type: object - properties: - subscription_id: - type: string - schedule_id: - type: string - dashboard_id: - type: string - subscriber: - type: object - properties: - user_subscriber: - type: object - properties: - user_id: - type: string - created_by_user_id: - type: string - etag: - type: string - create_time: - type: string - update_time: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.subscriptions - /api/2.0/lakeview/dashboards/{dashboard_id}/schedules/{schedule_id}/subscriptions/{subscription_id}: - delete: - operationId: dashboard-subscriptions-deletesubscription - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/deletesubscription - x-stackQL-resource: dashboard_subscriptions - x-stackQL-method: deletesubscription - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: dashboard_id - required: true - description: string - in: path - - name: schedule_id - required: true - description: string - in: path - - name: subscription_id - required: true - description: string - in: path - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - get: - operationId: dashboard-subscriptions-getsubscription - externalDocs: - url: https://docs.databricks.com/api/workspace/lakeview/getsubscription - x-stackQL-resource: dashboard_subscriptions - x-stackQL-method: getsubscription - x-stackQL-verb: select - x-numReqParams: 3 - parameters: - - name: dashboard_id - required: true - description: string - in: path - - name: schedule_id - required: true - description: string - in: path - - name: subscription_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - subscription_id: - type: string - schedule_id: - type: string - dashboard_id: - type: string - subscriber: - type: object - properties: - user_subscriber: - type: object - properties: - user_id: - type: string - created_by_user_id: - type: string - etag: - type: string - create_time: - type: string - update_time: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. -components: - x-stackQL-resources: - dashboards: - id: databricks_workspace.lakeview.dashboards - name: dashboards - title: Dashboards - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.dashboards - get: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - trash: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - migrate: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1migrate/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - publish: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1published/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - unpublish: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1published/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/get' - - $ref: '#/components/x-stackQL-resources/dashboards/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/trash' - exec: - - $ref: '#/components/x-stackQL-resources/dashboards/methods/unpublish' - - $ref: '#/components/x-stackQL-resources/dashboards/methods/publish' - - $ref: '#/components/x-stackQL-resources/dashboards/methods/migrate' - dashboards_published: - id: databricks_workspace.lakeview.dashboards_published - name: dashboards_published - title: Dashboards_published - methods: - getpublished: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1published/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dashboards_published/methods/getpublished' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - dashboard_schedules: - id: databricks_workspace.lakeview.dashboard_schedules - name: dashboard_schedules - title: Dashboard_schedules - methods: - createschedule: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listschedules: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.schedules - deleteschedule: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getschedule: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updateschedule: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dashboard_schedules/methods/getschedule' - - $ref: '#/components/x-stackQL-resources/dashboard_schedules/methods/listschedules' - insert: - - $ref: '#/components/x-stackQL-resources/dashboard_schedules/methods/createschedule' - update: - - $ref: '#/components/x-stackQL-resources/dashboard_schedules/methods/updateschedule' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dashboard_schedules/methods/deleteschedule' - exec: [] - dashboard_subscriptions: - id: databricks_workspace.lakeview.dashboard_subscriptions - name: dashboard_subscriptions - title: Dashboard_subscriptions - methods: - createsubscription: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listsubscriptions: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.subscriptions - deletesubscription: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions~1{subscription_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getsubscription: - operation: - $ref: '#/paths/~1api~12.0~1lakeview~1dashboards~1{dashboard_id}~1schedules~1{schedule_id}~1subscriptions~1{subscription_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dashboard_subscriptions/methods/getsubscription' - - $ref: '#/components/x-stackQL-resources/dashboard_subscriptions/methods/listsubscriptions' - insert: - - $ref: '#/components/x-stackQL-resources/dashboard_subscriptions/methods/createsubscription' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dashboard_subscriptions/methods/deletesubscription' - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/machinelearning.yaml b/providers/src/databricks_workspace/v00.00.00000/services/machinelearning.yaml deleted file mode 100644 index b80f95ec..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/machinelearning.yaml +++ /dev/null @@ -1,4891 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Machinelearning API - description: Machinelearning -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/mlflow/experiments/create: - post: - operationId: experiments-createexperiment - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/createexperiment - x-stackQL-resource: experiments - x-stackQL-method: createexperiment - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - artifact_location: - type: string - tags: - type: Array of object - example: - name: string - artifact_location: string - tags: - - key: string - value: string - description: Creates an experiment with a name. Returns the ID of the newly - created experiment. Validates that another experiment with the same name does - not already exist and fails if another experiment with the same name already - exists. - /api/2.0/mlflow/experiments/delete: - post: - operationId: experiments-deleteexperiment - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/deleteexperiment - x-stackQL-resource: experiments - x-stackQL-method: deleteexperiment - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: required - example: - experiment_id: string - description: Marks an experiment and associated metadata, runs, metrics, params, - and tags for deletion. If the experiment uses FileStore, artifacts associated - with experiment are also deleted. - /api/2.0/mlflow/experiments/get-by-name: - get: - operationId: experiments-getbyname - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/getbyname - x-stackQL-resource: experiments - x-stackQL-method: getbyname - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: experiment_name - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - experiment: - type: object - properties: - experiment_id: - type: string - name: - type: string - artifact_location: - type: string - lifecycle_stage: - type: string - last_update_time: - type: integer - creation_time: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.experiment - description: Gets metadata for an experiment. - /api/2.0/mlflow/experiments/get: - get: - operationId: experiments-getexperiment - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/getexperiment - x-stackQL-resource: experiments - x-stackQL-method: getexperiment - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: experiment_id - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - experiment: - type: object - properties: - experiment_id: - type: string - name: - type: string - artifact_location: - type: string - lifecycle_stage: - type: string - last_update_time: - type: integer - creation_time: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.experiment - description: Gets metadata for an experiment. This method works on deleted experiments. - /api/2.0/mlflow/experiments/list: - get: - operationId: experiments-listexperiments - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/listexperiments - x-stackQL-resource: experiments - x-stackQL-method: listexperiments - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: view_type - in: query - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - experiments: - type: array - items: - type: object - properties: - experiment_id: - type: string - name: - type: string - artifact_location: - type: string - lifecycle_stage: - type: string - last_update_time: - type: integer - creation_time: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.experiments - description: Gets a list of all experiments. - /api/2.0/mlflow/experiments/restore: - post: - operationId: experiments-restoreexperiment - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/restoreexperiment - x-stackQL-resource: experiments - x-stackQL-method: restoreexperiment - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: required - example: - experiment_id: string - description: Restore an experiment marked for deletion. This also restores associated - metadata, runs, metrics, params, and tags. If experiment uses FileStore, underlying - artifacts associated with experiment are also restored. - /api/2.0/mlflow/experiments/search: - post: - operationId: experiments-searchexperiments - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/searchexperiments - x-stackQL-resource: experiments - x-stackQL-method: searchexperiments - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - experiments: - type: array - items: - type: object - properties: - experiment_id: - type: string - name: - type: string - artifact_location: - type: string - lifecycle_stage: - type: string - last_update_time: - type: integer - creation_time: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.experiments - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - max_results: - type: integer - format: int64 - page_token: - type: string - filter: - type: string - order_by: - type: Array of string - view_type: - type: string - example: - max_results: 0 - page_token: string - filter: string - order_by: - - string - view_type: ACTIVE_ONLY - description: Searches for experiments that satisfy specified search criteria. - /api/2.0/mlflow/experiments/update: - post: - operationId: experiments-updateexperiment - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/updateexperiment - x-stackQL-resource: experiments - x-stackQL-method: updateexperiment - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: required - new_name: - type: string - example: - experiment_id: string - new_name: string - description: Updates experiment metadata. - /api/2.0/mlflow/metrics/get-history: - get: - operationId: experiments-history-gethistory - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/gethistory - x-stackQL-resource: experiments_history - x-stackQL-method: gethistory - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: run_id - in: query - - name: run_uuid - in: query - - name: metric_key - required: true - description: string - in: query - - name: page_token - in: query - - name: max_results - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - metrics: - type: array - items: - type: object - properties: - key: - type: string - value: - type: number - timestamp: - type: integer - step: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - x-stackQL-objectKey: $.metrics - description: Gets a list of all values for the specified metric for a given - run. - /api/2.0/mlflow/artifacts/list: - get: - operationId: experiment-artifacts-listartifacts - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/listartifacts - x-stackQL-resource: experiment_artifacts - x-stackQL-method: listartifacts - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: run_id - in: query - - name: run_uuid - in: query - - name: path - in: query - - name: page_token - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - root_uri: - type: string - files: - type: array - items: - type: object - properties: - path: - type: string - is_dir: - type: boolean - file_size: - type: integer - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: List artifacts for a run. Takes an optional - /api/2.0/permissions/experiments/{experiment_id}: - get: - operationId: experiment-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/getpermissions - x-stackQL-resource: experiment_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: experiment_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of an experiment. Experiments can inherit - permissions from their root object. - put: - operationId: experiment-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/setpermissions - x-stackQL-resource: experiment_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: experiment_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: experiment-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/updatepermissions - x-stackQL-resource: experiment_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: experiment_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on an experiment. Experiments can inherit - permissions from their root object. - /api/2.0/permissions/experiments/{experiment_id}/permissionLevels: - get: - operationId: experiment-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/getpermissionlevels - x-stackQL-resource: experiment_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: experiment_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/mlflow/runs/create: - post: - operationId: experiment-runs-createrun - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/createrun - x-stackQL-resource: experiment_runs - x-stackQL-method: createrun - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - run: - type: object - properties: - info: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - experiment_id: - type: string - user_id: - type: string - status: - type: string - start_time: - type: integer - end_time: - type: integer - artifact_uri: - type: string - lifecycle_stage: - type: string - data: - type: object - properties: - metrics: - type: array - items: - type: object - properties: - key: - type: string - value: - type: number - timestamp: - type: integer - step: - type: string - params: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - inputs: - type: object - properties: - dataset_inputs: - type: array - items: - type: object - properties: - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - dataset: - type: object - properties: - name: - type: string - digest: - type: string - source_type: - type: string - source: - type: string - schema: - type: string - profile: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: string - user_id: - type: string - start_time: - type: integer - format: int64 - tags: - type: Array of object - example: - experiment_id: string - user_id: string - start_time: 0 - tags: - - key: string - value: string - description: Creates a new run within an experiment. A run is usually a single - execution of a machine learning or data ETL pipeline. MLflow uses runs to - track the - /api/2.0/mlflow/runs/delete: - post: - operationId: experiment-runs-deleterun - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/deleterun - x-stackQL-resource: experiment_runs - x-stackQL-method: deleterun - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: required - example: - run_id: string - description: Marks a run for deletion. - /api/2.0/mlflow/databricks/runs/delete-runs: - post: - operationId: experiment-runs-deleteruns - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/deleteruns - x-stackQL-resource: experiment_runs - x-stackQL-method: deleteruns - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - runs_deleted: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: required - max_timestamp_millis: - type: string - max_runs: - type: required - example: - experiment_id: string - max_timestamp_millis: 0 - max_runs: '10000' - description: Bulk delete runs in an experiment that were created prior to or - at the specified timestamp. Deletes at most max_runs per request. To call - this API from a Databricks Notebook in Python, you can use the client code - snippet on - /api/2.0/mlflow/runs/get: - get: - operationId: experiment-runs-getrun - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/getrun - x-stackQL-resource: experiment_runs - x-stackQL-method: getrun - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: run_id - required: true - description: string - in: query - - name: run_uuid - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - run: - type: object - properties: - info: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - experiment_id: - type: string - user_id: - type: string - status: - type: string - start_time: - type: integer - end_time: - type: integer - artifact_uri: - type: string - lifecycle_stage: - type: string - data: - type: object - properties: - metrics: - type: array - items: - type: object - properties: - key: - type: string - value: - type: number - timestamp: - type: integer - step: - type: string - params: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - inputs: - type: object - properties: - dataset_inputs: - type: array - items: - type: object - properties: - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - dataset: - type: object - properties: - name: - type: string - digest: - type: string - source_type: - type: string - source: - type: string - schema: - type: string - profile: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.run - description: Gets the metadata, metrics, params, and tags for a run. In the - case where multiple metrics with the same key are logged for a run, return - only the value with the latest timestamp. - /api/2.0/mlflow/runs/log-batch: - post: - operationId: experiment-runs-logbatch - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/logbatch - x-stackQL-resource: experiment_runs - x-stackQL-method: logbatch - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: string - metrics: - type: Array of object - params: - type: Array of object - tags: - type: Array of object - example: - run_id: 2a14ed5c6a87499199e0106c3501eab8 - metrics: - - key: mae - value: 2.5 - timestamp: 1552550804 - - key: rmse - value: 2.7 - timestamp: 1552550804 - params: - - key: model_class - value: LogisticRegression - description: Logs a batch of metrics, params, and tags for a run. If any data - failed to be persisted, the server will respond with an error (non-200 status - code). - /api/2.0/mlflow/runs/log-inputs: - post: - operationId: experiment-runs-loginputs - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/loginputs - x-stackQL-resource: experiment_runs - x-stackQL-method: loginputs - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: string - datasets: - type: Array of object - example: - run_id: string - datasets: - - tags: - - key: string - value: string - dataset: - name: string - digest: string - source_type: string - source: string - schema: string - profile: string - description: 'Experimental: This API may change or be removed in a future release - without warning.' - /api/2.0/mlflow/runs/log-metric: - post: - operationId: experiment-runs-logmetric - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/logmetric - x-stackQL-resource: experiment_runs - x-stackQL-method: logmetric - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - key: - type: required - value: - type: string - timestamp: - type: required - step: - type: double - example: - run_id: string - run_uuid: string - key: string - value: 0.1 - timestamp: 0 - step: '0' - description: Logs a metric for a run. A metric is a key-value pair (string key, - float value) with an associated timestamp. Examples include the various metrics - that represent ML model accuracy. A metric can be logged multiple times. - /api/2.0/mlflow/runs/log-model: - post: - operationId: experiment-runs-logmodel - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/logmodel - x-stackQL-resource: experiment_runs - x-stackQL-method: logmodel - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: string - model_json: - type: string - example: - run_id: string - model_json: string - description: 'Experimental: This API may change or be removed in a future release - without warning.' - /api/2.0/mlflow/runs/log-parameter: - post: - operationId: experiment-runs-logparam - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/logparam - x-stackQL-resource: experiment_runs - x-stackQL-method: logparam - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - key: - type: required - value: - type: string - example: - run_id: string - run_uuid: string - key: string - value: string - description: Logs a param used for a run. A param is a key-value pair (string - key, string value). Examples include hyperparameters used for ML model training - and constant dates and values used in an ETL pipeline. A param can be logged - only once for a run. - /api/2.0/mlflow/runs/restore: - post: - operationId: experiment-runs-restorerun - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/restorerun - x-stackQL-resource: experiment_runs - x-stackQL-method: restorerun - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: required - example: - run_id: string - description: Restores a deleted run. - /api/2.0/mlflow/databricks/runs/restore-runs: - post: - operationId: experiment-runs-restoreruns - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/restoreruns - x-stackQL-resource: experiment_runs - x-stackQL-method: restoreruns - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - runs_restored: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: required - min_timestamp_millis: - type: string - max_runs: - type: required - example: - experiment_id: string - min_timestamp_millis: 0 - max_runs: '10000' - description: Bulk restore runs in an experiment that were deleted no earlier - than the specified timestamp. Restores at most max_runs per request. To call - this API from a Databricks Notebook in Python, you can use the client code - snippet on - /api/2.0/mlflow/runs/search: - post: - operationId: experiment-runs-searchruns - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/searchruns - x-stackQL-resource: experiment_runs - x-stackQL-method: searchruns - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - runs: - type: array - items: - type: object - properties: - info: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - experiment_id: - type: string - user_id: - type: string - status: - type: string - start_time: - type: integer - end_time: - type: integer - artifact_uri: - type: string - lifecycle_stage: - type: string - data: - type: object - properties: - metrics: - type: array - items: - type: object - properties: - key: - type: string - value: - type: number - timestamp: - type: integer - step: - type: string - params: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - inputs: - type: object - properties: - dataset_inputs: - type: array - items: - type: object - properties: - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - dataset: - type: object - properties: - name: - type: string - digest: - type: string - source_type: - type: string - source: - type: string - schema: - type: string - profile: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.runs - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_ids: - type: Array of string - filter: - type: string - run_view_type: - type: string - max_results: - type: int32 - order_by: - type: Array of string - page_token: - type: string - example: - experiment_ids: - - string - filter: string - run_view_type: ACTIVE_ONLY - max_results: '1000' - order_by: - - string - page_token: string - description: Searches for runs that satisfy expressions. - /api/2.0/mlflow/runs/update: - post: - operationId: experiment-runs-updaterun - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/updaterun - x-stackQL-resource: experiment_runs - x-stackQL-method: updaterun - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - run_info: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - experiment_id: - type: string - user_id: - type: string - status: - type: string - start_time: - type: integer - end_time: - type: integer - artifact_uri: - type: string - lifecycle_stage: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - status: - type: string - end_time: - type: integer - format: int64 - example: - run_id: string - run_uuid: string - status: RUNNING - end_time: 0 - description: Updates run metadata. - /api/2.0/mlflow/runs/set-tag: - post: - operationId: experiment-run-tags-settag - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/settag - x-stackQL-resource: experiment_run_tags - x-stackQL-method: settag - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: string - run_uuid: - type: string - key: - type: required - value: - type: string - example: - run_id: string - run_uuid: string - key: string - value: string - description: Sets a tag on a run. Tags are run metadata that can be updated - during a run and after a run completes. - /api/2.0/mlflow/runs/delete-tag: - post: - operationId: experiment-tags-deletetag - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/deletetag - x-stackQL-resource: experiment_tags - x-stackQL-method: deletetag - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: required - key: - type: string - example: - run_id: string - key: string - description: Deletes a tag on a run. Tags are run metadata that can be updated - during a run and after a run completes. - /api/2.0/mlflow/experiments/set-experiment-tag: - post: - operationId: experiment-tags-setexperimenttag - externalDocs: - url: https://docs.databricks.com/api/workspace/experiments/setexperimenttag - x-stackQL-resource: experiment_tags - x-stackQL-method: setexperimenttag - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - experiment_id: - type: required - key: - type: string - value: - type: required - example: - experiment_id: string - key: string - value: string - description: Sets a tag on an experiment. Experiment tags are metadata that - can be updated. - /api/2.0/mlflow/registered-models/create: - post: - operationId: models-createmodel - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/createmodel - x-stackQL-resource: models - x-stackQL-method: createmodel - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - registered_model: - type: object - properties: - name: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - description: - type: string - latest_versions: - type: array - items: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - description: - type: string - tags: - type: Array of object - example: - name: string - tags: - - key: string - value: string - description: string - description: Creates a new registered model with the name specified in the request - body. - /api/2.0/mlflow/registered-models/delete: - delete: - operationId: models-deletemodel - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/deletemodel - x-stackQL-resource: models - x-stackQL-method: deletemodel - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a registered model. - /api/2.0/mlflow/databricks/registered-models/get: - get: - operationId: models-getmodel - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/getmodel - x-stackQL-resource: models - x-stackQL-method: getmodel - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: query - responses: - '200': - description: Model details were returned successfully. - content: - application/json: - schema: - type: object - properties: - registered_model_databricks: - type: object - properties: - name: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - description: - type: string - latest_versions: - type: array - items: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - id: - type: string - permission_level: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get the details of a model. This is a Databricks workspace version - of the - /api/2.0/mlflow/registered-models/list: - get: - operationId: models-listmodels - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/listmodels - x-stackQL-resource: models - x-stackQL-method: listmodels - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - registered_models: - type: array - items: - type: object - properties: - name: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - description: - type: string - latest_versions: - type: array - items: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.registered_models - description: Lists all available registered models, up to the limit specified - in - /api/2.0/mlflow/registered-models/rename: - post: - operationId: models-renamemodel - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/renamemodel - x-stackQL-resource: models - x-stackQL-method: renamemodel - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - registered_model: - type: object - properties: - name: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - description: - type: string - latest_versions: - type: array - items: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - new_name: - type: string - example: - name: string - new_name: string - description: Renames a registered model. - /api/2.0/mlflow/registered-models/search: - get: - operationId: models-searchmodels - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/searchmodels - x-stackQL-resource: models - x-stackQL-method: searchmodels - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: filter - in: query - - name: max_results - in: query - - name: order_by - in: query - - name: page_token - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - registered_models: - type: array - items: - type: object - properties: - name: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - description: - type: string - latest_versions: - type: array - items: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.registered_models - description: Search for registered models based on the specified - /api/2.0/mlflow/registered-models/update: - patch: - operationId: models-updatemodel - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/updatemodel - x-stackQL-resource: models - x-stackQL-method: updatemodel - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - description: - type: string - example: - name: string - description: string - description: Updates a registered model. - /api/2.0/mlflow/registered-models/get-latest-versions: - post: - operationId: model-latest-versions-getlatestversions - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/getlatestversions - x-stackQL-resource: model_latest_versions - x-stackQL-method: getlatestversions - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - model_versions: - type: array - items: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - x-stackQL-objectKey: $.model_versions - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - stages: - type: string - example: - name: string - stages: - - string - description: Gets the latest version of a registered model. - /api/2.0/permissions/registered-models/{registered_model_id}: - get: - operationId: model-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/getpermissions - x-stackQL-resource: model_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: registered_model_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a registered model. Registered models can - inherit permissions from their root object. - put: - operationId: model-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/setpermissions - x-stackQL-resource: model_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: registered_model_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: model-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/updatepermissions - x-stackQL-resource: model_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: registered_model_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a registered model. Registered models - can inherit permissions from their root object. - /api/2.0/permissions/registered-models/{registered_model_id}/permissionLevels: - get: - operationId: model-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/getpermissionlevels - x-stackQL-resource: model_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: registered_model_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/mlflow/registered-models/delete-tag: - delete: - operationId: model-tags-deletemodeltag - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/deletemodeltag - x-stackQL-resource: model_tags - x-stackQL-method: deletemodeltag - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: query - - name: key - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - description: Deletes the tag for a registered model. - /api/2.0/mlflow/registered-models/set-tag: - post: - operationId: model-tags-setmodeltag - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/setmodeltag - x-stackQL-resource: model_tags - x-stackQL-method: setmodeltag - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - key: - type: string - value: - type: required - example: - name: string - key: string - value: string - description: Sets a tag on a registered model. - /api/2.0/mlflow/transition-requests/approve: - post: - operationId: model-transition-requests-approvetransitionrequest - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/approvetransitionrequest - x-stackQL-resource: model_transition_requests - x-stackQL-method: approvetransitionrequest - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Model version's stage was updated successfully. - content: - application/json: - schema: - type: object - properties: - activity: - type: object - properties: - id: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - activity_type: - type: string - from_stage: - type: string - to_stage: - type: string - comment: - type: string - system_comment: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - version: - type: string - stage: - type: required - archive_existing_versions: - type: string - comment: - type: required - example: - name: search_ads_model - version: '1' - stage: Staging - archive_existing_versions: true - comment: This version is great! - description: Details required to identify and approve a model version stage - transition request. - description: Approves a model version stage transition request. - /api/2.0/mlflow/transition-requests/create: - post: - operationId: model-transition-requests-createtransitionrequest - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/createtransitionrequest - x-stackQL-resource: model_transition_requests - x-stackQL-method: createtransitionrequest - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Transition request was made successfully. - content: - application/json: - schema: - type: object - properties: - request: - type: object - properties: - creation_timestamp: - type: integer - to_stage: - type: string - user_id: - type: string - comment: - type: string - available_actions: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - version: - type: string - stage: - type: required - comment: - type: string - example: - name: search_ads_model - version: '1' - stage: Staging - comment: This version is great! - description: Details required to create a model version stage transition request. - description: Creates a model version stage transition request. - /api/2.0/mlflow/transition-requests/delete: - delete: - operationId: model-transition-requests-deletetransitionrequest - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/deletetransitionrequest - x-stackQL-resource: model_transition_requests - x-stackQL-method: deletetransitionrequest - x-stackQL-verb: delete - x-numReqParams: 4 - parameters: - - name: name - required: true - description: string - in: query - - name: version - required: true - description: string - in: query - - name: stage - required: true - description: string - in: query - - name: creator - required: true - description: email - in: query - - name: comment - in: query - responses: - '200': - description: Transition request was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Cancels a model version stage transition request. - /api/2.0/mlflow/transition-requests/list: - get: - operationId: model-transition-requests-listtransitionrequests - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/listtransitionrequests - x-stackQL-resource: model_transition_requests - x-stackQL-method: listtransitionrequests - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: query - - name: version - required: true - description: string - in: query - responses: - '200': - description: Fetched all open requests successfully. - content: - application/json: - schema: - type: object - properties: - requests: - type: array - items: - type: object - properties: - id: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - activity_type: - type: string - from_stage: - type: string - to_stage: - type: string - comment: - type: string - system_comment: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.requests - description: Gets a list of all open stage transition requests for the model - version. - /api/2.0/mlflow/transition-requests/reject: - post: - operationId: model-transition-requests-rejecttransitionrequest - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/rejecttransitionrequest - x-stackQL-resource: model_transition_requests - x-stackQL-method: rejecttransitionrequest - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Transition request was rejected successfully. - content: - application/json: - schema: - type: object - properties: - activity: - type: object - properties: - id: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - activity_type: - type: string - from_stage: - type: string - to_stage: - type: string - comment: - type: string - system_comment: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - version: - type: string - stage: - type: required - comment: - type: string - example: - name: search_ads_model - version: '1' - stage: Staging - comment: This version is great! - description: Details required to identify and reject a model version stage - transition request. - description: Rejects a model version stage transition request. - /api/2.0/mlflow/databricks/model-versions/transition-stage: - post: - operationId: model-transition-requests-transitionstage - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/transitionstage - x-stackQL-resource: model_transition_requests - x-stackQL-method: transitionstage - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Model version's stage was updated successfully. - content: - application/json: - schema: - type: object - properties: - model_version: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - permission_level: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - version: - type: string - stage: - type: required - archive_existing_versions: - type: string - comment: - type: required - example: - name: search_ads_model - version: '1' - stage: Staging - archive_existing_versions: true - comment: This version is great! - description: Details required to transition a model version's stage. - description: Transition a model version's stage. This is a Databricks workspace - version of the - /api/2.0/mlflow/model-versions/create: - post: - operationId: model-versions-createmodelversion - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/createmodelversion - x-stackQL-resource: model_versions - x-stackQL-method: createmodelversion - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - model_version: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - source: - type: string - run_id: - type: required - run_link: - type: string - description: - type: string - tags: - type: Array of object - example: - name: string - source: string - run_id: string - tags: - - key: string - value: string - run_link: string - description: string - description: Creates a model version. - /api/2.0/mlflow/model-versions/delete: - delete: - operationId: model-versions-deletemodelversion - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/deletemodelversion - x-stackQL-resource: model_versions - x-stackQL-method: deletemodelversion - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: query - - name: version - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a model version. - /api/2.0/mlflow/model-versions/get: - get: - operationId: model-versions-getmodelversion - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/getmodelversion - x-stackQL-resource: model_versions - x-stackQL-method: getmodelversion - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: query - - name: version - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - model_version: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - x-stackQL-objectKey: $.model_version - description: Get a model version. - /api/2.0/mlflow/model-versions/search: - get: - operationId: model-versions-searchmodelversions - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/searchmodelversions - x-stackQL-resource: model_versions - x-stackQL-method: searchmodelversions - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: filter - in: query - - name: max_results - in: query - - name: order_by - in: query - - name: page_token - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - model_versions: - type: array - items: - type: object - properties: - name: - type: string - version: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - current_stage: - type: string - description: - type: string - source: - type: string - run_id: - type: string - status: - type: string - status_message: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - run_link: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.model_versions - description: Searches for specific model versions based on the supplied - /api/2.0/mlflow/model-versions/update: - patch: - operationId: model-versions-updatemodelversion - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/updatemodelversion - x-stackQL-resource: model_versions - x-stackQL-method: updatemodelversion - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - version: - type: string - description: - type: required - example: - name: string - version: string - description: string - description: Updates the model version. - /api/2.0/mlflow/comments/create: - post: - operationId: model-version-comments-createcomment - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/createcomment - x-stackQL-resource: model_version_comments - x-stackQL-method: createcomment - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Comment was made successfully. - content: - application/json: - schema: - type: object - properties: - comment: - type: object - properties: - id: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - comment: - type: string - available_actions: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - version: - type: string - comment: - type: required - example: - name: search_ads_model - version: '1' - comment: This version is great! - description: Details required to create a comment on a model version. - description: Posts a comment on a model version. A comment can be submitted - either by a user or programmatically to display relevant information about - the model. For example, test results or deployment errors. - /api/2.0/mlflow/comments/delete: - delete: - operationId: model-version-comments-deletecomment - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/deletecomment - x-stackQL-resource: model_version_comments - x-stackQL-method: deletecomment - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: uuid - in: query - responses: - '200': - description: Comment was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a comment on a model version. - /api/2.0/mlflow/comments/update: - patch: - operationId: model-version-comments-updatecomment - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/updatecomment - x-stackQL-resource: model_version_comments - x-stackQL-method: updatecomment - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Comment was updated successfully. - content: - application/json: - schema: - type: object - properties: - comment: - type: object - properties: - id: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - user_id: - type: string - comment: - type: string - available_actions: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: required - comment: - type: uuid - example: - id: 6fc74c92704341aaa49e74dcc6031057 - comment: This version is great! - description: Details required to edit a comment on a model version. - description: Post an edit to a comment on a model version. - /api/2.0/mlflow/model-versions/get-download-uri: - get: - operationId: model-version-download-urls-getmodelversiondownloaduri - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/getmodelversiondownloaduri - x-stackQL-resource: model_version_download_urls - x-stackQL-method: getmodelversiondownloaduri - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: query - - name: version - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - artifact_uri: - type: string - description: Gets a URI to download the model version. - /api/2.0/mlflow/model-versions/delete-tag: - delete: - operationId: model-version-tags-deletemodelversiontag - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/deletemodelversiontag - x-stackQL-resource: model_version_tags - x-stackQL-method: deletemodelversiontag - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: name - required: true - description: string - in: query - - name: version - required: true - description: string - in: query - - name: key - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a model version tag. - /api/2.0/mlflow/model-versions/set-tag: - post: - operationId: model-version-tags-setmodelversiontag - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/setmodelversiontag - x-stackQL-resource: model_version_tags - x-stackQL-method: setmodelversiontag - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - version: - type: string - key: - type: required - value: - type: string - example: - name: string - version: string - key: string - value: string - description: Sets a model version tag. - /api/2.0/mlflow/registry-webhooks/create: - post: - operationId: model-web-hooks-createwebhook - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/createwebhook - x-stackQL-resource: model_web_hooks - x-stackQL-method: createwebhook - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Registry webhook was created successfully. - content: - application/json: - schema: - type: object - properties: - webhook: - type: object - properties: - id: - type: string - events: - type: array - items: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - description: - type: string - status: - type: string - http_url_spec: - type: object - properties: - url: - type: string - enable_ssl_verification: - type: boolean - model_name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - events: - type: required - model_name: - type: Array of string - description: - type: string - status: - type: string - job_spec: - type: object - http_url_spec: - type: object - example: - job_spec: - job_id: '1' - access_token: dapi12345678935845824 - workspace_url: string - http_url_spec: - url: https://hooks.slack.com/services/... - secret: anyRandomString - enable_ssl_verification: true - authorization: Bearer - events: - - MODEL_VERSION_CREATED - - MODEL_VERSION_TRANSITIONED_TO_STAGING - - COMMENT_CREATED - model_name: registered-model-1 - description: Webhook for comment creation - status: ACTIVE - description: Details required to create a registry webhook. - description: ': This endpoint is in Public Preview.' - /api/2.0/mlflow/registry-webhooks/delete: - delete: - operationId: model-web-hooks-deletewebhook - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/deletewebhook - x-stackQL-resource: model_web_hooks - x-stackQL-method: deletewebhook - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: - - name: id - in: query - responses: - '200': - description: Registry webhook was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: This endpoint is in Public Preview. - /api/2.0/mlflow/registry-webhooks/list: - get: - operationId: model-web-hooks-listwebhooks - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/listwebhooks - x-stackQL-resource: model_web_hooks - x-stackQL-method: listwebhooks - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: model_name - in: query - - name: events - in: query - - name: page_token - in: query - responses: - '200': - description: Registry webhooks listed successfully. - content: - application/json: - schema: - type: object - properties: - webhooks: - type: array - items: - type: object - properties: - id: - type: string - events: - type: array - items: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - description: - type: string - status: - type: string - http_url_spec: - type: object - properties: - url: - type: string - enable_ssl_verification: - type: boolean - model_name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.webhooks - description: This endpoint is in Public Preview. - /api/2.0/mlflow/registry-webhooks/test: - post: - operationId: model-web-hooks-testregistrywebhook - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/testregistrywebhook - x-stackQL-resource: model_web_hooks - x-stackQL-method: testregistrywebhook - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Registry webhook was tested successfully. - content: - application/json: - schema: - type: object - properties: - webhook: - type: object - properties: - status_code: - type: integer - body: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: required - event: - type: string - example: - id: '124323' - event: MODEL_VERSION_CREATED - description: Details required to test a registry webhook. - description: This endpoint is in Public Preview. - /api/2.0/mlflow/registry-webhooks/update: - patch: - operationId: model-web-hooks-updatewebhook - externalDocs: - url: https://docs.databricks.com/api/workspace/modelregistry/updatewebhook - x-stackQL-resource: model_web_hooks - x-stackQL-method: updatewebhook - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Registry webhook was updated successfully. - content: - application/json: - schema: - type: object - properties: - webhook: - type: object - properties: - id: - type: string - events: - type: array - items: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - description: - type: string - status: - type: string - http_url_spec: - type: object - properties: - url: - type: string - enable_ssl_verification: - type: boolean - model_name: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - id: - type: required - events: - type: string - description: - type: Array of string - status: - type: string - http_url_spec: - type: object - job_spec: - type: object - example: - id: '124323' - status: DISABLED - description: Details required to update a registry webhook. Only the fields - that need to be updated should be specified, and both - description: This endpoint is in Public Preview. -components: - x-stackQL-resources: - experiments: - id: databricks_workspace.machinelearning.experiments - name: experiments - title: Experiments - methods: - createexperiment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deleteexperiment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getbyname: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1get-by-name/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.experiment - getexperiment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.experiment - listexperiments: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.experiments - restoreexperiment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1restore/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - searchexperiments: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1search/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.experiments - updateexperiment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1update/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/experiments/methods/searchexperiments' - - $ref: '#/components/x-stackQL-resources/experiments/methods/listexperiments' - - $ref: '#/components/x-stackQL-resources/experiments/methods/getexperiment' - - $ref: '#/components/x-stackQL-resources/experiments/methods/getbyname' - insert: - - $ref: '#/components/x-stackQL-resources/experiments/methods/createexperiment' - update: - - $ref: '#/components/x-stackQL-resources/experiments/methods/updateexperiment' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/experiments/methods/deleteexperiment' - exec: - - $ref: '#/components/x-stackQL-resources/experiments/methods/restoreexperiment' - experiments_history: - id: databricks_workspace.machinelearning.experiments_history - name: experiments_history - title: Experiments_history - methods: - gethistory: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1metrics~1get-history/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.metrics - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/experiments_history/methods/gethistory' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - experiment_artifacts: - id: databricks_workspace.machinelearning.experiment_artifacts - name: experiment_artifacts - title: Experiment_artifacts - methods: - listartifacts: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1artifacts~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/experiment_artifacts/methods/listartifacts' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - experiment_permissions: - id: databricks_workspace.machinelearning.experiment_permissions - name: experiment_permissions - title: Experiment_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/experiment_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/experiment_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/experiment_permissions/methods/setpermissions' - delete: [] - exec: [] - experiment_permission_levels: - id: databricks_workspace.machinelearning.experiment_permission_levels - name: experiment_permission_levels - title: Experiment_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/experiment_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - experiment_runs: - id: databricks_workspace.machinelearning.experiment_runs - name: experiment_runs - title: Experiment_runs - methods: - createrun: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deleterun: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deleteruns: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1databricks~1runs~1delete-runs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getrun: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.run - logbatch: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-batch/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - loginputs: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-inputs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - logmetric: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-metric/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - logmodel: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-model/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - logparam: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-parameter/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - restorerun: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1restore/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - restoreruns: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1databricks~1runs~1restore-runs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - searchruns: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1search/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.runs - updaterun: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1update/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/searchruns' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/getrun' - insert: - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/createrun' - update: - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/updaterun' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/deleteruns' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/deleterun' - exec: - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/restoreruns' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/restorerun' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/logparam' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/logmodel' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/logmetric' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/loginputs' - - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/logbatch' - experiment_run_tags: - id: databricks_workspace.machinelearning.experiment_run_tags - name: experiment_run_tags - title: Experiment_run_tags - methods: - settag: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1set-tag/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/experiment_run_tags/methods/settag' - delete: [] - exec: [] - experiment_tags: - id: databricks_workspace.machinelearning.experiment_tags - name: experiment_tags - title: Experiment_tags - methods: - deletetag: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1runs~1delete-tag/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setexperimenttag: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1set-experiment-tag/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/experiment_tags/methods/setexperimenttag' - delete: - - $ref: '#/components/x-stackQL-resources/experiment_tags/methods/deletetag' - exec: [] - models: - id: databricks_workspace.machinelearning.models - name: models - title: Models - methods: - createmodel: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletemodel: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1delete/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getmodel: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1databricks~1registered-models~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listmodels: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.registered_models - renamemodel: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1rename/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - searchmodels: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1search/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.registered_models - updatemodel: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1update/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/models/methods/searchmodels' - - $ref: '#/components/x-stackQL-resources/models/methods/listmodels' - - $ref: '#/components/x-stackQL-resources/models/methods/getmodel' - insert: - - $ref: '#/components/x-stackQL-resources/models/methods/createmodel' - update: - - $ref: '#/components/x-stackQL-resources/models/methods/updatemodel' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/models/methods/deletemodel' - exec: - - $ref: '#/components/x-stackQL-resources/models/methods/renamemodel' - model_latest_versions: - id: databricks_workspace.machinelearning.model_latest_versions - name: model_latest_versions - title: Model_latest_versions - methods: - getlatestversions: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1get-latest-versions/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.model_versions - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_latest_versions/methods/getlatestversions' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - model_permissions: - id: databricks_workspace.machinelearning.model_permissions - name: model_permissions - title: Model_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/model_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/model_permissions/methods/setpermissions' - delete: [] - exec: [] - model_permission_levels: - id: databricks_workspace.machinelearning.model_permission_levels - name: model_permission_levels - title: Model_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - model_tags: - id: databricks_workspace.machinelearning.model_tags - name: model_tags - title: Model_tags - methods: - deletemodeltag: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1delete-tag/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setmodeltag: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1set-tag/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/model_tags/methods/setmodeltag' - delete: - - $ref: '#/components/x-stackQL-resources/model_tags/methods/deletemodeltag' - exec: [] - model_transition_requests: - id: databricks_workspace.machinelearning.model_transition_requests - name: model_transition_requests - title: Model_transition_requests - methods: - approvetransitionrequest: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1approve/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - createtransitionrequest: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletetransitionrequest: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1delete/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listtransitionrequests: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.requests - rejecttransitionrequest: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1reject/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - transitionstage: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1databricks~1model-versions~1transition-stage/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_transition_requests/methods/listtransitionrequests' - insert: - - $ref: '#/components/x-stackQL-resources/model_transition_requests/methods/createtransitionrequest' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/model_transition_requests/methods/deletetransitionrequest' - exec: - - $ref: '#/components/x-stackQL-resources/model_transition_requests/methods/transitionstage' - - $ref: '#/components/x-stackQL-resources/model_transition_requests/methods/rejecttransitionrequest' - - $ref: '#/components/x-stackQL-resources/model_transition_requests/methods/approvetransitionrequest' - model_versions: - id: databricks_workspace.machinelearning.model_versions - name: model_versions - title: Model_versions - methods: - createmodelversion: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletemodelversion: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1delete/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getmodelversion: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.model_version - searchmodelversions: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1search/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.model_versions - updatemodelversion: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1update/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_versions/methods/searchmodelversions' - - $ref: '#/components/x-stackQL-resources/model_versions/methods/getmodelversion' - insert: - - $ref: '#/components/x-stackQL-resources/model_versions/methods/createmodelversion' - update: - - $ref: '#/components/x-stackQL-resources/model_versions/methods/updatemodelversion' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/model_versions/methods/deletemodelversion' - exec: [] - model_version_comments: - id: databricks_workspace.machinelearning.model_version_comments - name: model_version_comments - title: Model_version_comments - methods: - createcomment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1comments~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletecomment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1comments~1delete/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatecomment: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1comments~1update/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: - - $ref: '#/components/x-stackQL-resources/model_version_comments/methods/createcomment' - update: - - $ref: '#/components/x-stackQL-resources/model_version_comments/methods/updatecomment' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/model_version_comments/methods/deletecomment' - exec: [] - model_version_download_urls: - id: databricks_workspace.machinelearning.model_version_download_urls - name: model_version_download_urls - title: Model_version_download_urls - methods: - getmodelversiondownloaduri: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1get-download-uri/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_version_download_urls/methods/getmodelversiondownloaduri' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - model_version_tags: - id: databricks_workspace.machinelearning.model_version_tags - name: model_version_tags - title: Model_version_tags - methods: - deletemodelversiontag: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1delete-tag/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setmodelversiontag: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1set-tag/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/model_version_tags/methods/setmodelversiontag' - delete: - - $ref: '#/components/x-stackQL-resources/model_version_tags/methods/deletemodelversiontag' - exec: [] - model_web_hooks: - id: databricks_workspace.machinelearning.model_web_hooks - name: model_web_hooks - title: Model_web_hooks - methods: - createwebhook: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletewebhook: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1delete/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listwebhooks: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.webhooks - testregistrywebhook: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1test/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatewebhook: - operation: - $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1update/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_web_hooks/methods/listwebhooks' - insert: - - $ref: '#/components/x-stackQL-resources/model_web_hooks/methods/createwebhook' - update: - - $ref: '#/components/x-stackQL-resources/model_web_hooks/methods/updatewebhook' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/model_web_hooks/methods/deletewebhook' - exec: - - $ref: '#/components/x-stackQL-resources/model_web_hooks/methods/testregistrywebhook' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/marketplace.yaml b/providers/src/databricks_workspace/v00.00.00000/services/marketplace.yaml index 9ddf0000..ab661ef5 100644 --- a/providers/src/databricks_workspace/v00.00.00000/services/marketplace.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/services/marketplace.yaml @@ -1,4531 +1,4550 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Marketplace API - description: Marketplace -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.1/marketplace-consumer/listings/{listing_id}/fulfillments: - get: - operationId: consumer-fulfillments-list - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerfulfillments/list - x-stackQL-resource: consumer_fulfillments - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: listing_id - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - fulfillments: - type: array - items: - type: object - properties: - listing_id: - type: string - fulfillment_type: - type: string - share_info: - type: object - properties: - name: - type: string - type: - type: string - repo_info: - type: object - properties: - git_repo_url: - type: string - recipient_type: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.fulfillments - description: Get all listings fulfillments associated with a listing. A - /api/2.1/marketplace-consumer/listings/{listing_id}/installations: - post: - operationId: consumer-installations-create - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerinstallations/create - x-stackQL-resource: consumer_installations - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: listing_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - installation: - type: object - properties: - id: - type: string - listing_id: - type: string - share_name: - type: string - catalog_name: - type: string - installed_on: - type: integer - status: - type: string - error_message: - type: string - listing_name: - type: string - repo_name: - type: string - repo_path: - type: string - recipient_type: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - token_detail: - type: object - properties: - shareCredentialsVersion: - type: integer - bearerToken: - type: string - endpoint: - type: string - expirationTime: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - share_name: - type: string - catalog_name: - type: string - recipient_type: - type: string - repo_detail: - type: object - accepted_consumer_terms: - type: object - example: - share_name: string - catalog_name: string - repo_detail: - repo_name: string - repo_path: string - recipient_type: DELTA_SHARING_RECIPIENT_TYPE_DATABRICKS - accepted_consumer_terms: - version: string - description: Install payload associated with a Databricks Marketplace listing. - get: - operationId: listing-installations-listlistinginstallations - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerinstallations/listlistinginstallations - x-stackQL-resource: listing_installations - x-stackQL-method: listlistinginstallations - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: listing_id - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - installations: - type: array - items: - type: object - properties: - id: - type: string - listing_id: - type: string - share_name: - type: string - catalog_name: - type: string - installed_on: - type: integer - status: - type: string - error_message: - type: string - listing_name: - type: string - repo_name: - type: string - repo_path: - type: string - recipient_type: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - token_detail: - type: object - properties: - shareCredentialsVersion: - type: integer - bearerToken: - type: string - endpoint: - type: string - expirationTime: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.installations - description: List all installations for a particular listing. - /api/2.1/marketplace-consumer/listings/{listing_id}/installations/{installation_id}: - delete: - operationId: consumer-installations-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerinstallations/delete - x-stackQL-resource: consumer_installations - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: listing_id - required: true - description: string - in: path - - name: installation_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Uninstall an installation associated with a Databricks Marketplace - listing. - put: - operationId: consumer-installations-update - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerinstallations/update - x-stackQL-resource: consumer_installations - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: listing_id - required: true - description: string - in: path - - name: installation_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - installation: - type: object - properties: - id: - type: string - listing_id: - type: string - share_name: - type: string - catalog_name: - type: string - installed_on: - type: integer - status: - type: string - error_message: - type: string - listing_name: - type: string - repo_name: - type: string - repo_path: - type: string - recipient_type: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - token_detail: - type: object - properties: - shareCredentialsVersion: - type: integer - bearerToken: - type: string - endpoint: - type: string - expirationTime: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - rotate_token: - type: boolean - installation: - type: required - example: - installation: - id: string - listing_id: string - share_name: string - catalog_name: string - installed_on: 0 - status: INSTALLED - error_message: string - listing_name: string - repo_name: string - repo_path: string - recipient_type: DELTA_SHARING_RECIPIENT_TYPE_DATABRICKS - tokens: - - id: string - created_at: 0 - created_by: string - activation_url: string - expiration_time: 0 - updated_at: 0 - updated_by: string - token_detail: - shareCredentialsVersion: 0 - bearerToken: string - endpoint: string - expirationTime: string - rotate_token: true - description: This is a update API that will update the part of the fields defined - in the installation table as well as interact with external services according - to the fields not included in the installation table - /api/2.1/marketplace-consumer/installations: - get: - operationId: consumer-installations-list - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerinstallations/list - x-stackQL-resource: consumer_installations - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - installations: - type: array - items: - type: object - properties: - id: - type: string - listing_id: - type: string - share_name: - type: string - catalog_name: - type: string - installed_on: - type: integer - status: - type: string - error_message: - type: string - listing_name: - type: string - repo_name: - type: string - repo_path: - type: string - recipient_type: - type: string - tokens: - type: array - items: - type: object - properties: - id: - type: string - created_at: - type: integer - created_by: - type: string - activation_url: - type: string - expiration_time: - type: integer - updated_at: - type: integer - updated_by: - type: string - token_detail: - type: object - properties: - shareCredentialsVersion: - type: integer - bearerToken: - type: string - endpoint: - type: string - expirationTime: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.installations - description: List all installations across all listings. - /api/2.1/marketplace-consumer/listings:batchGet: - get: - operationId: consumer-listings-batchget - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerlistings/batchget - x-stackQL-resource: consumer_listings - x-stackQL-method: batchget - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: - - name: ids - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listings: - type: array - items: - type: object - properties: - id: - type: string - summary: - type: object - properties: - name: - type: string - subtitle: - type: string - status: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - provider_region: - type: object - properties: - cloud: - type: string - region: - type: string - setting: - type: object - properties: - visibility: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - published_at: - type: integer - published_by: - type: string - categories: - type: array - items: - type: string - listingType: - type: string - created_by_id: - type: integer - updated_by_id: - type: integer - provider_id: - type: string - exchange_ids: - type: array - items: - type: string - git_repo: - type: object - properties: - git_repo_url: - type: string - detail: - type: object - properties: - description: - type: string - terms_of_service: - type: string - documentation_link: - type: string - support_link: - type: string - file_ids: - type: array - items: - type: string - privacy_policy_link: - type: string - embedded_notebook_file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - geographical_coverage: - type: string - cost: - type: string - pricing_model: - type: string - update_frequency: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_granularity: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_date_start: - type: integer - collection_date_end: - type: integer - data_source: - type: string - size: - type: number - assets: - type: array - items: - type: string - license: - type: string - tags: - type: array - items: - type: object - properties: - tag_name: - type: string - tag_values: - type: array - items: - type: string - description: Batch get a published listing in the Databricks Marketplace that - the consumer has access to. - /api/2.1/marketplace-consumer/listings/{id}: - get: - operationId: consumer-listings-get - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerlistings/get - x-stackQL-resource: consumer_listings - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listing: - type: object - properties: - id: - type: string - summary: - type: object - properties: - name: - type: string - subtitle: - type: string - status: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - provider_region: - type: object - properties: - cloud: - type: string - region: - type: string - setting: - type: object - properties: - visibility: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - published_at: - type: integer - published_by: - type: string - categories: - type: array - items: - type: string - listingType: - type: string - created_by_id: - type: integer - updated_by_id: - type: integer - provider_id: - type: string - exchange_ids: - type: array - items: - type: string - git_repo: - type: object - properties: - git_repo_url: - type: string - detail: - type: object - properties: - description: - type: string - terms_of_service: - type: string - documentation_link: - type: string - support_link: - type: string - file_ids: - type: array - items: - type: string - privacy_policy_link: - type: string - embedded_notebook_file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - geographical_coverage: - type: string - cost: - type: string - pricing_model: - type: string - update_frequency: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_granularity: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_date_start: - type: integer - collection_date_end: - type: integer - data_source: - type: string - size: - type: number - assets: - type: array - items: - type: string - license: - type: string - tags: - type: array - items: - type: object - properties: - tag_name: - type: string - tag_values: - type: array - items: - type: string - x-stackQL-objectKey: $.listing - description: Get a published listing in the Databricks Marketplace that the - consumer has access to. - /api/2.1/marketplace-consumer/listings: - get: - operationId: consumer-listings-list - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerlistings/list - x-stackQL-resource: consumer_listings - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - - name: assets - in: query - - name: categories - in: query - - name: null - in: query - - name: is_free - in: query - - name: is_private_exchange - in: query - - name: is_staff_pick - in: query - - name: provider_ids - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listings: - type: array - items: - type: object - properties: - id: - type: string - summary: - type: object - properties: - name: - type: string - subtitle: - type: string - status: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - provider_region: - type: object - properties: - cloud: - type: string - region: - type: string - setting: - type: object - properties: - visibility: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - published_at: - type: integer - published_by: - type: string - categories: - type: array - items: - type: string - listingType: - type: string - created_by_id: - type: integer - updated_by_id: - type: integer - provider_id: - type: string - exchange_ids: - type: array - items: - type: string - git_repo: - type: object - properties: - git_repo_url: - type: string - detail: - type: object - properties: - description: - type: string - terms_of_service: - type: string - documentation_link: - type: string - support_link: - type: string - file_ids: - type: array - items: - type: string - privacy_policy_link: - type: string - embedded_notebook_file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - geographical_coverage: - type: string - cost: - type: string - pricing_model: - type: string - update_frequency: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_granularity: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_date_start: - type: integer - collection_date_end: - type: integer - data_source: - type: string - size: - type: number - assets: - type: array - items: - type: string - license: - type: string - tags: - type: array - items: - type: object - properties: - tag_name: - type: string - tag_values: - type: array - items: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.listings - description: List all published listings in the Databricks Marketplace that - the consumer has access to. - /api/2.1/marketplace-consumer/search-listings: - get: - operationId: consumer-listings-search - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerlistings/search - x-stackQL-resource: consumer_listings - x-stackQL-method: search - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: query - required: true - description: string - in: query - - name: is_free - in: query - - name: is_private_exchange - in: query - - name: provider_ids - in: query - - name: categories - in: query - - name: assets - in: query - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listings: - type: array - items: - type: object - properties: - id: - type: string - summary: - type: object - properties: - name: - type: string - subtitle: - type: string - status: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - provider_region: - type: object - properties: - cloud: - type: string - region: - type: string - setting: - type: object - properties: - visibility: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - published_at: - type: integer - published_by: - type: string - categories: - type: array - items: - type: string - listingType: - type: string - created_by_id: - type: integer - updated_by_id: - type: integer - provider_id: - type: string - exchange_ids: - type: array - items: - type: string - git_repo: - type: object - properties: - git_repo_url: - type: string - detail: - type: object - properties: - description: - type: string - terms_of_service: - type: string - documentation_link: - type: string - support_link: - type: string - file_ids: - type: array - items: - type: string - privacy_policy_link: - type: string - embedded_notebook_file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - geographical_coverage: - type: string - cost: - type: string - pricing_model: - type: string - update_frequency: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_granularity: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_date_start: - type: integer - collection_date_end: - type: integer - data_source: - type: string - size: - type: number - assets: - type: array - items: - type: string - license: - type: string - tags: - type: array - items: - type: object - properties: - tag_name: - type: string - tag_values: - type: array - items: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.listings - description: Search published listings in the Databricks Marketplace that the - consumer has access to. This query supports a variety of different search - parameters and performs fuzzy matching. - /api/2.1/marketplace-consumer/listings/{listing_id}/content: - get: - operationId: consumer-listing-content-get - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerfulfillments/get - x-stackQL-resource: consumer_listing_content - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: listing_id - required: true - description: string - in: path - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - shared_data_objects: - type: array - items: - type: object - properties: - name: - type: string - data_object_type: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.shared_data_objects - description: Get a high level preview of the metadata of listing installable - content. - /api/2.1/marketplace-consumer/listings/{listing_id}/personalization-requests: - post: - operationId: consumer-personalization-requests-create - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerpersonalizationrequests/create - x-stackQL-resource: consumer_personalization_requests - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: listing_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - comment: - type: string - intended_use: - type: required - first_name: - type: string - last_name: - type: string - company: - type: string - is_from_lighthouse: - type: string - recipient_type: - type: boolean - accepted_consumer_terms: - type: object - example: - comment: string - intended_use: string - first_name: string - last_name: string - company: string - is_from_lighthouse: true - recipient_type: DELTA_SHARING_RECIPIENT_TYPE_DATABRICKS - accepted_consumer_terms: - version: string - description: Data request messages also creates a lead (maybe) - description: Create a personalization request for a listing. - get: - operationId: consumer-personalization-requests-get - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerpersonalizationrequests/get - x-stackQL-resource: consumer_personalization_requests - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: listing_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - personalization_requests: - type: array - items: - type: object - properties: - id: - type: string - consumer_region: - type: object - properties: - cloud: - type: string - region: - type: string - contact_info: - type: object - properties: - first_name: - type: string - last_name: - type: string - email: - type: string - company: - type: string - comment: - type: string - intended_use: - type: string - status: - type: string - status_message: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - created_at: - type: integer - listing_id: - type: string - updated_at: - type: integer - metastore_id: - type: string - listing_name: - type: string - is_from_lighthouse: - type: boolean - provider_id: - type: string - recipient_type: - type: string - x-stackQL-objectKey: $.personalization_requests - description: Get the personalization request for a listing. Each consumer can - make at - /api/2.1/marketplace-consumer/personalization-requests: - get: - operationId: consumer-personalization-requests-list - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerpersonalizationrequests/list - x-stackQL-resource: consumer_personalization_requests - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - personalization_requests: - type: array - items: - type: object - properties: - id: - type: string - consumer_region: - type: object - properties: - cloud: - type: string - region: - type: string - contact_info: - type: object - properties: - first_name: - type: string - last_name: - type: string - email: - type: string - company: - type: string - comment: - type: string - intended_use: - type: string - status: - type: string - status_message: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - created_at: - type: integer - listing_id: - type: string - updated_at: - type: integer - metastore_id: - type: string - listing_name: - type: string - is_from_lighthouse: - type: boolean - provider_id: - type: string - recipient_type: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.personalization_requests - description: List personalization requests for a consumer across all listings. - /api/2.1/marketplace-consumer/providers:batchGet: - get: - operationId: consumer-providers-batchget - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerproviders/batchget - x-stackQL-resource: consumer_providers - x-stackQL-method: batchget - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: - - name: ids - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - providers: - type: array - items: - type: object - properties: - id: - type: string - name: - type: string - description: - type: string - icon_file_path: - type: string - business_contact_email: - type: string - support_contact_email: - type: string - is_featured: - type: boolean - published_by: - type: string - company_website_link: - type: string - icon_file_id: - type: string - term_of_service_link: - type: string - privacy_policy_link: - type: string - dark_mode_icon_file_id: - type: string - dark_mode_icon_file_path: - type: string - description: Batch get a provider in the Databricks Marketplace with at least - one visible listing. - /api/2.1/marketplace-consumer/providers/{id}: - get: - operationId: consumer-providers-get - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerproviders/get - x-stackQL-resource: consumer_providers - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - provider: - type: object - properties: - id: - type: string - name: - type: string - description: - type: string - icon_file_path: - type: string - business_contact_email: - type: string - support_contact_email: - type: string - is_featured: - type: boolean - published_by: - type: string - company_website_link: - type: string - icon_file_id: - type: string - term_of_service_link: - type: string - privacy_policy_link: - type: string - dark_mode_icon_file_id: - type: string - dark_mode_icon_file_path: - type: string - x-stackQL-objectKey: $.provider - description: Get a provider in the Databricks Marketplace with at least one - visible listing. - /api/2.1/marketplace-consumer/providers: - get: - operationId: consumer-providers-list - externalDocs: - url: https://docs.databricks.com/api/workspace/consumerproviders/list - x-stackQL-resource: consumer_providers - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - - name: is_featured - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - providers: - type: array - items: - type: object - properties: - id: - type: string - name: - type: string - description: - type: string - icon_file_path: - type: string - business_contact_email: - type: string - support_contact_email: - type: string - is_featured: - type: boolean - published_by: - type: string - company_website_link: - type: string - icon_file_id: - type: string - term_of_service_link: - type: string - privacy_policy_link: - type: string - dark_mode_icon_file_id: - type: string - dark_mode_icon_file_path: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.providers - description: List all providers in the Databricks Marketplace with at least - one visible listing. - /api/2.0/marketplace-provider/provider: - post: - operationId: providers-create - externalDocs: - url: https://docs.databricks.com/api/workspace/providerproviders/create - x-stackQL-resource: providers - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - provider: - type: object - example: - provider: - id: string - name: string - description: string - icon_file_path: string - business_contact_email: string - support_contact_email: string - is_featured: true - published_by: string - company_website_link: string - icon_file_id: string - term_of_service_link: string - privacy_policy_link: string - dark_mode_icon_file_id: string - dark_mode_icon_file_path: string - description: Create a provider - /api/2.0/marketplace-provider/providers/{id}: - delete: - operationId: providers-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/providerproviders/delete - x-stackQL-resource: providers - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Delete provider - get: - operationId: providers-get - externalDocs: - url: https://docs.databricks.com/api/workspace/providerproviders/get - x-stackQL-resource: providers - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - provider: - type: object - properties: - id: - type: string - name: - type: string - description: - type: string - icon_file_path: - type: string - business_contact_email: - type: string - support_contact_email: - type: string - is_featured: - type: boolean - published_by: - type: string - company_website_link: - type: string - icon_file_id: - type: string - term_of_service_link: - type: string - privacy_policy_link: - type: string - dark_mode_icon_file_id: - type: string - dark_mode_icon_file_path: - type: string - x-stackQL-objectKey: $.provider - description: Get provider profile - put: - operationId: providers-update - externalDocs: - url: https://docs.databricks.com/api/workspace/providerproviders/update - x-stackQL-resource: providers - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - provider: - type: object - properties: - id: - type: string - name: - type: string - description: - type: string - icon_file_path: - type: string - business_contact_email: - type: string - support_contact_email: - type: string - is_featured: - type: boolean - published_by: - type: string - company_website_link: - type: string - icon_file_id: - type: string - term_of_service_link: - type: string - privacy_policy_link: - type: string - dark_mode_icon_file_id: - type: string - dark_mode_icon_file_path: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - provider: - type: object - example: - provider: - id: string - name: string - description: string - icon_file_path: string - business_contact_email: string - support_contact_email: string - is_featured: true - published_by: string - company_website_link: string - icon_file_id: string - term_of_service_link: string - privacy_policy_link: string - dark_mode_icon_file_id: string - dark_mode_icon_file_path: string - description: Update provider profile - /api/2.0/marketplace-provider/providers: - get: - operationId: providers-list - externalDocs: - url: https://docs.databricks.com/api/workspace/providerproviders/list - x-stackQL-resource: providers - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - providers: - type: array - items: - type: object - properties: - id: - type: string - name: - type: string - description: - type: string - icon_file_path: - type: string - business_contact_email: - type: string - support_contact_email: - type: string - is_featured: - type: boolean - published_by: - type: string - company_website_link: - type: string - icon_file_id: - type: string - term_of_service_link: - type: string - privacy_policy_link: - type: string - dark_mode_icon_file_id: - type: string - dark_mode_icon_file_path: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.providers - description: List provider profiles for account. - /api/2.0/marketplace-provider/analytics_dashboard: - post: - operationId: provider-analytics-dashboards-create - externalDocs: - url: https://docs.databricks.com/api/workspace/providerprovideranalyticsdashboards/create - x-stackQL-resource: provider_analytics_dashboards - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - description: Create provider analytics dashboard. Returns Marketplace specific - get: - operationId: provider-analytics-dashboards-get - externalDocs: - url: https://docs.databricks.com/api/workspace/providerprovideranalyticsdashboards/get - x-stackQL-resource: provider_analytics_dashboards - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - version: - type: integer - dashboard_id: - type: string - description: Get provider analytics dashboard. - /api/2.0/marketplace-provider/analytics_dashboard/{id}: - put: - operationId: provider-analytics-dashboards-update - externalDocs: - url: https://docs.databricks.com/api/workspace/providerprovideranalyticsdashboards/update - x-stackQL-resource: provider_analytics_dashboards - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - id: - type: string - version: - type: integer - dashboard_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - version: - type: integer - format: int64 - example: - version: 0 - description: Update provider analytics dashboard. - /api/2.0/marketplace-provider/analytics_dashboard/latest: - get: - operationId: provider-analytics-dashboards-latest-getlatestversion - externalDocs: - url: https://docs.databricks.com/api/workspace/providerprovideranalyticsdashboards/getlatestversion - x-stackQL-resource: provider_analytics_dashboards_latest - x-stackQL-method: getlatestversion - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - version: - type: integer - description: Get latest version of provider analytics dashboard. - /api/2.0/marketplace-exchange/exchanges: - post: - operationId: provider-exchanges-create - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/create - x-stackQL-resource: provider_exchanges - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - exchange_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - exchange: - type: object - example: - exchange: - id: string - name: string - comment: string - filters: - - id: string - exchange_id: string - filter_value: string - name: string - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - filter_type: GLOBAL_METASTORE_ID - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - linked_listings: - - id: string - exchange_id: string - exchange_name: string - listing_id: string - listing_name: string - created_at: 0 - created_by: string - description: Create an exchange - get: - operationId: provider-exchanges-list - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/list - x-stackQL-resource: provider_exchanges - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - exchanges: - type: array - items: - type: object - properties: - id: - type: string - name: - type: string - comment: - type: string - filters: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - filter_value: - type: string - name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - filter_type: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - linked_listings: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - exchange_name: - type: string - listing_id: - type: string - listing_name: - type: string - created_at: - type: integer - created_by: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.exchanges - description: List exchanges visible to provider - /api/2.0/marketplace-exchange/exchanges/{id}: - delete: - operationId: provider-exchanges-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/delete - x-stackQL-resource: provider_exchanges - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: This removes a listing from marketplace. - get: - operationId: provider-exchanges-get - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/get - x-stackQL-resource: provider_exchanges - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - exchange: - type: object - properties: - id: - type: string - name: - type: string - comment: - type: string - filters: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - filter_value: - type: string - name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - filter_type: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - linked_listings: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - exchange_name: - type: string - listing_id: - type: string - listing_name: - type: string - created_at: - type: integer - created_by: - type: string - x-stackQL-objectKey: $.exchange - description: Get an exchange. - put: - operationId: provider-exchanges-update - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/update - x-stackQL-resource: provider_exchanges - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - exchange: - type: object - properties: - id: - type: string - name: - type: string - comment: - type: string - filters: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - filter_value: - type: string - name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - filter_type: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - linked_listings: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - exchange_name: - type: string - listing_id: - type: string - listing_name: - type: string - created_at: - type: integer - created_by: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - exchange: - type: object - example: - exchange: - id: string - name: string - comment: string - filters: - - id: string - exchange_id: string - filter_value: string - name: string - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - filter_type: GLOBAL_METASTORE_ID - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - linked_listings: - - id: string - exchange_id: string - exchange_name: string - listing_id: string - listing_name: string - created_at: 0 - created_by: string - description: Update an exchange - /api/2.0/marketplace-exchange/filters: - post: - operationId: provider-exchange-filters-create - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchangefilters/create - x-stackQL-resource: provider_exchange_filters - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - filter_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - filter: - type: object - example: - filter: - id: string - exchange_id: string - filter_value: string - name: string - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - filter_type: GLOBAL_METASTORE_ID - description: Add an exchange filter. - get: - operationId: provider-exchange-filters-list - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchangefilters/list - x-stackQL-resource: provider_exchange_filters - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: exchange_id - required: true - description: string - in: query - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - filters: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - filter_value: - type: string - name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - filter_type: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.filters - description: List exchange filter - /api/2.0/marketplace-exchange/filters/{id}: - delete: - operationId: provider-exchange-filters-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchangefilters/delete - x-stackQL-resource: provider_exchange_filters - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Delete an exchange filter - put: - operationId: provider-exchange-filters-update - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchangefilters/update - x-stackQL-resource: provider_exchange_filters - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - filter: - type: object - properties: - id: - type: string - exchange_id: - type: string - filter_value: - type: string - name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - filter_type: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - filter: - type: object - example: - filter: - id: string - exchange_id: string - filter_value: string - name: string - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - filter_type: GLOBAL_METASTORE_ID - description: Update an exchange filter. - /api/2.0/marketplace-exchange/exchanges-for-listing: - post: - operationId: provider-exchange-listings-addlistingtoexchange - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/addlistingtoexchange - x-stackQL-resource: provider_exchange_listings - x-stackQL-method: addlistingtoexchange - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - exchange_for_listing: - type: object - properties: - id: - type: string - exchange_id: - type: string - exchange_name: - type: string - listing_id: - type: string - listing_name: - type: string - created_at: - type: integer - created_by: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - listing_id: - type: required - exchange_id: - type: string - example: - listing_id: string - exchange_id: string - description: Associate an exchange with a listing - get: - operationId: provider-exchange-listings-listexchangesforlisting - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/listexchangesforlisting - x-stackQL-resource: provider_exchange_listings - x-stackQL-method: listexchangesforlisting - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: listing_id - required: true - description: string - in: query - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - exchange_listing: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - exchange_name: - type: string - listing_id: - type: string - listing_name: - type: string - created_at: - type: integer - created_by: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.exchange_listing - description: List exchanges associated with a listing - /api/2.0/marketplace-exchange/exchanges-for-listing/{id}: - delete: - operationId: provider-exchange-listings-deletelistingfromexchange - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/deletelistingfromexchange - x-stackQL-resource: provider_exchange_listings - x-stackQL-method: deletelistingfromexchange - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Disassociate an exchange with a listing - /api/2.0/marketplace-exchange/listings-for-exchange: - get: - operationId: provider-exchange-listings-listlistingsforexchange - externalDocs: - url: https://docs.databricks.com/api/workspace/providerexchanges/listlistingsforexchange - x-stackQL-resource: provider_exchange_listings - x-stackQL-method: listlistingsforexchange - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: exchange_id - required: true - description: string - in: query - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - exchange_listings: - type: array - items: - type: object - properties: - id: - type: string - exchange_id: - type: string - exchange_name: - type: string - listing_id: - type: string - listing_name: - type: string - created_at: - type: integer - created_by: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.exchange_listings - description: List listings associated with an exchange - /api/2.0/marketplace-provider/files: - post: - operationId: provider-files-create - externalDocs: - url: https://docs.databricks.com/api/workspace/providerfiles/create - x-stackQL-resource: provider_files - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - upload_url: - type: string - file_info: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - marketplace_file_type: - type: required - mime_type: - type: string - display_name: - type: required - file_parent: - type: required - example: - file_parent: - parent_id: string - file_parent_type: PROVIDER - marketplace_file_type: PROVIDER_ICON - mime_type: string - display_name: string - description: Create a file. Currently, only provider icons and attached notebooks - are supported. - get: - operationId: provider-files-list - externalDocs: - url: https://docs.databricks.com/api/workspace/providerfiles/list - x-stackQL-resource: provider_files - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: null - in: query - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.file_infos - description: List files attached to a parent entity. - /api/2.0/marketplace-provider/files/{file_id}: - delete: - operationId: provider-files-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/providerfiles/delete - x-stackQL-resource: provider_files - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: file_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Delete a file - get: - operationId: provider-files-get - externalDocs: - url: https://docs.databricks.com/api/workspace/providerfiles/get - x-stackQL-resource: provider_files - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: file_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - file_info: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - x-stackQL-objectKey: $.file_info - description: Get a file - /api/2.0/marketplace-provider/listing: - post: - operationId: provider-listings-create - externalDocs: - url: https://docs.databricks.com/api/workspace/providerlistings/create - x-stackQL-resource: provider_listings - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listing_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - listing: - type: object - example: - listing: - id: string - summary: - name: string - subtitle: string - status: DRAFT - share: - name: string - type: SAMPLE - provider_region: - cloud: string - region: string - setting: - visibility: PUBLIC - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - published_at: 0 - published_by: string - categories: - - ADVERTISING_AND_MARKETING - listingType: STANDARD - created_by_id: 0 - updated_by_id: 0 - provider_id: string - exchange_ids: - - string - git_repo: - git_repo_url: string - detail: - description: string - terms_of_service: string - documentation_link: string - support_link: string - file_ids: - - string - privacy_policy_link: string - embedded_notebook_file_infos: - - id: string - marketplace_file_type: PROVIDER_ICON - file_parent: - parent_id: string - file_parent_type: PROVIDER - mime_type: string - download_link: string - created_at: 0 - updated_at: 0 - display_name: string - status: FILE_STATUS_PUBLISHED - status_message: string - geographical_coverage: string - cost: FREE - pricing_model: string - update_frequency: - interval: 0 - unit: NONE - collection_granularity: - interval: 0 - unit: NONE - collection_date_start: 0 - collection_date_end: 0 - data_source: string - size: 0.1 - assets: - - ASSET_TYPE_GIT_REPO - license: string - tags: - - tag_name: LISTING_TAG_TYPE_LANGUAGE - tag_values: - - string - description: Create a new listing - /api/2.0/marketplace-provider/listings/{id}: - delete: - operationId: provider-listings-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/providerlistings/delete - x-stackQL-resource: provider_listings - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Delete a listing - get: - operationId: provider-listings-get - externalDocs: - url: https://docs.databricks.com/api/workspace/providerlistings/get - x-stackQL-resource: provider_listings - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listing: - type: object - properties: - id: - type: string - summary: - type: object - properties: - name: - type: string - subtitle: - type: string - status: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - provider_region: - type: object - properties: - cloud: - type: string - region: - type: string - setting: - type: object - properties: - visibility: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - published_at: - type: integer - published_by: - type: string - categories: - type: array - items: - type: string - listingType: - type: string - created_by_id: - type: integer - updated_by_id: - type: integer - provider_id: - type: string - exchange_ids: - type: array - items: - type: string - git_repo: - type: object - properties: - git_repo_url: - type: string - detail: - type: object - properties: - description: - type: string - terms_of_service: - type: string - documentation_link: - type: string - support_link: - type: string - file_ids: - type: array - items: - type: string - privacy_policy_link: - type: string - embedded_notebook_file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - geographical_coverage: - type: string - cost: - type: string - pricing_model: - type: string - update_frequency: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_granularity: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_date_start: - type: integer - collection_date_end: - type: integer - data_source: - type: string - size: - type: number - assets: - type: array - items: - type: string - license: - type: string - tags: - type: array - items: - type: object - properties: - tag_name: - type: string - tag_values: - type: array - items: - type: string - x-stackQL-objectKey: $.listing - description: Get a listing - put: - operationId: provider-listings-update - externalDocs: - url: https://docs.databricks.com/api/workspace/providerlistings/update - x-stackQL-resource: provider_listings - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listing: - type: object - properties: - id: - type: string - summary: - type: object - properties: - name: - type: string - subtitle: - type: string - status: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - provider_region: - type: object - properties: - cloud: - type: string - region: - type: string - setting: - type: object - properties: - visibility: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - published_at: - type: integer - published_by: - type: string - categories: - type: array - items: - type: string - listingType: - type: string - created_by_id: - type: integer - updated_by_id: - type: integer - provider_id: - type: string - exchange_ids: - type: array - items: - type: string - git_repo: - type: object - properties: - git_repo_url: - type: string - detail: - type: object - properties: - description: - type: string - terms_of_service: - type: string - documentation_link: - type: string - support_link: - type: string - file_ids: - type: array - items: - type: string - privacy_policy_link: - type: string - embedded_notebook_file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - geographical_coverage: - type: string - cost: - type: string - pricing_model: - type: string - update_frequency: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_granularity: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_date_start: - type: integer - collection_date_end: - type: integer - data_source: - type: string - size: - type: number - assets: - type: array - items: - type: string - license: - type: string - tags: - type: array - items: - type: object - properties: - tag_name: - type: string - tag_values: - type: array - items: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - listing: - type: object - example: - listing: - id: string - summary: - name: string - subtitle: string - status: DRAFT - share: - name: string - type: SAMPLE - provider_region: - cloud: string - region: string - setting: - visibility: PUBLIC - created_at: 0 - created_by: string - updated_at: 0 - updated_by: string - published_at: 0 - published_by: string - categories: - - ADVERTISING_AND_MARKETING - listingType: STANDARD - created_by_id: 0 - updated_by_id: 0 - provider_id: string - exchange_ids: - - string - git_repo: - git_repo_url: string - detail: - description: string - terms_of_service: string - documentation_link: string - support_link: string - file_ids: - - string - privacy_policy_link: string - embedded_notebook_file_infos: - - id: string - marketplace_file_type: PROVIDER_ICON - file_parent: - parent_id: string - file_parent_type: PROVIDER - mime_type: string - download_link: string - created_at: 0 - updated_at: 0 - display_name: string - status: FILE_STATUS_PUBLISHED - status_message: string - geographical_coverage: string - cost: FREE - pricing_model: string - update_frequency: - interval: 0 - unit: NONE - collection_granularity: - interval: 0 - unit: NONE - collection_date_start: 0 - collection_date_end: 0 - data_source: string - size: 0.1 - assets: - - ASSET_TYPE_GIT_REPO - license: string - tags: - - tag_name: LISTING_TAG_TYPE_LANGUAGE - tag_values: - - string - description: Update a listing - /api/2.0/marketplace-provider/listings: - get: - operationId: provider-listings-list - externalDocs: - url: https://docs.databricks.com/api/workspace/providerlistings/list - x-stackQL-resource: provider_listings - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - listings: - type: array - items: - type: object - properties: - id: - type: string - summary: - type: object - properties: - name: - type: string - subtitle: - type: string - status: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - provider_region: - type: object - properties: - cloud: - type: string - region: - type: string - setting: - type: object - properties: - visibility: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - published_at: - type: integer - published_by: - type: string - categories: - type: array - items: - type: string - listingType: - type: string - created_by_id: - type: integer - updated_by_id: - type: integer - provider_id: - type: string - exchange_ids: - type: array - items: - type: string - git_repo: - type: object - properties: - git_repo_url: - type: string - detail: - type: object - properties: - description: - type: string - terms_of_service: - type: string - documentation_link: - type: string - support_link: - type: string - file_ids: - type: array - items: - type: string - privacy_policy_link: - type: string - embedded_notebook_file_infos: - type: array - items: - type: object - properties: - id: - type: string - marketplace_file_type: - type: string - file_parent: - type: object - properties: - parent_id: - type: string - file_parent_type: - type: string - mime_type: - type: string - download_link: - type: string - created_at: - type: integer - updated_at: - type: integer - display_name: - type: string - status: - type: string - status_message: - type: string - geographical_coverage: - type: string - cost: - type: string - pricing_model: - type: string - update_frequency: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_granularity: - type: object - properties: - interval: - type: integer - unit: - type: string - collection_date_start: - type: integer - collection_date_end: - type: integer - data_source: - type: string - size: - type: number - assets: - type: array - items: - type: string - license: - type: string - tags: - type: array - items: - type: object - properties: - tag_name: - type: string - tag_values: - type: array - items: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.listings - description: List listings owned by this provider - /api/2.0/marketplace-provider/personalization-requests: - get: - operationId: provider-personalization-requests-list - externalDocs: - url: https://docs.databricks.com/api/workspace/providerpersonalizationrequests/list - x-stackQL-resource: provider_personalization_requests - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - personalization_requests: - type: array - items: - type: object - properties: - id: - type: string - consumer_region: - type: object - properties: - cloud: - type: string - region: - type: string - contact_info: - type: object - properties: - first_name: - type: string - last_name: - type: string - email: - type: string - company: - type: string - comment: - type: string - intended_use: - type: string - status: - type: string - status_message: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - created_at: - type: integer - listing_id: - type: string - updated_at: - type: integer - metastore_id: - type: string - listing_name: - type: string - is_from_lighthouse: - type: boolean - provider_id: - type: string - recipient_type: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.personalization_requests - description: List personalization requests to this provider. This will return - all personalization requests, regardless of which listing they are for. - /api/2.0/marketplace-provider/listings/{listing_id}/personalization-requests/{request_id}/request-status: - put: - operationId: provider-personalization-requests-update - externalDocs: - url: https://docs.databricks.com/api/workspace/providerpersonalizationrequests/update - x-stackQL-resource: provider_personalization_requests - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: listing_id - required: true - description: string - in: path - - name: request_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - request: - type: object - properties: - id: - type: string - consumer_region: - type: object - properties: - cloud: - type: string - region: - type: string - contact_info: - type: object - properties: - first_name: - type: string - last_name: - type: string - email: - type: string - company: - type: string - comment: - type: string - intended_use: - type: string - status: - type: string - status_message: - type: string - share: - type: object - properties: - name: - type: string - type: - type: string - created_at: - type: integer - listing_id: - type: string - updated_at: - type: integer - metastore_id: - type: string - listing_name: - type: string - is_from_lighthouse: - type: boolean - provider_id: - type: string - recipient_type: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - status: - type: required - reason: - type: string - share: - type: object - example: - status: NEW - reason: string - share: - name: string - type: SAMPLE - description: Update personalization request. This method only permits updating - the status of the request. -components: - x-stackQL-resources: - consumer_fulfillments: - id: databricks_workspace.marketplace.consumer_fulfillments - name: consumer_fulfillments - title: Consumer_fulfillments - methods: - list: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1fulfillments/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.fulfillments - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/consumer_fulfillments/methods/list' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - consumer_installations: - id: databricks_workspace.marketplace.consumer_installations - name: consumer_installations - title: Consumer_installations - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations~1{installation_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations~1{installation_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1installations/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.installations - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/consumer_installations/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/consumer_installations/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/consumer_installations/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/consumer_installations/methods/delete' - exec: [] - listing_installations: - id: databricks_workspace.marketplace.listing_installations - name: listing_installations - title: Listing_installations - methods: - listlistinginstallations: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.installations - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/listing_installations/methods/listlistinginstallations' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - consumer_listings: - id: databricks_workspace.marketplace.consumer_listings - name: consumer_listings - title: Consumer_listings - methods: - batchget: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings:batchGet/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.listing - list: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.listings - search: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1search-listings/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.listings - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/consumer_listings/methods/search' - - $ref: '#/components/x-stackQL-resources/consumer_listings/methods/list' - - $ref: '#/components/x-stackQL-resources/consumer_listings/methods/get' - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/consumer_listings/methods/batchget' - consumer_listing_content: - id: databricks_workspace.marketplace.consumer_listing_content - name: consumer_listing_content - title: Consumer_listing_content - methods: - get: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1content/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.shared_data_objects - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/consumer_listing_content/methods/get' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - consumer_personalization_requests: - id: databricks_workspace.marketplace.consumer_personalization_requests - name: consumer_personalization_requests - title: Consumer_personalization_requests - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1personalization-requests/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1personalization-requests/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.personalization_requests - list: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1personalization-requests/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.personalization_requests - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/consumer_personalization_requests/methods/list' - - $ref: '#/components/x-stackQL-resources/consumer_personalization_requests/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/consumer_personalization_requests/methods/create' - update: [] - replace: [] - delete: [] - exec: [] - consumer_providers: - id: databricks_workspace.marketplace.consumer_providers - name: consumer_providers - title: Consumer_providers - methods: - batchget: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1providers:batchGet/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1providers~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.provider - list: - operation: - $ref: '#/paths/~1api~12.1~1marketplace-consumer~1providers/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.providers - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/consumer_providers/methods/list' - - $ref: '#/components/x-stackQL-resources/consumer_providers/methods/get' - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/consumer_providers/methods/batchget' - providers: - id: databricks_workspace.marketplace.providers - name: providers - title: Providers - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1provider/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.provider - update: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.providers - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/providers/methods/list' - - $ref: '#/components/x-stackQL-resources/providers/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/providers/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/providers/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/providers/methods/delete' - exec: [] - provider_analytics_dashboards: - id: databricks_workspace.marketplace.provider_analytics_dashboards - name: provider_analytics_dashboards - title: Provider_analytics_dashboards - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_analytics_dashboards/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/provider_analytics_dashboards/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/provider_analytics_dashboards/methods/update' - replace: [] - delete: [] - exec: [] - provider_analytics_dashboards_latest: - id: databricks_workspace.marketplace.provider_analytics_dashboards_latest - name: provider_analytics_dashboards_latest - title: Provider_analytics_dashboards_latest - methods: - getlatestversion: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard~1latest/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_analytics_dashboards_latest/methods/getlatestversion' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - provider_exchanges: - id: databricks_workspace.marketplace.provider_exchanges - name: provider_exchanges - title: Provider_exchanges - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.exchanges - delete: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.exchange - update: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/get' - - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/delete' - exec: [] - provider_exchange_filters: - id: databricks_workspace.marketplace.provider_exchange_filters - name: provider_exchange_filters - title: Provider_exchange_filters - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.filters - delete: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_exchange_filters/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/provider_exchange_filters/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/provider_exchange_filters/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/provider_exchange_filters/methods/delete' - exec: [] - provider_exchange_listings: - id: databricks_workspace.marketplace.provider_exchange_listings - name: provider_exchange_listings - title: Provider_exchange_listings - methods: - addlistingtoexchange: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges-for-listing/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listexchangesforlisting: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges-for-listing/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.exchange_listing - deletelistingfromexchange: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges-for-listing~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listlistingsforexchange: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-exchange~1listings-for-exchange/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.exchange_listings - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_exchange_listings/methods/listlistingsforexchange' - - $ref: '#/components/x-stackQL-resources/provider_exchange_listings/methods/listexchangesforlisting' - insert: - - $ref: '#/components/x-stackQL-resources/provider_exchange_listings/methods/addlistingtoexchange' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/provider_exchange_listings/methods/deletelistingfromexchange' - exec: [] - provider_files: - id: databricks_workspace.marketplace.provider_files - name: provider_files - title: Provider_files - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1files/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1files/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.file_infos - delete: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1files~1{file_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1files~1{file_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.file_info - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_files/methods/get' - - $ref: '#/components/x-stackQL-resources/provider_files/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/provider_files/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/provider_files/methods/delete' - exec: [] - provider_listings: - id: databricks_workspace.marketplace.provider_listings - name: provider_listings - title: Provider_listings - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1listing/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.listing - update: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings~1{id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.listings - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_listings/methods/list' - - $ref: '#/components/x-stackQL-resources/provider_listings/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/provider_listings/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/provider_listings/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/provider_listings/methods/delete' - exec: [] - provider_personalization_requests: - id: databricks_workspace.marketplace.provider_personalization_requests - name: provider_personalization_requests - title: Provider_personalization_requests - methods: - list: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1personalization-requests/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.personalization_requests - update: - operation: - $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings~1{listing_id}~1personalization-requests~1{request_id}~1request-status/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/provider_personalization_requests/methods/list' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/provider_personalization_requests/methods/update' - replace: [] - delete: [] - exec: [] +openapi: 3.0.0 +info: + title: Databricks Marketplace API (workspace) + description: >- + OpenAPI specification for the Databricks marketplace service + (workspace-level APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.marketplace +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.1/marketplace-consumer/listings/{listing_id}/content: + get: + operationId: consumer_fulfillments_get + summary: Get a high level preview of the metadata of listing installable content. + tags: + - marketplace + - consumer_fulfillments + description: |- + Get a high level preview of the metadata of listing installable content. + + :param listing_id: str + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`SharedDataObject` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`SharedDataObject`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetListingContentMetadataResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerFulfillmentsAPI + /api/2.1/marketplace-consumer/listings/{listing_id}/fulfillments: + get: + operationId: consumer_fulfillments_list + summary: >- + Get all listings fulfillments associated with a listing. A _fulfillment_ + is a potential installation. + tags: + - marketplace + - consumer_fulfillments + description: >- + Get all listings fulfillments associated with a listing. A _fulfillment_ + is a potential installation. + + Standard installations contain metadata about the attached share or git + repo. Only one of these fields + + will be present. Personalized installations contain metadata about the + attached share or git repo, as + + well as the Delta Sharing recipient type. + + + :param listing_id: str + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`ListingFulfillment` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`ListingFulfillment`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFulfillmentsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerFulfillmentsAPI + /api/2.1/marketplace-consumer/listings/{listing_id}/installations: + post: + operationId: consumer_installations_create + summary: Install payload associated with a Databricks Marketplace listing. + tags: + - marketplace + - consumer_installations + description: |- + Install payload associated with a Databricks Marketplace listing. + + :param listing_id: str + :param accepted_consumer_terms: :class:`ConsumerTerms` (optional) + :param catalog_name: str (optional) + :param recipient_type: :class:`DeltaSharingRecipientType` (optional) + :param repo_detail: :class:`RepoInstallation` (optional) + for git repo installations + :param share_name: str (optional) + + :returns: :class:`Installation` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param accepted_consumer_terms: :class:`ConsumerTerms` (optional)' + requestBody: + content: + application/json: + schema: + type: object + properties: + accepted_consumer_terms: + type: string + catalog_name: + type: string + description: >- + :param recipient_type: :class:`DeltaSharingRecipientType` + (optional) + recipient_type: + type: string + repo_detail: + type: string + description: for git repo installations + share_name: + type: string + description: ':returns: :class:`Installation`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Installation' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerInstallationsAPI + get: + operationId: consumer_installations_list_listing_installations + summary: List all installations for a particular listing. + tags: + - marketplace + - consumer_installations + description: |- + List all installations for a particular listing. + + :param listing_id: str + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`InstallationDetail` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`InstallationDetail`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAllInstallationsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerInstallationsAPI + /api/2.1/marketplace-consumer/listings/{listing_id}/installations/{installation_id}: + delete: + operationId: consumer_installations_delete + summary: >- + Uninstall an installation associated with a Databricks Marketplace + listing. + tags: + - marketplace + - consumer_installations + description: >- + Uninstall an installation associated with a Databricks Marketplace + listing. + + + :param listing_id: str + + :param installation_id: str + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param installation_id: str' + - name: installation_id + in: path + required: true + schema: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerInstallationsAPI + put: + operationId: consumer_installations_update + summary: >- + This is a update API that will update the part of the fields defined in + the installation table as well + tags: + - marketplace + - consumer_installations + description: >- + This is a update API that will update the part of the fields defined in + the installation table as well + + as interact with external services according to the fields not included + in the installation table 1. + + the token will be rotate if the rotateToken flag is true 2. the token + will be forcibly rotate if the + + rotateToken flag is true and the tokenInfo field is empty + + + :param listing_id: str + + :param installation_id: str + + :param installation: :class:`InstallationDetail` + + :param rotate_token: bool (optional) + + + :returns: :class:`UpdateInstallationResponse` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param installation_id: str' + - name: installation_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + installation: + type: string + description: ':param rotate_token: bool (optional)' + rotate_token: + type: string + required: + - installation + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateInstallationResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerInstallationsAPI + /api/2.1/marketplace-consumer/installations: + get: + operationId: consumer_installations_list + summary: List all installations across all listings. + tags: + - marketplace + - consumer_installations + description: |- + List all installations across all listings. + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`InstallationDetail` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAllInstallationsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerInstallationsAPI + /api/2.1/marketplace-consumer/listings:batchGet: + get: + operationId: consumer_listings_batch_get + summary: >- + Batch get a published listing in the Databricks Marketplace that the + consumer has access to. + tags: + - marketplace + - consumer_listings + description: >- + Batch get a published listing in the Databricks Marketplace that the + consumer has access to. + + + :param ids: List[str] (optional) + + + :returns: :class:`BatchGetListingsResponse` + parameters: + - name: ids + in: query + required: false + schema: + type: string + description: ':returns: :class:`BatchGetListingsResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchGetListingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerListingsAPI + /api/2.1/marketplace-consumer/listings/{id}: + get: + operationId: consumer_listings_get + summary: >- + Get a published listing in the Databricks Marketplace that the consumer + has access to. + tags: + - marketplace + - consumer_listings + description: >- + Get a published listing in the Databricks Marketplace that the consumer + has access to. + + + :param id: str + + + :returns: :class:`GetListingResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetListingResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetListingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerListingsAPI + /api/2.1/marketplace-consumer/listings: + get: + operationId: consumer_listings_list + summary: >- + List all published listings in the Databricks Marketplace that the + consumer has access to. + tags: + - marketplace + - consumer_listings + description: >- + List all published listings in the Databricks Marketplace that the + consumer has access to. + + + :param assets: List[:class:`AssetType`] (optional) + Matches any of the following asset types + :param categories: List[:class:`Category`] (optional) + Matches any of the following categories + :param is_free: bool (optional) + Filters each listing based on if it is free. + :param is_private_exchange: bool (optional) + Filters each listing based on if it is a private exchange. + :param is_staff_pick: bool (optional) + Filters each listing based on whether it is a staff pick. + :param page_size: int (optional) + + :param page_token: str (optional) + + :param provider_ids: List[str] (optional) + Matches any of the following provider ids + :param tags: List[:class:`ListingTag`] (optional) + Matches any of the following tags + + :returns: Iterator over :class:`Listing` + parameters: + - name: assets + in: query + required: false + schema: + type: string + description: Matches any of the following asset types + - name: categories + in: query + required: false + schema: + type: string + description: Matches any of the following categories + - name: is_free + in: query + required: false + schema: + type: string + description: Filters each listing based on if it is free. + - name: is_private_exchange + in: query + required: false + schema: + type: string + description: Filters each listing based on if it is a private exchange. + - name: is_staff_pick + in: query + required: false + schema: + type: string + description: Filters each listing based on whether it is a staff pick. + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + - name: provider_ids + in: query + required: false + schema: + type: string + description: Matches any of the following provider ids + - name: tags + in: query + required: false + schema: + type: string + description: Matches any of the following tags + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchGetListingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerListingsAPI + /api/2.1/marketplace-consumer/search-listings: + get: + operationId: consumer_listings_search + summary: >- + Search published listings in the Databricks Marketplace that the + consumer has access to. This query + tags: + - marketplace + - consumer_listings + description: >- + Search published listings in the Databricks Marketplace that the + consumer has access to. This query + + supports a variety of different search parameters and performs fuzzy + matching. + + + :param query: str + Fuzzy matches query + :param assets: List[:class:`AssetType`] (optional) + Matches any of the following asset types + :param categories: List[:class:`Category`] (optional) + Matches any of the following categories + :param is_free: bool (optional) + + :param is_private_exchange: bool (optional) + + :param page_size: int (optional) + + :param page_token: str (optional) + + :param provider_ids: List[str] (optional) + Matches any of the following provider ids + + :returns: Iterator over :class:`Listing` + parameters: + - name: query + in: query + required: true + schema: + type: string + description: Fuzzy matches query + - name: assets + in: query + required: false + schema: + type: string + description: Matches any of the following asset types + - name: categories + in: query + required: false + schema: + type: string + description: Matches any of the following categories + - name: is_free + in: query + required: false + schema: + type: string + description: ':param is_private_exchange: bool (optional)' + - name: is_private_exchange + in: query + required: false + schema: + type: string + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + - name: provider_ids + in: query + required: false + schema: + type: string + description: Matches any of the following provider ids + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchGetListingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerListingsAPI + /api/2.1/marketplace-consumer/listings/{listing_id}/personalization-requests: + post: + operationId: consumer_personalization_requests_create + summary: Create a personalization request for a listing. + tags: + - marketplace + - consumer_personalization_requests + description: |- + Create a personalization request for a listing. + + :param listing_id: str + :param intended_use: str + :param accepted_consumer_terms: :class:`ConsumerTerms` + :param comment: str (optional) + :param company: str (optional) + :param first_name: str (optional) + :param is_from_lighthouse: bool (optional) + :param last_name: str (optional) + :param recipient_type: :class:`DeltaSharingRecipientType` (optional) + + :returns: :class:`CreatePersonalizationRequestResponse` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param intended_use: str' + requestBody: + content: + application/json: + schema: + type: object + properties: + intended_use: + type: string + accepted_consumer_terms: + type: string + description: ':param comment: str (optional)' + comment: + type: string + company: + type: string + description: ':param first_name: str (optional)' + first_name: + type: string + is_from_lighthouse: + type: string + description: ':param last_name: str (optional)' + last_name: + type: string + recipient_type: + type: string + description: ':returns: :class:`CreatePersonalizationRequestResponse`' + required: + - intended_use + - accepted_consumer_terms + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreatePersonalizationRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerPersonalizationRequestsAPI + get: + operationId: consumer_personalization_requests_get + summary: >- + Get the personalization request for a listing. Each consumer can make at + *most* one personalization + tags: + - marketplace + - consumer_personalization_requests + description: >- + Get the personalization request for a listing. Each consumer can make at + *most* one personalization + + request for a listing. + + + :param listing_id: str + + + :returns: :class:`GetPersonalizationRequestResponse` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetPersonalizationRequestResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPersonalizationRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerPersonalizationRequestsAPI + /api/2.1/marketplace-consumer/personalization-requests: + get: + operationId: consumer_personalization_requests_list + summary: List personalization requests for a consumer across all listings. + tags: + - marketplace + - consumer_personalization_requests + description: |- + List personalization requests for a consumer across all listings. + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`PersonalizationRequest` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPersonalizationRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerPersonalizationRequestsAPI + /api/2.1/marketplace-consumer/providers:batchGet: + get: + operationId: consumer_providers_batch_get + summary: >- + Batch get a provider in the Databricks Marketplace with at least one + visible listing. + tags: + - marketplace + - consumer_providers + description: >- + Batch get a provider in the Databricks Marketplace with at least one + visible listing. + + + :param ids: List[str] (optional) + + + :returns: :class:`BatchGetProvidersResponse` + parameters: + - name: ids + in: query + required: false + schema: + type: string + description: ':returns: :class:`BatchGetProvidersResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchGetProvidersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerProvidersAPI + /api/2.1/marketplace-consumer/providers/{id}: + get: + operationId: consumer_providers_get + summary: >- + Get a provider in the Databricks Marketplace with at least one visible + listing. + tags: + - marketplace + - consumer_providers + description: >- + Get a provider in the Databricks Marketplace with at least one visible + listing. + + + :param id: str + + + :returns: :class:`GetProviderResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetProviderResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetProviderResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerProvidersAPI + /api/2.1/marketplace-consumer/providers: + get: + operationId: consumer_providers_list + summary: >- + List all providers in the Databricks Marketplace with at least one + visible listing. + tags: + - marketplace + - consumer_providers + description: >- + List all providers in the Databricks Marketplace with at least one + visible listing. + + + :param is_featured: bool (optional) + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`ProviderInfo` + parameters: + - name: is_featured + in: query + required: false + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`ProviderInfo`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchGetProvidersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ConsumerProvidersAPI + /api/2.0/marketplace-exchange/filters: + post: + operationId: provider_exchange_filters_create + summary: Add an exchange filter. + tags: + - marketplace + - provider_exchange_filters + description: |- + Add an exchange filter. + + :param filter: :class:`ExchangeFilter` + + :returns: :class:`CreateExchangeFilterResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + filter: + type: string + description: ':returns: :class:`CreateExchangeFilterResponse`' + required: + - filter + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateExchangeFilterResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangeFiltersAPI + get: + operationId: provider_exchange_filters_list + summary: List exchange filter + tags: + - marketplace + - provider_exchange_filters + description: |- + List exchange filter + + :param exchange_id: str + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`ExchangeFilter` + parameters: + - name: exchange_id + in: query + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`ExchangeFilter`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExchangeFiltersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangeFiltersAPI + /api/2.0/marketplace-exchange/filters/{id}: + delete: + operationId: provider_exchange_filters_delete + summary: Delete an exchange filter + tags: + - marketplace + - provider_exchange_filters + description: |- + Delete an exchange filter + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangeFiltersAPI + put: + operationId: provider_exchange_filters_update + summary: Update an exchange filter. + tags: + - marketplace + - provider_exchange_filters + description: |- + Update an exchange filter. + + :param id: str + :param filter: :class:`ExchangeFilter` + + :returns: :class:`UpdateExchangeFilterResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':param filter: :class:`ExchangeFilter`' + requestBody: + content: + application/json: + schema: + type: object + properties: + filter: + type: string + required: + - filter + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateExchangeFilterResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangeFiltersAPI + /api/2.0/marketplace-exchange/exchanges-for-listing: + post: + operationId: provider_exchanges_add_listing_to_exchange + summary: Associate an exchange with a listing + tags: + - marketplace + - provider_exchanges + description: |- + Associate an exchange with a listing + + :param listing_id: str + :param exchange_id: str + + :returns: :class:`AddExchangeForListingResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + listing_id: + type: string + description: ':param exchange_id: str' + exchange_id: + type: string + required: + - listing_id + - exchange_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AddExchangeForListingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + get: + operationId: provider_exchanges_list_exchanges_for_listing + summary: List exchanges associated with a listing + tags: + - marketplace + - provider_exchanges + description: |- + List exchanges associated with a listing + + :param listing_id: str + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`ExchangeListing` + parameters: + - name: listing_id + in: query + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`ExchangeListing`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExchangesForListingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + /api/2.0/marketplace-exchange/exchanges: + post: + operationId: provider_exchanges_create + summary: Create an exchange + tags: + - marketplace + - provider_exchanges + description: |- + Create an exchange + + :param exchange: :class:`Exchange` + + :returns: :class:`CreateExchangeResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + exchange: + type: string + description: ':returns: :class:`CreateExchangeResponse`' + required: + - exchange + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateExchangeResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + get: + operationId: provider_exchanges_list + summary: List exchanges visible to provider + tags: + - marketplace + - provider_exchanges + description: |- + List exchanges visible to provider + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`Exchange` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExchangesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + /api/2.0/marketplace-exchange/exchanges/{id}: + delete: + operationId: provider_exchanges_delete + summary: This removes a listing from marketplace. + tags: + - marketplace + - provider_exchanges + description: |- + This removes a listing from marketplace. + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + get: + operationId: provider_exchanges_get + summary: Get an exchange. + tags: + - marketplace + - provider_exchanges + description: |- + Get an exchange. + + :param id: str + + :returns: :class:`GetExchangeResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetExchangeResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetExchangeResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + put: + operationId: provider_exchanges_update + summary: Update an exchange + tags: + - marketplace + - provider_exchanges + description: |- + Update an exchange + + :param id: str + :param exchange: :class:`Exchange` + + :returns: :class:`UpdateExchangeResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':param exchange: :class:`Exchange`' + requestBody: + content: + application/json: + schema: + type: object + properties: + exchange: + type: string + required: + - exchange + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateExchangeResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + /api/2.0/marketplace-exchange/exchanges-for-listing/{id}: + delete: + operationId: provider_exchanges_delete_listing_from_exchange + summary: Disassociate an exchange with a listing + tags: + - marketplace + - provider_exchanges + description: |- + Disassociate an exchange with a listing + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + /api/2.0/marketplace-exchange/listings-for-exchange: + get: + operationId: provider_exchanges_list_listings_for_exchange + summary: List listings associated with an exchange + tags: + - marketplace + - provider_exchanges + description: |- + List listings associated with an exchange + + :param exchange_id: str + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`ExchangeListing` + parameters: + - name: exchange_id + in: query + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`ExchangeListing`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExchangesForListingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderExchangesAPI + /api/2.0/marketplace-provider/files: + post: + operationId: provider_files_create + summary: >- + Create a file. Currently, only provider icons and attached notebooks are + supported. + tags: + - marketplace + - provider_files + description: >- + Create a file. Currently, only provider icons and attached notebooks are + supported. + + + :param file_parent: :class:`FileParent` + + :param marketplace_file_type: :class:`MarketplaceFileType` + + :param mime_type: str + + :param display_name: str (optional) + + + :returns: :class:`CreateFileResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + file_parent: + type: string + description: ':param marketplace_file_type: :class:`MarketplaceFileType`' + marketplace_file_type: + type: string + mime_type: + type: string + description: ':param display_name: str (optional)' + display_name: + type: string + required: + - file_parent + - marketplace_file_type + - mime_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateFileResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderFilesAPI + get: + operationId: provider_files_list + summary: List files attached to a parent entity. + tags: + - marketplace + - provider_files + description: |- + List files attached to a parent entity. + + :param file_parent: :class:`FileParent` + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`FileInfo` + parameters: + - name: file_parent + in: query + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`FileInfo`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFilesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderFilesAPI + /api/2.0/marketplace-provider/files/{file_id}: + delete: + operationId: provider_files_delete + summary: Delete a file + tags: + - marketplace + - provider_files + description: |- + Delete a file + + :param file_id: str + parameters: + - name: file_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderFilesAPI + get: + operationId: provider_files_get + summary: Get a file + tags: + - marketplace + - provider_files + description: |- + Get a file + + :param file_id: str + + :returns: :class:`GetFileResponse` + parameters: + - name: file_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetFileResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetFileResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderFilesAPI + /api/2.0/marketplace-provider/listing: + post: + operationId: provider_listings_create + summary: Create a new listing + tags: + - marketplace + - provider_listings + description: |- + Create a new listing + + :param listing: :class:`Listing` + + :returns: :class:`CreateListingResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + listing: + type: string + description: ':returns: :class:`CreateListingResponse`' + required: + - listing + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateListingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderListingsAPI + /api/2.0/marketplace-provider/listings/{id}: + delete: + operationId: provider_listings_delete + summary: Delete a listing + tags: + - marketplace + - provider_listings + description: |- + Delete a listing + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderListingsAPI + get: + operationId: provider_listings_get + summary: Get a listing + tags: + - marketplace + - provider_listings + description: |- + Get a listing + + :param id: str + + :returns: :class:`GetListingResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetListingResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetListingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderListingsAPI + put: + operationId: provider_listings_update + summary: Update a listing + tags: + - marketplace + - provider_listings + description: |- + Update a listing + + :param id: str + :param listing: :class:`Listing` + + :returns: :class:`UpdateListingResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':param listing: :class:`Listing`' + requestBody: + content: + application/json: + schema: + type: object + properties: + listing: + type: string + required: + - listing + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateListingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderListingsAPI + /api/2.0/marketplace-provider/listings: + get: + operationId: provider_listings_list + summary: List listings owned by this provider + tags: + - marketplace + - provider_listings + description: |- + List listings owned by this provider + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`Listing` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchGetListingsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderListingsAPI + /api/2.0/marketplace-provider/personalization-requests: + get: + operationId: provider_personalization_requests_list + summary: >- + List personalization requests to this provider. This will return all + personalization requests, + tags: + - marketplace + - provider_personalization_requests + description: >- + List personalization requests to this provider. This will return all + personalization requests, + + regardless of which listing they are for. + + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`PersonalizationRequest` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPersonalizationRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderPersonalizationRequestsAPI + /api/2.0/marketplace-provider/listings/{listing_id}/personalization-requests/{request_id}/request-status: + put: + operationId: provider_personalization_requests_update + summary: >- + Update personalization request. This method only permits updating the + status of the request. + tags: + - marketplace + - provider_personalization_requests + description: >- + Update personalization request. This method only permits updating the + status of the request. + + + :param listing_id: str + + :param request_id: str + + :param status: :class:`PersonalizationRequestStatus` + + :param reason: str (optional) + + :param share: :class:`ShareInfo` (optional) + + + :returns: :class:`UpdatePersonalizationRequestResponse` + parameters: + - name: listing_id + in: path + required: true + schema: + type: string + description: ':param request_id: str' + - name: request_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + status: + type: string + description: ':param reason: str (optional)' + reason: + type: string + share: + type: string + description: ':returns: :class:`UpdatePersonalizationRequestResponse`' + required: + - status + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdatePersonalizationRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderPersonalizationRequestsAPI + /api/2.0/marketplace-provider/analytics_dashboard: + post: + operationId: provider_provider_analytics_dashboards_create + summary: >- + Create provider analytics dashboard. Returns Marketplace specific `id`. + Not to be confused with the + tags: + - marketplace + - provider_provider_analytics_dashboards + description: >- + Create provider analytics dashboard. Returns Marketplace specific `id`. + Not to be confused with the + + Lakeview dashboard id. + + + + :returns: :class:`ProviderAnalyticsDashboard` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ProviderAnalyticsDashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProviderAnalyticsDashboardsAPI + get: + operationId: provider_provider_analytics_dashboards_get + summary: Get provider analytics dashboard. + tags: + - marketplace + - provider_provider_analytics_dashboards + description: |- + Get provider analytics dashboard. + + + :returns: :class:`ListProviderAnalyticsDashboardResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListProviderAnalyticsDashboardResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProviderAnalyticsDashboardsAPI + /api/2.0/marketplace-provider/analytics_dashboard/latest: + get: + operationId: provider_provider_analytics_dashboards_get_latest_version + summary: Get latest version of provider analytics dashboard. + tags: + - marketplace + - provider_provider_analytics_dashboards + description: |- + Get latest version of provider analytics dashboard. + + + :returns: :class:`GetLatestVersionProviderAnalyticsDashboardResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/GetLatestVersionProviderAnalyticsDashboardResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProviderAnalyticsDashboardsAPI + /api/2.0/marketplace-provider/analytics_dashboard/{id}: + put: + operationId: provider_provider_analytics_dashboards_update + summary: Update provider analytics dashboard. + tags: + - marketplace + - provider_provider_analytics_dashboards + description: |- + Update provider analytics dashboard. + + :param id: str + id is immutable property and can't be updated. + :param version: int (optional) + this is the version of the dashboard template we want to update our user to current expectation is + that it should be equal to latest version of the dashboard template + + :returns: :class:`UpdateProviderAnalyticsDashboardResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: id is immutable property and can't be updated. + requestBody: + content: + application/json: + schema: + type: object + properties: + version: + type: string + description: >- + this is the version of the dashboard template we want to + update our user to current expectation is that it should be + equal to latest version of the dashboard template + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateProviderAnalyticsDashboardResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProviderAnalyticsDashboardsAPI + /api/2.0/marketplace-provider/provider: + post: + operationId: provider_providers_create + summary: Create a provider + tags: + - marketplace + - provider_providers + description: |- + Create a provider + + :param provider: :class:`ProviderInfo` + + :returns: :class:`CreateProviderResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + provider: + type: string + description: ':returns: :class:`CreateProviderResponse`' + required: + - provider + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateProviderResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProvidersAPI + /api/2.0/marketplace-provider/providers/{id}: + delete: + operationId: provider_providers_delete + summary: Delete provider + tags: + - marketplace + - provider_providers + description: |- + Delete provider + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProvidersAPI + get: + operationId: provider_providers_get + summary: Get provider profile + tags: + - marketplace + - provider_providers + description: |- + Get provider profile + + :param id: str + + :returns: :class:`GetProviderResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetProviderResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetProviderResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProvidersAPI + put: + operationId: provider_providers_update + summary: Update provider profile + tags: + - marketplace + - provider_providers + description: |- + Update provider profile + + :param id: str + :param provider: :class:`ProviderInfo` + + :returns: :class:`UpdateProviderResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':param provider: :class:`ProviderInfo`' + requestBody: + content: + application/json: + schema: + type: object + properties: + provider: + type: string + required: + - provider + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateProviderResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProvidersAPI + /api/2.0/marketplace-provider/providers: + get: + operationId: provider_providers_list + summary: List provider profiles for account. + tags: + - marketplace + - provider_providers + description: |- + List provider profiles for account. + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`ProviderInfo` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchGetProvidersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProviderProvidersAPI +components: + schemas: + AddExchangeForListingResponse: + type: object + properties: + exchange_for_listing: + $ref: '#/components/schemas/ExchangeListing' + BatchGetListingsResponse: + type: object + properties: + listings: + type: array + items: + $ref: '#/components/schemas/Listing' + BatchGetProvidersResponse: + type: object + properties: + providers: + type: array + items: + $ref: '#/components/schemas/ProviderInfo' + ConsumerTerms: + type: object + properties: + version: + type: string + required: + - version + ContactInfo: + type: object + properties: + company: + type: string + email: + type: string + first_name: + type: string + last_name: + type: string + description: >- + contact info for the consumer requesting data or performing a listing + installation + CreateExchangeFilterResponse: + type: object + properties: + filter_id: + type: string + CreateExchangeResponse: + type: object + properties: + exchange_id: + type: string + CreateFileResponse: + type: object + properties: + file_info: + $ref: '#/components/schemas/FileInfo' + upload_url: + type: string + description: Pre-signed POST URL to blob storage + CreateListingResponse: + type: object + properties: + listing_id: + type: string + CreatePersonalizationRequestResponse: + type: object + properties: + id: + type: string + CreateProviderResponse: + type: object + properties: + id: + type: string + DataRefreshInfo: + type: object + properties: + interval: + type: integer + unit: + $ref: '#/components/schemas/DataRefresh' + required: + - interval + - unit + DeleteExchangeFilterResponse: + type: object + properties: {} + DeleteExchangeResponse: + type: object + properties: {} + DeleteFileResponse: + type: object + properties: {} + DeleteInstallationResponse: + type: object + properties: {} + DeleteListingResponse: + type: object + properties: {} + DeleteProviderResponse: + type: object + properties: {} + Exchange: + type: object + properties: + name: + type: string + comment: + type: string + created_at: + type: integer + created_by: + type: string + filters: + type: array + items: + $ref: '#/components/schemas/ExchangeFilter' + id: + type: string + linked_listings: + type: array + items: + $ref: '#/components/schemas/ExchangeListing' + updated_at: + type: integer + updated_by: + type: string + required: + - name + ExchangeFilter: + type: object + properties: + exchange_id: + type: string + filter_value: + type: string + filter_type: + $ref: '#/components/schemas/ExchangeFilterType' + created_at: + type: integer + created_by: + type: string + id: + type: string + name: + type: string + updated_at: + type: integer + updated_by: + type: string + required: + - exchange_id + - filter_value + - filter_type + ExchangeListing: + type: object + properties: + created_at: + type: integer + created_by: + type: string + exchange_id: + type: string + exchange_name: + type: string + id: + type: string + listing_id: + type: string + listing_name: + type: string + FileInfo: + type: object + properties: + created_at: + type: integer + display_name: + type: string + description: >- + Name displayed to users for applicable files, e.g. embedded + notebooks + download_link: + type: string + file_parent: + $ref: '#/components/schemas/FileParent' + id: + type: string + marketplace_file_type: + $ref: '#/components/schemas/MarketplaceFileType' + mime_type: + type: string + status: + $ref: '#/components/schemas/FileStatus' + status_message: + type: string + description: >- + Populated if status is in a failed state with more information on + reason for the failure. + updated_at: + type: integer + FileParent: + type: object + properties: + file_parent_type: + $ref: '#/components/schemas/FileParentType' + parent_id: + type: string + description: TODO make the following fields required + GetExchangeResponse: + type: object + properties: + exchange: + $ref: '#/components/schemas/Exchange' + GetFileResponse: + type: object + properties: + file_info: + $ref: '#/components/schemas/FileInfo' + GetLatestVersionProviderAnalyticsDashboardResponse: + type: object + properties: + version: + type: integer + GetListingContentMetadataResponse: + type: object + properties: + next_page_token: + type: string + shared_data_objects: + type: array + items: + $ref: '#/components/schemas/SharedDataObject' + GetListingResponse: + type: object + properties: + listing: + $ref: '#/components/schemas/Listing' + GetListingsResponse: + type: object + properties: + listings: + type: array + items: + $ref: '#/components/schemas/Listing' + next_page_token: + type: string + GetPersonalizationRequestResponse: + type: object + properties: + personalization_requests: + type: array + items: + $ref: '#/components/schemas/PersonalizationRequest' + GetProviderResponse: + type: object + properties: + provider: + $ref: '#/components/schemas/ProviderInfo' + Installation: + type: object + properties: + installation: + $ref: '#/components/schemas/InstallationDetail' + InstallationDetail: + type: object + properties: + catalog_name: + type: string + error_message: + type: string + id: + type: string + installed_on: + type: integer + listing_id: + type: string + listing_name: + type: string + recipient_type: + $ref: '#/components/schemas/DeltaSharingRecipientType' + repo_name: + type: string + repo_path: + type: string + share_name: + type: string + status: + $ref: '#/components/schemas/InstallationStatus' + token_detail: + $ref: '#/components/schemas/TokenDetail' + tokens: + type: array + items: + $ref: '#/components/schemas/TokenInfo' + ListAllInstallationsResponse: + type: object + properties: + installations: + type: array + items: + $ref: '#/components/schemas/InstallationDetail' + next_page_token: + type: string + ListAllPersonalizationRequestsResponse: + type: object + properties: + next_page_token: + type: string + personalization_requests: + type: array + items: + $ref: '#/components/schemas/PersonalizationRequest' + ListExchangeFiltersResponse: + type: object + properties: + filters: + type: array + items: + $ref: '#/components/schemas/ExchangeFilter' + next_page_token: + type: string + ListExchangesForListingResponse: + type: object + properties: + exchange_listing: + type: array + items: + $ref: '#/components/schemas/ExchangeListing' + next_page_token: + type: string + ListExchangesResponse: + type: object + properties: + exchanges: + type: array + items: + $ref: '#/components/schemas/Exchange' + next_page_token: + type: string + ListFilesResponse: + type: object + properties: + file_infos: + type: array + items: + $ref: '#/components/schemas/FileInfo' + next_page_token: + type: string + ListFulfillmentsResponse: + type: object + properties: + fulfillments: + type: array + items: + $ref: '#/components/schemas/ListingFulfillment' + next_page_token: + type: string + ListInstallationsResponse: + type: object + properties: + installations: + type: array + items: + $ref: '#/components/schemas/InstallationDetail' + next_page_token: + type: string + ListListingsForExchangeResponse: + type: object + properties: + exchange_listings: + type: array + items: + $ref: '#/components/schemas/ExchangeListing' + next_page_token: + type: string + ListListingsResponse: + type: object + properties: + listings: + type: array + items: + $ref: '#/components/schemas/Listing' + next_page_token: + type: string + ListProviderAnalyticsDashboardResponse: + type: object + properties: + id: + type: string + dashboard_id: + type: string + description: dashboard_id will be used to open Lakeview dashboard. + version: + type: integer + required: + - id + - dashboard_id + ListProvidersResponse: + type: object + properties: + next_page_token: + type: string + providers: + type: array + items: + $ref: '#/components/schemas/ProviderInfo' + Listing: + type: object + properties: + summary: + $ref: '#/components/schemas/ListingSummary' + detail: + $ref: '#/components/schemas/ListingDetail' + id: + type: string + required: + - summary + ListingDetail: + type: object + properties: + assets: + type: array + items: + $ref: '#/components/schemas/AssetType' + collection_date_end: + type: integer + description: The ending date timestamp for when the data spans + collection_date_start: + type: integer + description: The starting date timestamp for when the data spans + collection_granularity: + $ref: '#/components/schemas/DataRefreshInfo' + description: Smallest unit of time in the dataset + cost: + $ref: '#/components/schemas/Cost' + description: Whether the dataset is free or paid + data_source: + type: string + description: Where/how the data is sourced + description: + type: string + documentation_link: + type: string + embedded_notebook_file_infos: + type: array + items: + $ref: '#/components/schemas/FileInfo' + file_ids: + type: array + items: + type: string + geographical_coverage: + type: string + description: Which geo region the listing data is collected from + license: + type: string + description: >- + ID 20, 21 removed don't use License of the data asset - Required for + listings with model based assets + pricing_model: + type: string + description: >- + What the pricing model is (e.g. paid, subscription, paid upfront); + should only be present if cost is paid TODO: Not used yet, should + deprecate if we will never use it + privacy_policy_link: + type: string + size: + type: number + description: size of the dataset in GB + support_link: + type: string + tags: + type: array + items: + $ref: '#/components/schemas/ListingTag' + description: >- + Listing tags - Simple key value pair to annotate listings. When + should I use tags vs dedicated fields? Using tags avoids the need to + add new columns in the database for new annotations. However, this + should be used sparingly since tags are stored as key value pair. + Use tags only: 1. If the field is optional and won't need to have + NOT NULL integrity check 2. The value is fairly fixed, static and + low cardinality (eg. enums). 3. The value won't be used in filters + or joins with other tables. + terms_of_service: + type: string + update_frequency: + $ref: '#/components/schemas/DataRefreshInfo' + description: How often data is updated + ListingFulfillment: + type: object + properties: + listing_id: + type: string + fulfillment_type: + $ref: '#/components/schemas/FulfillmentType' + recipient_type: + $ref: '#/components/schemas/DeltaSharingRecipientType' + repo_info: + $ref: '#/components/schemas/RepoInfo' + share_info: + $ref: '#/components/schemas/ShareInfo' + required: + - listing_id + ListingSetting: + type: object + properties: + visibility: + $ref: '#/components/schemas/Visibility' + ListingSummary: + type: object + properties: + name: + type: string + listingType: + $ref: '#/components/schemas/ListingType' + categories: + type: array + items: + $ref: '#/components/schemas/Category' + created_at: + type: integer + created_by: + type: string + created_by_id: + type: integer + exchange_ids: + type: array + items: + type: string + git_repo: + $ref: '#/components/schemas/RepoInfo' + description: >- + if a git repo is being created, a listing will be initialized with + this field as opposed to a share + provider_id: + type: string + provider_region: + $ref: '#/components/schemas/RegionInfo' + published_at: + type: integer + published_by: + type: string + setting: + $ref: '#/components/schemas/ListingSetting' + share: + $ref: '#/components/schemas/ShareInfo' + status: + $ref: '#/components/schemas/ListingStatus' + subtitle: + type: string + updated_at: + type: integer + updated_by: + type: string + updated_by_id: + type: integer + required: + - name + - listingType + ListingTag: + type: object + properties: + tag_name: + $ref: '#/components/schemas/ListingTagType' + tag_values: + type: array + items: + type: string + description: >- + String representation of the tag value. Values should be string + literals (no complex types) + PersonalizationRequest: + type: object + properties: + consumer_region: + $ref: '#/components/schemas/RegionInfo' + comment: + type: string + contact_info: + $ref: '#/components/schemas/ContactInfo' + created_at: + type: integer + id: + type: string + intended_use: + type: string + is_from_lighthouse: + type: boolean + listing_id: + type: string + listing_name: + type: string + metastore_id: + type: string + provider_id: + type: string + recipient_type: + $ref: '#/components/schemas/DeltaSharingRecipientType' + share: + $ref: '#/components/schemas/ShareInfo' + description: >- + Share information is required for data listings but should be + empty/ignored for non-data listings (MCP and App). + status: + $ref: '#/components/schemas/PersonalizationRequestStatus' + status_message: + type: string + updated_at: + type: integer + required: + - consumer_region + ProviderAnalyticsDashboard: + type: object + properties: + id: + type: string + required: + - id + ProviderInfo: + type: object + properties: + name: + type: string + business_contact_email: + type: string + term_of_service_link: + type: string + privacy_policy_link: + type: string + company_website_link: + type: string + dark_mode_icon_file_id: + type: string + dark_mode_icon_file_path: + type: string + description: + type: string + icon_file_id: + type: string + icon_file_path: + type: string + id: + type: string + is_featured: + type: boolean + description: is_featured is accessible by consumers only + published_by: + type: string + description: published_by is only applicable to data aggregators (e.g. Crux) + support_contact_email: + type: string + required: + - name + - business_contact_email + - term_of_service_link + - privacy_policy_link + RegionInfo: + type: object + properties: + cloud: + type: string + region: + type: string + RemoveExchangeForListingResponse: + type: object + properties: {} + RepoInfo: + type: object + properties: + git_repo_url: + type: string + required: + - git_repo_url + RepoInstallation: + type: object + properties: + repo_name: + type: string + repo_path: + type: string + description: >- + refers to the full url file path that navigates the user to the + repo's entrypoint (e.g. a README.md file, or the repo file view in + the unified UI) should just be a relative path + required: + - repo_name + - repo_path + SearchListingsResponse: + type: object + properties: + listings: + type: array + items: + $ref: '#/components/schemas/Listing' + next_page_token: + type: string + ShareInfo: + type: object + properties: + name: + type: string + type: + $ref: '#/components/schemas/ListingShareType' + required: + - name + - type + SharedDataObject: + type: object + properties: + data_object_type: + type: string + name: + type: string + description: Name of the shared object + TokenDetail: + type: object + properties: + bearerToken: + type: string + endpoint: + type: string + expirationTime: + type: string + shareCredentialsVersion: + type: integer + description: >- + These field names must follow the delta sharing protocol. Original + message: RetrieveToken.Response in + managed-catalog/api/messages/recipient.proto + TokenInfo: + type: object + properties: + activation_url: + type: string + created_at: + type: integer + description: >- + Time at which this Recipient Token was created, in epoch + milliseconds. + created_by: + type: string + description: Username of Recipient Token creator. + expiration_time: + type: integer + description: Expiration timestamp of the token in epoch milliseconds. + id: + type: string + description: Unique id of the Recipient Token. + updated_at: + type: integer + description: >- + Time at which this Recipient Token was updated, in epoch + milliseconds. + updated_by: + type: string + description: Username of Recipient Token updater. + UpdateExchangeFilterResponse: + type: object + properties: + filter: + $ref: '#/components/schemas/ExchangeFilter' + UpdateExchangeResponse: + type: object + properties: + exchange: + $ref: '#/components/schemas/Exchange' + UpdateInstallationResponse: + type: object + properties: + installation: + $ref: '#/components/schemas/InstallationDetail' + UpdateListingResponse: + type: object + properties: + listing: + $ref: '#/components/schemas/Listing' + UpdatePersonalizationRequestResponse: + type: object + properties: + request: + $ref: '#/components/schemas/PersonalizationRequest' + UpdateProviderAnalyticsDashboardResponse: + type: object + properties: + id: + type: string + dashboard_id: + type: string + description: this is newly created Lakeview dashboard for the user + version: + type: integer + required: + - id + - dashboard_id + UpdateProviderResponse: + type: object + properties: + provider: + $ref: '#/components/schemas/ProviderInfo' + AssetType: + type: string + enum: + - ASSET_TYPE_APP + - ASSET_TYPE_DATA_TABLE + - ASSET_TYPE_GIT_REPO + - ASSET_TYPE_MCP + - ASSET_TYPE_MEDIA + - ASSET_TYPE_MODEL + - ASSET_TYPE_NOTEBOOK + - ASSET_TYPE_PARTNER_INTEGRATION + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + Category: + type: string + enum: + - ADVERTISING_AND_MARKETING + - CLIMATE_AND_ENVIRONMENT + - COMMERCE + - DEMOGRAPHICS + - ECONOMICS + - EDUCATION + - ENERGY + - FINANCIAL + - GAMING + - GEOSPATIAL + - HEALTH + - LOOKUP_TABLES + - MANUFACTURING + - MEDIA + - OTHER + - PUBLIC_SECTOR + - RETAIL + - SCIENCE_AND_RESEARCH + - SECURITY + - SPORTS + - TRANSPORTATION_AND_LOGISTICS + - TRAVEL_AND_TOURISM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + Cost: + type: string + enum: + - FREE + - PAID + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DataRefresh: + type: string + enum: + - DAILY + - HOURLY + - MINUTE + - MONTHLY + - NONE + - QUARTERLY + - SECOND + - WEEKLY + - YEARLY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DeltaSharingRecipientType: + type: string + enum: + - DELTA_SHARING_RECIPIENT_TYPE_DATABRICKS + - DELTA_SHARING_RECIPIENT_TYPE_OPEN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ExchangeFilterType: + type: string + enum: + - GLOBAL_METASTORE_ID + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FileParentType: + type: string + enum: + - LISTING + - LISTING_RESOURCE + - PROVIDER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FileStatus: + type: string + enum: + - FILE_STATUS_PUBLISHED + - FILE_STATUS_SANITIZATION_FAILED + - FILE_STATUS_SANITIZING + - FILE_STATUS_STAGING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FulfillmentType: + type: string + enum: + - INSTALL + - REQUEST_ACCESS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + InstallationStatus: + type: string + enum: + - FAILED + - INSTALLED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListingShareType: + type: string + enum: + - FULL + - SAMPLE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListingStatus: + type: string + enum: + - DRAFT + - PENDING + - PUBLISHED + - SUSPENDED + description: Enums + ListingTagType: + type: string + enum: + - LISTING_TAG_TYPE_LANGUAGE + - LISTING_TAG_TYPE_TASK + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListingType: + type: string + enum: + - PERSONALIZED + - STANDARD + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MarketplaceFileType: + type: string + enum: + - APP + - EMBEDDED_NOTEBOOK + - PROVIDER_ICON + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PersonalizationRequestStatus: + type: string + enum: + - DENIED + - FULFILLED + - NEW + - REQUEST_PENDING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + Visibility: + type: string + enum: + - PRIVATE + - PUBLIC + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + consumer_listing_content: + id: databricks_workspace.marketplace.consumer_listing_content + name: consumer_listing_content + title: Consumer Listing Content + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1content/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.shared_data_objects + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/consumer_listing_content/methods/get + insert: [] + update: [] + delete: [] + replace: [] + consumer_fulfillments: + id: databricks_workspace.marketplace.consumer_fulfillments + name: consumer_fulfillments + title: Consumer Fulfillments + methods: + list: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1fulfillments/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.fulfillments + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/consumer_fulfillments/methods/list + insert: [] + update: [] + delete: [] + replace: [] + consumer_installations: + id: databricks_workspace.marketplace.consumer_installations + name: consumer_installations + title: Consumer Installations + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations/post + response: + mediaType: application/json + openAPIDocKey: '200' + list_for_listing: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.installations + delete: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations~1{installation_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1installations~1{installation_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1installations/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.installations + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/consumer_installations/methods/list_for_listing + - $ref: >- + #/components/x-stackQL-resources/consumer_installations/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/consumer_installations/methods/create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/consumer_installations/methods/delete + replace: + - $ref: >- + #/components/x-stackQL-resources/consumer_installations/methods/update + consumer_listings: + id: databricks_workspace.marketplace.consumer_listings + name: consumer_listings + title: Consumer Listings + methods: + batch_get: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings:batchGet/get' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1listings/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.listings + search: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1search-listings/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.listings + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/consumer_listings/methods/get' + - $ref: '#/components/x-stackQL-resources/consumer_listings/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + consumer_personalization_requests: + id: databricks_workspace.marketplace.consumer_personalization_requests + name: consumer_personalization_requests + title: Consumer Personalization Requests + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1personalization-requests/post + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1listings~1{listing_id}~1personalization-requests/get + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.1~1marketplace-consumer~1personalization-requests/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.personalization_requests + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/consumer_personalization_requests/methods/get + - $ref: >- + #/components/x-stackQL-resources/consumer_personalization_requests/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/consumer_personalization_requests/methods/create + update: [] + delete: [] + replace: [] + consumer_providers: + id: databricks_workspace.marketplace.consumer_providers + name: consumer_providers + title: Consumer Providers + methods: + batch_get: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1providers:batchGet/get' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1providers~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1marketplace-consumer~1providers/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.providers + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/consumer_providers/methods/get' + - $ref: '#/components/x-stackQL-resources/consumer_providers/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + provider_exchange_filters: + id: databricks_workspace.marketplace.provider_exchange_filters + name: provider_exchange_filters + title: Provider Exchange Filters + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.filters + delete: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1filters~1{id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_filters/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_filters/methods/create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_filters/methods/delete + replace: + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_filters/methods/update + provider_exchange_listings: + id: databricks_workspace.marketplace.provider_exchange_listings + name: provider_exchange_listings + title: Provider Exchange Listings + methods: + add: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-exchange~1exchanges-for-listing/post + response: + mediaType: application/json + openAPIDocKey: '200' + list_for_listing: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-exchange~1exchanges-for-listing/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.exchange_listing + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-exchange~1exchanges-for-listing~1{id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + list_for_exchange: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-exchange~1listings-for-exchange/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.exchange_listings + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_listings/methods/list_for_listing + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_listings/methods/list_for_exchange + insert: + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_listings/methods/add + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/provider_exchange_listings/methods/delete + replace: [] + provider_exchanges: + id: databricks_workspace.marketplace.provider_exchanges + name: provider_exchanges + title: Provider Exchanges + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.exchanges + delete: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-exchange~1exchanges~1{id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/get' + - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/provider_exchanges/methods/update' + provider_files: + id: databricks_workspace.marketplace.provider_files + name: provider_files + title: Provider Files + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1files/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1files/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.file_infos + delete: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1files~1{file_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1files~1{file_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/provider_files/methods/get' + - $ref: '#/components/x-stackQL-resources/provider_files/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/provider_files/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/provider_files/methods/delete' + replace: [] + provider_listings: + id: databricks_workspace.marketplace.provider_listings + name: provider_listings + title: Provider Listings + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1listing/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings~1{id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1listings/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.listings + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/provider_listings/methods/get' + - $ref: '#/components/x-stackQL-resources/provider_listings/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/provider_listings/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/provider_listings/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/provider_listings/methods/update' + provider_personalization_requests: + id: databricks_workspace.marketplace.provider_personalization_requests + name: provider_personalization_requests + title: Provider Personalization Requests + methods: + list: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-provider~1personalization-requests/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.personalization_requests + update: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-provider~1listings~1{listing_id}~1personalization-requests~1{request_id}~1request-status/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/provider_personalization_requests/methods/list + insert: [] + update: [] + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/provider_personalization_requests/methods/update + provider_provider_analytics_dashboards: + id: databricks_workspace.marketplace.provider_provider_analytics_dashboards + name: provider_provider_analytics_dashboards + title: Provider Provider Analytics Dashboards + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard/get' + response: + mediaType: application/json + openAPIDocKey: '200' + get_latest_version: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard~1latest/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1marketplace-provider~1analytics_dashboard~1{id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/provider_provider_analytics_dashboards/methods/get + insert: + - $ref: >- + #/components/x-stackQL-resources/provider_provider_analytics_dashboards/methods/create + update: [] + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/provider_provider_analytics_dashboards/methods/update + provider_providers: + id: databricks_workspace.marketplace.provider_providers + name: provider_providers + title: Provider Providers + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1provider/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers~1{id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1marketplace-provider~1providers/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.providers + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/provider_providers/methods/get' + - $ref: '#/components/x-stackQL-resources/provider_providers/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/provider_providers/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/provider_providers/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/provider_providers/methods/update' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/ml.yaml b/providers/src/databricks_workspace/v00.00.00000/services/ml.yaml new file mode 100644 index 00000000..757b21e0 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/ml.yaml @@ -0,0 +1,9830 @@ +openapi: 3.0.0 +info: + title: Databricks Ml API (workspace) + description: >- + OpenAPI specification for the Databricks ml service (workspace-level APIs), + generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.ml +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/mlflow/experiments/create: + post: + operationId: experiments_create_experiment + summary: >- + Creates an experiment with a name. Returns the ID of the newly created + experiment. Validates that + tags: + - ml + - experiments + description: >- + Creates an experiment with a name. Returns the ID of the newly created + experiment. Validates that + + another experiment with the same name does not already exist and fails + if another experiment with the + + same name already exists. + + + Throws `RESOURCE_ALREADY_EXISTS` if an experiment with the given name + exists. + + + :param name: str + Experiment name. + :param artifact_location: str (optional) + Location where all artifacts for the experiment are stored. If not provided, the remote server will + select an appropriate default. + :param tags: List[:class:`ExperimentTag`] (optional) + A collection of tags to set on the experiment. Maximum tag size and number of tags per request + depends on the storage backend. All storage backends are guaranteed to support tag keys up to 250 + bytes in size and tag values up to 5000 bytes in size. All storage backends are also guaranteed to + support up to 20 tags per request. + + :returns: :class:`CreateExperimentResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Experiment name. + artifact_location: + type: string + description: >- + Location where all artifacts for the experiment are stored. + If not provided, the remote server will select an + appropriate default. + tags: + type: string + description: >- + A collection of tags to set on the experiment. Maximum tag + size and number of tags per request depends on the storage + backend. All storage backends are guaranteed to support tag + keys up to 250 bytes in size and tag values up to 5000 bytes + in size. All storage backends are also guaranteed to support + up to 20 tags per request. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateExperimentResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/logged-models: + post: + operationId: experiments_create_logged_model + summary: Create a logged model. + tags: + - ml + - experiments + description: |- + Create a logged model. + + :param experiment_id: str + The ID of the experiment that owns the model. + :param model_type: str (optional) + The type of the model, such as ``"Agent"``, ``"Classifier"``, ``"LLM"``. + :param name: str (optional) + The name of the model (optional). If not specified one will be generated. + :param params: List[:class:`LoggedModelParameter`] (optional) + Parameters attached to the model. + :param source_run_id: str (optional) + The ID of the run that created the model. + :param tags: List[:class:`LoggedModelTag`] (optional) + Tags attached to the model. + + :returns: :class:`CreateLoggedModelResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: The ID of the experiment that owns the model. + model_type: + type: string + description: >- + The type of the model, such as ``"Agent"``, + ``"Classifier"``, ``"LLM"``. + name: + type: string + description: >- + The name of the model (optional). If not specified one will + be generated. + params: + type: string + description: Parameters attached to the model. + source_run_id: + type: string + description: The ID of the run that created the model. + tags: + type: string + description: Tags attached to the model. + required: + - experiment_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateLoggedModelResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/create: + post: + operationId: experiments_create_run + summary: >- + Creates a new run within an experiment. A run is usually a single + execution of a machine learning or + tags: + - ml + - experiments + description: >- + Creates a new run within an experiment. A run is usually a single + execution of a machine learning or + + data ETL pipeline. MLflow uses runs to track the `mlflowParam`, + `mlflowMetric`, and `mlflowRunTag` + + associated with a single execution. + + + :param experiment_id: str (optional) + ID of the associated experiment. + :param run_name: str (optional) + The name of the run. + :param start_time: int (optional) + Unix timestamp in milliseconds of when the run started. + :param tags: List[:class:`RunTag`] (optional) + Additional metadata for run. + :param user_id: str (optional) + ID of the user executing the run. This field is deprecated as of MLflow 1.0, and will be removed in + a future MLflow release. Use 'mlflow.user' tag instead. + + :returns: :class:`CreateRunResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: ID of the associated experiment. + run_name: + type: string + description: The name of the run. + start_time: + type: string + description: Unix timestamp in milliseconds of when the run started. + tags: + type: string + description: Additional metadata for run. + user_id: + type: string + description: >- + ID of the user executing the run. This field is deprecated + as of MLflow 1.0, and will be removed in a future MLflow + release. Use 'mlflow.user' tag instead. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateRunResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/delete: + post: + operationId: experiments_delete_experiment + summary: >- + Marks an experiment and associated metadata, runs, metrics, params, and + tags for deletion. If the + tags: + - ml + - experiments + description: >- + Marks an experiment and associated metadata, runs, metrics, params, and + tags for deletion. If the + + experiment uses FileStore, artifacts associated with the experiment are + also deleted. + + + :param experiment_id: str + ID of the associated experiment. + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: ID of the associated experiment. + required: + - experiment_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/logged-models/{model_id}: + delete: + operationId: experiments_delete_logged_model + summary: Delete a logged model. + tags: + - ml + - experiments + description: |- + Delete a logged model. + + :param model_id: str + The ID of the logged model to delete. + parameters: + - name: model_id + in: path + required: true + schema: + type: string + description: The ID of the logged model to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + patch: + operationId: experiments_finalize_logged_model + summary: Finalize a logged model. + tags: + - ml + - experiments + description: |- + Finalize a logged model. + + :param model_id: str + The ID of the logged model to finalize. + :param status: :class:`LoggedModelStatus` + Whether or not the model is ready for use. ``"LOGGED_MODEL_UPLOAD_FAILED"`` indicates that something + went wrong when logging the model weights / agent code. + + :returns: :class:`FinalizeLoggedModelResponse` + parameters: + - name: model_id + in: path + required: true + schema: + type: string + description: The ID of the logged model to finalize. + requestBody: + content: + application/json: + schema: + type: object + properties: + status: + type: string + description: >- + Whether or not the model is ready for use. + ``"LOGGED_MODEL_UPLOAD_FAILED"`` indicates that something + went wrong when logging the model weights / agent code. + required: + - status + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FinalizeLoggedModelResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + get: + operationId: experiments_get_logged_model + summary: Get a logged model. + tags: + - ml + - experiments + description: |- + Get a logged model. + + :param model_id: str + The ID of the logged model to retrieve. + + :returns: :class:`GetLoggedModelResponse` + parameters: + - name: model_id + in: path + required: true + schema: + type: string + description: The ID of the logged model to retrieve. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetLoggedModelResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/logged-models/{model_id}/tags/{tag_key}: + delete: + operationId: experiments_delete_logged_model_tag + summary: Delete a tag on a logged model. + tags: + - ml + - experiments + description: |- + Delete a tag on a logged model. + + :param model_id: str + The ID of the logged model to delete the tag from. + :param tag_key: str + The tag key. + parameters: + - name: model_id + in: path + required: true + schema: + type: string + description: The ID of the logged model to delete the tag from. + - name: tag_key + in: path + required: true + schema: + type: string + description: The tag key. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/delete: + post: + operationId: experiments_delete_run + summary: Marks a run for deletion. + tags: + - ml + - experiments + description: |- + Marks a run for deletion. + + :param run_id: str + ID of the run to delete. + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: string + description: ID of the run to delete. + required: + - run_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/databricks/runs/delete-runs: + post: + operationId: experiments_delete_runs + summary: >- + Bulk delete runs in an experiment that were created prior to or at the + specified timestamp. Deletes at + tags: + - ml + - experiments + description: >- + Bulk delete runs in an experiment that were created prior to or at the + specified timestamp. Deletes at + + most max_runs per request. To call this API from a Databricks Notebook + in Python, you can use the + + client code snippet on + + + :param experiment_id: str + The ID of the experiment containing the runs to delete. + :param max_timestamp_millis: int + The maximum creation timestamp in milliseconds since the UNIX epoch for deleting runs. Only runs + created prior to or at this timestamp are deleted. + :param max_runs: int (optional) + An optional positive integer indicating the maximum number of runs to delete. The maximum allowed + value for max_runs is 10000. + + :returns: :class:`DeleteRunsResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: The ID of the experiment containing the runs to delete. + max_timestamp_millis: + type: integer + description: >- + The maximum creation timestamp in milliseconds since the + UNIX epoch for deleting runs. Only runs created prior to or + at this timestamp are deleted. + max_runs: + type: string + description: >- + An optional positive integer indicating the maximum number + of runs to delete. The maximum allowed value for max_runs is + 10000. + required: + - experiment_id + - max_timestamp_millis + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteRunsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/delete-tag: + post: + operationId: experiments_delete_tag + summary: >- + Deletes a tag on a run. Tags are run metadata that can be updated during + a run and after a run + tags: + - ml + - experiments + description: >- + Deletes a tag on a run. Tags are run metadata that can be updated during + a run and after a run + + completes. + + + :param run_id: str + ID of the run that the tag was logged under. Must be provided. + :param key: str + Name of the tag. Maximum size is 255 bytes. Must be provided. + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: string + description: >- + ID of the run that the tag was logged under. Must be + provided. + key: + type: string + description: >- + Name of the tag. Maximum size is 255 bytes. Must be + provided. + required: + - run_id + - key + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/get-by-name: + get: + operationId: experiments_get_by_name + summary: Gets metadata for an experiment. + tags: + - ml + - experiments + description: >- + Gets metadata for an experiment. + + + This endpoint will return deleted experiments, but prefers the active + experiment if an active and + + deleted experiment share the same name. If multiple deleted experiments + share the same name, the API + + will return one of them. + + + Throws `RESOURCE_DOES_NOT_EXIST` if no experiment with the specified + name exists. + + + :param experiment_name: str + Name of the associated experiment. + + :returns: :class:`GetExperimentByNameResponse` + parameters: + - name: experiment_name + in: query + required: true + schema: + type: string + description: Name of the associated experiment. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetExperimentByNameResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/get: + get: + operationId: experiments_get_experiment + summary: >- + Gets metadata for an experiment. This method works on deleted + experiments. + tags: + - ml + - experiments + description: >- + Gets metadata for an experiment. This method works on deleted + experiments. + + + :param experiment_id: str + ID of the associated experiment. + + :returns: :class:`GetExperimentResponse` + parameters: + - name: experiment_id + in: query + required: true + schema: + type: string + description: ID of the associated experiment. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetExperimentResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/metrics/get-history: + get: + operationId: experiments_get_history + summary: Gets a list of all values for the specified metric for a given run. + tags: + - ml + - experiments + description: |- + Gets a list of all values for the specified metric for a given run. + + :param metric_key: str + Name of the metric. + :param max_results: int (optional) + Maximum number of Metric records to return per paginated request. Default is set to 25,000. If set + higher than 25,000, a request Exception will be raised. + :param page_token: str (optional) + Token indicating the page of metric histories to fetch. + :param run_id: str (optional) + ID of the run from which to fetch metric values. Must be provided. + :param run_uuid: str (optional) + [Deprecated, use `run_id` instead] ID of the run from which to fetch metric values. This field will + be removed in a future MLflow version. + + :returns: Iterator over :class:`Metric` + parameters: + - name: metric_key + in: query + required: true + schema: + type: string + description: Name of the metric. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of Metric records to return per paginated request. + Default is set to 25,000. If set higher than 25,000, a request + Exception will be raised. + - name: page_token + in: query + required: false + schema: + type: string + description: Token indicating the page of metric histories to fetch. + - name: run_id + in: query + required: false + schema: + type: string + description: ID of the run from which to fetch metric values. Must be provided. + - name: run_uuid + in: query + required: false + schema: + type: string + description: >- + [Deprecated, use `run_id` instead] ID of the run from which to fetch + metric values. This field will be removed in a future MLflow + version. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetMetricHistoryResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/permissions/experiments/{experiment_id}/permissionLevels: + get: + operationId: experiments_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - ml + - experiments + description: |- + Gets the permission levels that a user can have on an object. + + :param experiment_id: str + The experiment for which to get or manage permissions. + + :returns: :class:`GetExperimentPermissionLevelsResponse` + parameters: + - name: experiment_id + in: path + required: true + schema: + type: string + description: The experiment for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetExperimentPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/permissions/experiments/{experiment_id}: + get: + operationId: experiments_get_permissions + summary: >- + Gets the permissions of an experiment. Experiments can inherit + permissions from their root object. + tags: + - ml + - experiments + description: >- + Gets the permissions of an experiment. Experiments can inherit + permissions from their root object. + + + :param experiment_id: str + The experiment for which to get or manage permissions. + + :returns: :class:`ExperimentPermissions` + parameters: + - name: experiment_id + in: path + required: true + schema: + type: string + description: The experiment for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExperimentPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + put: + operationId: experiments_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - ml + - experiments + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param experiment_id: str + The experiment for which to get or manage permissions. + :param access_control_list: + List[:class:`ExperimentAccessControlRequest`] (optional) + + + :returns: :class:`ExperimentPermissions` + parameters: + - name: experiment_id + in: path + required: true + schema: + type: string + description: The experiment for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ExperimentPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExperimentPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + patch: + operationId: experiments_update_permissions + summary: >- + Updates the permissions on an experiment. Experiments can inherit + permissions from their root object. + tags: + - ml + - experiments + description: >- + Updates the permissions on an experiment. Experiments can inherit + permissions from their root object. + + + :param experiment_id: str + The experiment for which to get or manage permissions. + :param access_control_list: + List[:class:`ExperimentAccessControlRequest`] (optional) + + + :returns: :class:`ExperimentPermissions` + parameters: + - name: experiment_id + in: path + required: true + schema: + type: string + description: The experiment for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ExperimentPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExperimentPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/get: + get: + operationId: experiments_get_run + summary: >- + Gets the metadata, metrics, params, and tags for a run. In the case + where multiple metrics with the + tags: + - ml + - experiments + description: >- + Gets the metadata, metrics, params, and tags for a run. In the case + where multiple metrics with the + + same key are logged for a run, return only the value with the latest + timestamp. + + + If there are multiple values with the latest timestamp, return the + maximum of these values. + + + :param run_id: str + ID of the run to fetch. Must be provided. + :param run_uuid: str (optional) + [Deprecated, use `run_id` instead] ID of the run to fetch. This field will be removed in a future + MLflow version. + + :returns: :class:`GetRunResponse` + parameters: + - name: run_id + in: query + required: true + schema: + type: string + description: ID of the run to fetch. Must be provided. + - name: run_uuid + in: query + required: false + schema: + type: string + description: >- + [Deprecated, use `run_id` instead] ID of the run to fetch. This + field will be removed in a future MLflow version. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetRunResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/artifacts/list: + get: + operationId: experiments_list_artifacts + summary: >- + List artifacts for a run. Takes an optional `artifact_path` prefix which + if specified, the response + tags: + - ml + - experiments + description: >- + List artifacts for a run. Takes an optional `artifact_path` prefix which + if specified, the response + + contains only artifacts with the specified prefix. A maximum of 1000 + artifacts will be retrieved for + + UC Volumes. Please call `/api/2.0/fs/directories{directory_path}` for + listing artifacts in UC Volumes, + + which supports pagination. See [List directory contents | Files + + API](/api/workspace/files/listdirectorycontents). + + + :param page_token: str (optional) + The token indicating the page of artifact results to fetch. `page_token` is not supported when + listing artifacts in UC Volumes. A maximum of 1000 artifacts will be retrieved for UC Volumes. + Please call `/api/2.0/fs/directories{directory_path}` for listing artifacts in UC Volumes, which + supports pagination. See [List directory contents | Files + API](/api/workspace/files/listdirectorycontents). + :param path: str (optional) + Filter artifacts matching this path (a relative path from the root artifact directory). + :param run_id: str (optional) + ID of the run whose artifacts to list. Must be provided. + :param run_uuid: str (optional) + [Deprecated, use `run_id` instead] ID of the run whose artifacts to list. This field will be removed + in a future MLflow version. + + :returns: Iterator over :class:`FileInfo` + parameters: + - name: page_token + in: query + required: false + schema: + type: string + description: >- + The token indicating the page of artifact results to fetch. + `page_token` is not supported when listing artifacts in UC Volumes. + A maximum of 1000 artifacts will be retrieved for UC Volumes. Please + call `/api/2.0/fs/directories{directory_path}` for listing artifacts + in UC Volumes, which supports pagination. See [List directory + contents | Files API](/api/workspace/files/listdirectorycontents). + - name: path + in: query + required: false + schema: + type: string + description: >- + Filter artifacts matching this path (a relative path from the root + artifact directory). + - name: run_id + in: query + required: false + schema: + type: string + description: ID of the run whose artifacts to list. Must be provided. + - name: run_uuid + in: query + required: false + schema: + type: string + description: >- + [Deprecated, use `run_id` instead] ID of the run whose artifacts to + list. This field will be removed in a future MLflow version. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListArtifactsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/list: + get: + operationId: experiments_list_experiments + summary: Gets a list of all experiments. + tags: + - ml + - experiments + description: |- + Gets a list of all experiments. + + :param max_results: int (optional) + Maximum number of experiments desired. If `max_results` is unspecified, return all experiments. If + `max_results` is too large, it'll be automatically capped at 1000. Callers of this endpoint are + encouraged to pass max_results explicitly and leverage page_token to iterate through experiments. + :param page_token: str (optional) + Token indicating the page of experiments to fetch + :param view_type: :class:`ViewType` (optional) + Qualifier for type of experiments to be returned. If unspecified, return only active experiments. + + :returns: Iterator over :class:`Experiment` + parameters: + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of experiments desired. If `max_results` is + unspecified, return all experiments. If `max_results` is too large, + it'll be automatically capped at 1000. Callers of this endpoint are + encouraged to pass max_results explicitly and leverage page_token to + iterate through experiments. + - name: page_token + in: query + required: false + schema: + type: string + description: Token indicating the page of experiments to fetch + - name: view_type + in: query + required: false + schema: + type: string + description: >- + Qualifier for type of experiments to be returned. If unspecified, + return only active experiments. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExperimentsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/log-batch: + post: + operationId: experiments_log_batch + summary: >- + Logs a batch of metrics, params, and tags for a run. If any data failed + to be persisted, the server + tags: + - ml + - experiments + description: >- + Logs a batch of metrics, params, and tags for a run. If any data failed + to be persisted, the server + + will respond with an error (non-200 status code). + + + In case of error (due to internal server error or an invalid request), + partial data may be written. + + + You can write metrics, params, and tags in interleaving fashion, but + within a given entity type are + + guaranteed to follow the order specified in the request body. + + + The overwrite behavior for metrics, params, and tags is as follows: + + + * Metrics: metric values are never overwritten. Logging a metric (key, + value, timestamp) appends to + + the set of values for the metric with the provided key. + + + * Tags: tag values can be overwritten by successive writes to the same + tag key. That is, if multiple + + tag values with the same key are provided in the same API request, the + last-provided tag value is + + written. Logging the same tag (key, value) is permitted. Specifically, + logging a tag is idempotent. + + + * Parameters: once written, param values cannot be changed (attempting + to overwrite a param value will + + result in an error). However, logging the same param (key, value) is + permitted. Specifically, logging + + a param is idempotent. + + + Request Limits ------------------------------- A single JSON-serialized + API request may be up to 1 MB + + in size and contain: + + + * No more than 1000 metrics, params, and tags in total + + + * Up to 1000 metrics + + + * Up to 100 params + + + * Up to 100 tags + + + For example, a valid request might contain 900 metrics, 50 params, and + 50 tags, but logging 900 + + metrics, 50 params, and 51 tags is invalid. + + + The following limits also apply to metric, param, and tag keys and + values: + + + * Metric keys, param keys, and tag keys can be up to 250 characters in + length + + + * Parameter and tag values can be up to 250 characters in length + + + :param metrics: List[:class:`Metric`] (optional) + Metrics to log. A single request can contain up to 1000 metrics, and up to 1000 metrics, params, and + tags in total. + :param params: List[:class:`Param`] (optional) + Params to log. A single request can contain up to 100 params, and up to 1000 metrics, params, and + tags in total. + :param run_id: str (optional) + ID of the run to log under + :param tags: List[:class:`RunTag`] (optional) + Tags to log. A single request can contain up to 100 tags, and up to 1000 metrics, params, and tags + in total. + requestBody: + content: + application/json: + schema: + type: object + properties: + metrics: + type: string + description: >- + Metrics to log. A single request can contain up to 1000 + metrics, and up to 1000 metrics, params, and tags in total. + params: + type: string + description: >- + Params to log. A single request can contain up to 100 + params, and up to 1000 metrics, params, and tags in total. + run_id: + type: string + description: ID of the run to log under + tags: + type: string + description: >- + Tags to log. A single request can contain up to 100 tags, + and up to 1000 metrics, params, and tags in total. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/log-inputs: + post: + operationId: experiments_log_inputs + summary: Logs inputs, such as datasets and models, to an MLflow Run. + tags: + - ml + - experiments + description: |- + Logs inputs, such as datasets and models, to an MLflow Run. + + :param run_id: str + ID of the run to log under + :param datasets: List[:class:`DatasetInput`] (optional) + Dataset inputs + :param models: List[:class:`ModelInput`] (optional) + Model inputs + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: string + description: ID of the run to log under + datasets: + type: string + description: Dataset inputs + models: + type: string + description: Model inputs + required: + - run_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/logged-models/{model_id}/params: + post: + operationId: experiments_log_logged_model_params + summary: >- + Logs params for a logged model. A param is a key-value pair (string key, + string value). Examples + tags: + - ml + - experiments + description: >- + Logs params for a logged model. A param is a key-value pair (string key, + string value). Examples + + include hyperparameters used for ML model training. A param can be + logged only once for a logged + + model, and attempting to overwrite an existing param with a different + value will result in an error + + + :param model_id: str + The ID of the logged model to log params for. + :param params: List[:class:`LoggedModelParameter`] (optional) + Parameters to attach to the model. + parameters: + - name: model_id + in: path + required: true + schema: + type: string + description: The ID of the logged model to log params for. + requestBody: + content: + application/json: + schema: + type: object + properties: + params: + type: string + description: Parameters to attach to the model. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/log-metric: + post: + operationId: experiments_log_metric + summary: >- + Log a metric for a run. A metric is a key-value pair (string key, float + value) with an associated + tags: + - ml + - experiments + description: >- + Log a metric for a run. A metric is a key-value pair (string key, float + value) with an associated + + timestamp. Examples include the various metrics that represent ML model + accuracy. A metric can be + + logged multiple times. + + + :param key: str + Name of the metric. + :param value: float + Double value of the metric being logged. + :param timestamp: int + Unix timestamp in milliseconds at the time metric was logged. + :param dataset_digest: str (optional) + Dataset digest of the dataset associated with the metric, e.g. an md5 hash of the dataset that + uniquely identifies it within datasets of the same name. + :param dataset_name: str (optional) + The name of the dataset associated with the metric. E.g. “my.uc.table@2” “nyc-taxi-dataset”, + “fantastic-elk-3” + :param model_id: str (optional) + ID of the logged model associated with the metric, if applicable + :param run_id: str (optional) + ID of the run under which to log the metric. Must be provided. + :param run_uuid: str (optional) + [Deprecated, use `run_id` instead] ID of the run under which to log the metric. This field will be + removed in a future MLflow version. + :param step: int (optional) + Step at which to log the metric + requestBody: + content: + application/json: + schema: + type: object + properties: + key: + type: string + description: Name of the metric. + value: + type: number + description: Double value of the metric being logged. + timestamp: + type: integer + description: >- + Unix timestamp in milliseconds at the time metric was + logged. + dataset_digest: + type: string + description: >- + Dataset digest of the dataset associated with the metric, + e.g. an md5 hash of the dataset that uniquely identifies it + within datasets of the same name. + dataset_name: + type: string + description: >- + The name of the dataset associated with the metric. E.g. + “my.uc.table@2” “nyc-taxi-dataset”, “fantastic-elk-3” + model_id: + type: string + description: >- + ID of the logged model associated with the metric, if + applicable + run_id: + type: string + description: >- + ID of the run under which to log the metric. Must be + provided. + run_uuid: + type: string + description: >- + [Deprecated, use `run_id` instead] ID of the run under which + to log the metric. This field will be removed in a future + MLflow version. + step: + type: string + description: Step at which to log the metric + required: + - key + - value + - timestamp + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/log-model: + post: + operationId: experiments_log_model + summary: >- + **Note:** the [Create a logged + model](/api/workspace/experiments/createloggedmodel) API replaces this + tags: + - ml + - experiments + description: >- + **Note:** the [Create a logged + model](/api/workspace/experiments/createloggedmodel) API replaces this + + endpoint. + + + Log a model to an MLflow Run. + + + :param model_json: str (optional) + MLmodel file in json format. + :param run_id: str (optional) + ID of the run to log under + requestBody: + content: + application/json: + schema: + type: object + properties: + model_json: + type: string + description: MLmodel file in json format. + run_id: + type: string + description: ID of the run to log under + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/outputs: + post: + operationId: experiments_log_outputs + summary: Logs outputs, such as models, from an MLflow Run. + tags: + - ml + - experiments + description: |- + Logs outputs, such as models, from an MLflow Run. + + :param run_id: str + The ID of the Run from which to log outputs. + :param models: List[:class:`ModelOutput`] (optional) + The model outputs from the Run. + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: string + description: The ID of the Run from which to log outputs. + models: + type: string + description: The model outputs from the Run. + required: + - run_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/log-parameter: + post: + operationId: experiments_log_param + summary: >- + Logs a param used for a run. A param is a key-value pair (string key, + string value). Examples include + tags: + - ml + - experiments + description: >- + Logs a param used for a run. A param is a key-value pair (string key, + string value). Examples include + + hyperparameters used for ML model training and constant dates and values + used in an ETL pipeline. A + + param can be logged only once for a run. + + + :param key: str + Name of the param. Maximum size is 255 bytes. + :param value: str + String value of the param being logged. Maximum size is 500 bytes. + :param run_id: str (optional) + ID of the run under which to log the param. Must be provided. + :param run_uuid: str (optional) + [Deprecated, use `run_id` instead] ID of the run under which to log the param. This field will be + removed in a future MLflow version. + requestBody: + content: + application/json: + schema: + type: object + properties: + key: + type: string + description: Name of the param. Maximum size is 255 bytes. + value: + type: string + description: >- + String value of the param being logged. Maximum size is 500 + bytes. + run_id: + type: string + description: >- + ID of the run under which to log the param. Must be + provided. + run_uuid: + type: string + description: >- + [Deprecated, use `run_id` instead] ID of the run under which + to log the param. This field will be removed in a future + MLflow version. + required: + - key + - value + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/restore: + post: + operationId: experiments_restore_experiment + summary: >- + Restore an experiment marked for deletion. This also restores associated + metadata, runs, metrics, + tags: + - ml + - experiments + description: >- + Restore an experiment marked for deletion. This also restores associated + metadata, runs, metrics, + + params, and tags. If experiment uses FileStore, underlying artifacts + associated with experiment are + + also restored. + + + Throws `RESOURCE_DOES_NOT_EXIST` if experiment was never created or was + permanently deleted. + + + :param experiment_id: str + ID of the associated experiment. + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: ID of the associated experiment. + required: + - experiment_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/restore: + post: + operationId: experiments_restore_run + summary: >- + Restores a deleted run. This also restores associated metadata, runs, + metrics, params, and tags. + tags: + - ml + - experiments + description: >- + Restores a deleted run. This also restores associated metadata, runs, + metrics, params, and tags. + + + Throws `RESOURCE_DOES_NOT_EXIST` if the run was never created or was + permanently deleted. + + + :param run_id: str + ID of the run to restore. + requestBody: + content: + application/json: + schema: + type: object + properties: + run_id: + type: string + description: ID of the run to restore. + required: + - run_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/databricks/runs/restore-runs: + post: + operationId: experiments_restore_runs + summary: >- + Bulk restore runs in an experiment that were deleted no earlier than the + specified timestamp. Restores + tags: + - ml + - experiments + description: >- + Bulk restore runs in an experiment that were deleted no earlier than the + specified timestamp. Restores + + at most max_runs per request. To call this API from a Databricks + Notebook in Python, you can use the + + client code snippet on + + + :param experiment_id: str + The ID of the experiment containing the runs to restore. + :param min_timestamp_millis: int + The minimum deletion timestamp in milliseconds since the UNIX epoch for restoring runs. Only runs + deleted no earlier than this timestamp are restored. + :param max_runs: int (optional) + An optional positive integer indicating the maximum number of runs to restore. The maximum allowed + value for max_runs is 10000. + + :returns: :class:`RestoreRunsResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: The ID of the experiment containing the runs to restore. + min_timestamp_millis: + type: integer + description: >- + The minimum deletion timestamp in milliseconds since the + UNIX epoch for restoring runs. Only runs deleted no earlier + than this timestamp are restored. + max_runs: + type: string + description: >- + An optional positive integer indicating the maximum number + of runs to restore. The maximum allowed value for max_runs + is 10000. + required: + - experiment_id + - min_timestamp_millis + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RestoreRunsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/search: + post: + operationId: experiments_search_experiments + summary: Searches for experiments that satisfy specified search criteria. + tags: + - ml + - experiments + description: |- + Searches for experiments that satisfy specified search criteria. + + :param filter: str (optional) + String representing a SQL filter condition (e.g. "name ILIKE 'my-experiment%'") + :param max_results: int (optional) + Maximum number of experiments desired. Max threshold is 3000. + :param order_by: List[str] (optional) + List of columns for ordering search results, which can include experiment name and last updated + timestamp with an optional "DESC" or "ASC" annotation, where "ASC" is the default. Tiebreaks are + done by experiment id DESC. + :param page_token: str (optional) + Token indicating the page of experiments to fetch + :param view_type: :class:`ViewType` (optional) + Qualifier for type of experiments to be returned. If unspecified, return only active experiments. + + :returns: Iterator over :class:`Experiment` + requestBody: + content: + application/json: + schema: + type: object + properties: + filter: + type: string + description: >- + String representing a SQL filter condition (e.g. "name ILIKE + 'my-experiment%'") + max_results: + type: string + description: >- + Maximum number of experiments desired. Max threshold is + 3000. + order_by: + type: string + description: >- + List of columns for ordering search results, which can + include experiment name and last updated timestamp with an + optional "DESC" or "ASC" annotation, where "ASC" is the + default. Tiebreaks are done by experiment id DESC. + page_token: + type: string + description: Token indicating the page of experiments to fetch + view_type: + type: string + description: >- + Qualifier for type of experiments to be returned. If + unspecified, return only active experiments. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListExperimentsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/logged-models/search: + post: + operationId: experiments_search_logged_models + summary: Search for Logged Models that satisfy specified search criteria. + tags: + - ml + - experiments + description: |- + Search for Logged Models that satisfy specified search criteria. + + :param datasets: List[:class:`SearchLoggedModelsDataset`] (optional) + List of datasets on which to apply the metrics filter clauses. For example, a filter with + `metrics.accuracy > 0.9` and dataset info with name "test_dataset" means we will return all logged + models with accuracy > 0.9 on the test_dataset. Metric values from ANY dataset matching the criteria + are considered. If no datasets are specified, then metrics across all datasets are considered in the + filter. + :param experiment_ids: List[str] (optional) + The IDs of the experiments in which to search for logged models. + :param filter: str (optional) + A filter expression over logged model info and data that allows returning a subset of logged models. + The syntax is a subset of SQL that supports AND'ing together binary operations. + + Example: ``params.alpha < 0.3 AND metrics.accuracy > 0.9``. + :param max_results: int (optional) + The maximum number of Logged Models to return. The maximum limit is 50. + :param order_by: List[:class:`SearchLoggedModelsOrderBy`] (optional) + The list of columns for ordering the results, with additional fields for sorting criteria. + :param page_token: str (optional) + The token indicating the page of logged models to fetch. + + :returns: :class:`SearchLoggedModelsResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + datasets: + type: string + description: >- + List of datasets on which to apply the metrics filter + clauses. For example, a filter with `metrics.accuracy > 0.9` + and dataset info with name "test_dataset" means we will + return all logged models with accuracy > 0.9 on the + test_dataset. Metric values from ANY dataset matching the + criteria are considered. If no datasets are specified, then + metrics across all datasets are considered in the filter. + experiment_ids: + type: string + description: >- + The IDs of the experiments in which to search for logged + models. + filter: + type: string + description: >- + A filter expression over logged model info and data that + allows returning a subset of logged models. The syntax is a + subset of SQL that supports AND'ing together binary + operations. Example: ``params.alpha < 0.3 AND + metrics.accuracy > 0.9``. + max_results: + type: string + description: >- + The maximum number of Logged Models to return. The maximum + limit is 50. + order_by: + type: string + description: >- + The list of columns for ordering the results, with + additional fields for sorting criteria. + page_token: + type: string + description: The token indicating the page of logged models to fetch. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SearchLoggedModelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/search: + post: + operationId: experiments_search_runs + summary: Searches for runs that satisfy expressions. + tags: + - ml + - experiments + description: |- + Searches for runs that satisfy expressions. + + Search expressions can use `mlflowMetric` and `mlflowParam` keys. + + :param experiment_ids: List[str] (optional) + List of experiment IDs to search over. + :param filter: str (optional) + A filter expression over params, metrics, and tags, that allows returning a subset of runs. The + syntax is a subset of SQL that supports ANDing together binary operations between a param, metric, + or tag and a constant. + + Example: `metrics.rmse < 1 and params.model_class = 'LogisticRegression'` + + You can select columns with special characters (hyphen, space, period, etc.) by using double quotes: + `metrics."model class" = 'LinearRegression' and tags."user-name" = 'Tomas'` + + Supported operators are `=`, `!=`, `>`, `>=`, `<`, and `<=`. + :param max_results: int (optional) + Maximum number of runs desired. Max threshold is 50000 + :param order_by: List[str] (optional) + List of columns to be ordered by, including attributes, params, metrics, and tags with an optional + `"DESC"` or `"ASC"` annotation, where `"ASC"` is the default. Example: `["params.input DESC", + "metrics.alpha ASC", "metrics.rmse"]`. Tiebreaks are done by start_time `DESC` followed by `run_id` + for runs with the same start time (and this is the default ordering criterion if order_by is not + provided). + :param page_token: str (optional) + Token for the current page of runs. + :param run_view_type: :class:`ViewType` (optional) + Whether to display only active, only deleted, or all runs. Defaults to only active runs. + + :returns: Iterator over :class:`Run` + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_ids: + type: string + description: List of experiment IDs to search over. + filter: + type: string + description: >- + A filter expression over params, metrics, and tags, that + allows returning a subset of runs. The syntax is a subset of + SQL that supports ANDing together binary operations between + a param, metric, or tag and a constant. Example: + `metrics.rmse < 1 and params.model_class = + 'LogisticRegression'` You can select columns with special + characters (hyphen, space, period, etc.) by using double + quotes: `metrics."model class" = 'LinearRegression' and + tags."user-name" = 'Tomas'` Supported operators are `=`, + `!=`, `>`, `>=`, `<`, and `<=`. + max_results: + type: string + description: Maximum number of runs desired. Max threshold is 50000 + order_by: + type: string + description: >- + List of columns to be ordered by, including attributes, + params, metrics, and tags with an optional `"DESC"` or + `"ASC"` annotation, where `"ASC"` is the default. Example: + `["params.input DESC", "metrics.alpha ASC", + "metrics.rmse"]`. Tiebreaks are done by start_time `DESC` + followed by `run_id` for runs with the same start time (and + this is the default ordering criterion if order_by is not + provided). + page_token: + type: string + description: Token for the current page of runs. + run_view_type: + type: string + description: >- + Whether to display only active, only deleted, or all runs. + Defaults to only active runs. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SearchRunsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/set-experiment-tag: + post: + operationId: experiments_set_experiment_tag + summary: >- + Sets a tag on an experiment. Experiment tags are metadata that can be + updated. + tags: + - ml + - experiments + description: >- + Sets a tag on an experiment. Experiment tags are metadata that can be + updated. + + + :param experiment_id: str + ID of the experiment under which to log the tag. Must be provided. + :param key: str + Name of the tag. Keys up to 250 bytes in size are supported. + :param value: str + String value of the tag being logged. Values up to 64KB in size are supported. + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: >- + ID of the experiment under which to log the tag. Must be + provided. + key: + type: string + description: Name of the tag. Keys up to 250 bytes in size are supported. + value: + type: string + description: >- + String value of the tag being logged. Values up to 64KB in + size are supported. + required: + - experiment_id + - key + - value + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/logged-models/{model_id}/tags: + patch: + operationId: experiments_set_logged_model_tags + summary: Set tags for a logged model. + tags: + - ml + - experiments + description: |- + Set tags for a logged model. + + :param model_id: str + The ID of the logged model to set the tags on. + :param tags: List[:class:`LoggedModelTag`] (optional) + The tags to set on the logged model. + parameters: + - name: model_id + in: path + required: true + schema: + type: string + description: The ID of the logged model to set the tags on. + requestBody: + content: + application/json: + schema: + type: object + properties: + tags: + type: string + description: The tags to set on the logged model. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/set-tag: + post: + operationId: experiments_set_tag + summary: >- + Sets a tag on a run. Tags are run metadata that can be updated during a + run and after a run completes. + tags: + - ml + - experiments + description: >- + Sets a tag on a run. Tags are run metadata that can be updated during a + run and after a run completes. + + + :param key: str + Name of the tag. Keys up to 250 bytes in size are supported. + :param value: str + String value of the tag being logged. Values up to 64KB in size are supported. + :param run_id: str (optional) + ID of the run under which to log the tag. Must be provided. + :param run_uuid: str (optional) + [Deprecated, use `run_id` instead] ID of the run under which to log the tag. This field will be + removed in a future MLflow version. + requestBody: + content: + application/json: + schema: + type: object + properties: + key: + type: string + description: Name of the tag. Keys up to 250 bytes in size are supported. + value: + type: string + description: >- + String value of the tag being logged. Values up to 64KB in + size are supported. + run_id: + type: string + description: ID of the run under which to log the tag. Must be provided. + run_uuid: + type: string + description: >- + [Deprecated, use `run_id` instead] ID of the run under which + to log the tag. This field will be removed in a future + MLflow version. + required: + - key + - value + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/experiments/update: + post: + operationId: experiments_update_experiment + summary: Updates experiment metadata. + tags: + - ml + - experiments + description: |- + Updates experiment metadata. + + :param experiment_id: str + ID of the associated experiment. + :param new_name: str (optional) + If provided, the experiment's name is changed to the new name. The new name must be unique. + requestBody: + content: + application/json: + schema: + type: object + properties: + experiment_id: + type: string + description: ID of the associated experiment. + new_name: + type: string + description: >- + If provided, the experiment's name is changed to the new + name. The new name must be unique. + required: + - experiment_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/mlflow/runs/update: + post: + operationId: experiments_update_run + summary: Updates run metadata. + tags: + - ml + - experiments + description: |- + Updates run metadata. + + :param end_time: int (optional) + Unix timestamp in milliseconds of when the run ended. + :param run_id: str (optional) + ID of the run to update. Must be provided. + :param run_name: str (optional) + Updated name of the run. + :param run_uuid: str (optional) + [Deprecated, use `run_id` instead] ID of the run to update. This field will be removed in a future + MLflow version. + :param status: :class:`UpdateRunStatus` (optional) + Updated status of the run. + + :returns: :class:`UpdateRunResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + end_time: + type: string + description: Unix timestamp in milliseconds of when the run ended. + run_id: + type: string + description: ID of the run to update. Must be provided. + run_name: + type: string + description: Updated name of the run. + run_uuid: + type: string + description: >- + [Deprecated, use `run_id` instead] ID of the run to update. + This field will be removed in a future MLflow version. + status: + type: string + description: Updated status of the run. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateRunResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ExperimentsAPI + /api/2.0/feature-engineering/materialized-features:batchCreate: + post: + operationId: feature_engineering_batch_create_materialized_features + summary: Batch create materialized features. + tags: + - ml + - feature_engineering + description: |- + Batch create materialized features. + + :param requests: List[:class:`CreateMaterializedFeatureRequest`] + The requests to create materialized features. + + :returns: :class:`BatchCreateMaterializedFeaturesResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + requests: + type: string + description: The requests to create materialized features. + required: + - requests + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchCreateMaterializedFeaturesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + /api/2.0/feature-engineering/features: + post: + operationId: feature_engineering_create_feature + summary: Create a Feature. + tags: + - ml + - feature_engineering + description: |- + Create a Feature. + + :param feature: :class:`Feature` + Feature to create. + + :returns: :class:`Feature` + requestBody: + content: + application/json: + schema: + type: object + properties: + feature: + type: string + description: Feature to create. + required: + - feature + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Feature' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + get: + operationId: feature_engineering_list_features + summary: List Features. + tags: + - ml + - feature_engineering + description: |- + List Features. + + :param page_size: int (optional) + The maximum number of results to return. + :param page_token: str (optional) + Pagination token to go to the next page based on a previous query. + + :returns: Iterator over :class:`Feature` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: The maximum number of results to return. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to the next page based on a previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFeaturesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + /api/2.0/feature-engineering/features/kafka-configs: + post: + operationId: feature_engineering_create_kafka_config + summary: >- + Create a Kafka config. During PrPr, Kafka configs can be read and used + when creating features under + tags: + - ml + - feature_engineering + description: >- + Create a Kafka config. During PrPr, Kafka configs can be read and used + when creating features under + + the entire metastore. Only the creator of the Kafka config can delete + it. + + + :param kafka_config: :class:`KafkaConfig` + + + :returns: :class:`KafkaConfig` + requestBody: + content: + application/json: + schema: + type: object + properties: + kafka_config: + type: string + description: ':returns: :class:`KafkaConfig`' + required: + - kafka_config + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/KafkaConfig' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + get: + operationId: feature_engineering_list_kafka_configs + summary: >- + List Kafka configs. During PrPr, Kafka configs can be read and used when + creating features under the + tags: + - ml + - feature_engineering + description: >- + List Kafka configs. During PrPr, Kafka configs can be read and used when + creating features under the + + entire metastore. Only the creator of the Kafka config can delete it. + + + :param page_size: int (optional) + The maximum number of results to return. + :param page_token: str (optional) + Pagination token to go to the next page based on a previous query. + + :returns: Iterator over :class:`KafkaConfig` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: The maximum number of results to return. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to the next page based on a previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListKafkaConfigsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + /api/2.0/feature-engineering/materialized-features: + post: + operationId: feature_engineering_create_materialized_feature + summary: Create a materialized feature. + tags: + - ml + - feature_engineering + description: |- + Create a materialized feature. + + :param materialized_feature: :class:`MaterializedFeature` + The materialized feature to create. + + :returns: :class:`MaterializedFeature` + requestBody: + content: + application/json: + schema: + type: object + properties: + materialized_feature: + type: string + description: The materialized feature to create. + required: + - materialized_feature + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MaterializedFeature' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + get: + operationId: feature_engineering_list_materialized_features + summary: List materialized features. + tags: + - ml + - feature_engineering + description: |- + List materialized features. + + :param feature_name: str (optional) + Filter by feature name. If specified, only materialized features materialized from this feature will + be returned. + :param page_size: int (optional) + The maximum number of results to return. Defaults to 100 if not specified. Cannot be greater than + 1000. + :param page_token: str (optional) + Pagination token to go to the next page based on a previous query. + + :returns: Iterator over :class:`MaterializedFeature` + parameters: + - name: feature_name + in: query + required: false + schema: + type: string + description: >- + Filter by feature name. If specified, only materialized features + materialized from this feature will be returned. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of results to return. Defaults to 100 if not + specified. Cannot be greater than 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to the next page based on a previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BatchCreateMaterializedFeaturesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + /api/2.0/feature-engineering/features/{full_name}: + delete: + operationId: feature_engineering_delete_feature + summary: Delete a Feature. + tags: + - ml + - feature_engineering + description: |- + Delete a Feature. + + :param full_name: str + Name of the feature to delete. + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Name of the feature to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + get: + operationId: feature_engineering_get_feature + summary: Get a Feature. + tags: + - ml + - feature_engineering + description: |- + Get a Feature. + + :param full_name: str + Name of the feature to get. + + :returns: :class:`Feature` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: Name of the feature to get. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Feature' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + patch: + operationId: feature_engineering_update_feature + summary: Update a Feature. + tags: + - ml + - feature_engineering + description: |- + Update a Feature. + + :param full_name: str + The full three-part name (catalog, schema, name) of the feature. + :param feature: :class:`Feature` + Feature to update. + :param update_mask: str + The list of fields to update. + + :returns: :class:`Feature` + parameters: + - name: full_name + in: path + required: true + schema: + type: string + description: The full three-part name (catalog, schema, name) of the feature. + - name: update_mask + in: query + required: true + schema: + type: string + description: The list of fields to update. + requestBody: + content: + application/json: + schema: + type: object + properties: + feature: + type: string + description: Feature to update. + required: + - feature + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Feature' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + /api/2.0/feature-engineering/features/kafka-configs/{name}: + delete: + operationId: feature_engineering_delete_kafka_config + summary: >- + Delete a Kafka config. During PrPr, Kafka configs can be read and used + when creating features under + tags: + - ml + - feature_engineering + description: >- + Delete a Kafka config. During PrPr, Kafka configs can be read and used + when creating features under + + the entire metastore. Only the creator of the Kafka config can delete + it. + + + :param name: str + Name of the Kafka config to delete. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the Kafka config to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + get: + operationId: feature_engineering_get_kafka_config + summary: >- + Get a Kafka config. During PrPr, Kafka configs can be read and used when + creating features under the + tags: + - ml + - feature_engineering + description: >- + Get a Kafka config. During PrPr, Kafka configs can be read and used when + creating features under the + + entire metastore. Only the creator of the Kafka config can delete it. + + + :param name: str + Name of the Kafka config to get. + + :returns: :class:`KafkaConfig` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the Kafka config to get. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/KafkaConfig' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + patch: + operationId: feature_engineering_update_kafka_config + summary: >- + Update a Kafka config. During PrPr, Kafka configs can be read and used + when creating features under + tags: + - ml + - feature_engineering + description: >- + Update a Kafka config. During PrPr, Kafka configs can be read and used + when creating features under + + the entire metastore. Only the creator of the Kafka config can delete + it. + + + :param name: str + Name that uniquely identifies this Kafka config within the metastore. This will be the identifier + used from the Feature object to reference these configs for a feature. Can be distinct from topic + name. + :param kafka_config: :class:`KafkaConfig` + The Kafka config to update. + :param update_mask: FieldMask + The list of fields to update. + + :returns: :class:`KafkaConfig` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + Name that uniquely identifies this Kafka config within the + metastore. This will be the identifier used from the Feature object + to reference these configs for a feature. Can be distinct from topic + name. + - name: update_mask + in: query + required: true + schema: + type: string + description: The list of fields to update. + requestBody: + content: + application/json: + schema: + type: object + properties: + kafka_config: + type: string + description: The Kafka config to update. + required: + - kafka_config + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/KafkaConfig' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + /api/2.0/feature-engineering/materialized-features/{materialized_feature_id}: + delete: + operationId: feature_engineering_delete_materialized_feature + summary: Delete a materialized feature. + tags: + - ml + - feature_engineering + description: |- + Delete a materialized feature. + + :param materialized_feature_id: str + The ID of the materialized feature to delete. + parameters: + - name: materialized_feature_id + in: path + required: true + schema: + type: string + description: The ID of the materialized feature to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + get: + operationId: feature_engineering_get_materialized_feature + summary: Get a materialized feature. + tags: + - ml + - feature_engineering + description: |- + Get a materialized feature. + + :param materialized_feature_id: str + The ID of the materialized feature. + + :returns: :class:`MaterializedFeature` + parameters: + - name: materialized_feature_id + in: path + required: true + schema: + type: string + description: The ID of the materialized feature. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MaterializedFeature' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + patch: + operationId: feature_engineering_update_materialized_feature + summary: Update a materialized feature (pause/resume). + tags: + - ml + - feature_engineering + description: |- + Update a materialized feature (pause/resume). + + :param materialized_feature_id: str + Unique identifier for the materialized feature. + :param materialized_feature: :class:`MaterializedFeature` + The materialized feature to update. + :param update_mask: str + Provide the materialization feature fields which should be updated. Currently, only the + pipeline_state field can be updated. + + :returns: :class:`MaterializedFeature` + parameters: + - name: materialized_feature_id + in: path + required: true + schema: + type: string + description: Unique identifier for the materialized feature. + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + Provide the materialization feature fields which should be updated. + Currently, only the pipeline_state field can be updated. + requestBody: + content: + application/json: + schema: + type: object + properties: + materialized_feature: + type: string + description: The materialized feature to update. + required: + - materialized_feature + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/MaterializedFeature' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureEngineeringAPI + /api/2.0/feature-store/online-stores: + post: + operationId: feature_store_create_online_store + summary: Create an Online Feature Store. + tags: + - ml + - feature_store + description: |- + Create an Online Feature Store. + + :param online_store: :class:`OnlineStore` + Online store to create. + + :returns: :class:`OnlineStore` + requestBody: + content: + application/json: + schema: + type: object + properties: + online_store: + type: string + description: Online store to create. + required: + - online_store + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/OnlineStore' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureStoreAPI + get: + operationId: feature_store_list_online_stores + summary: List Online Feature Stores. + tags: + - ml + - feature_store + description: |- + List Online Feature Stores. + + :param page_size: int (optional) + The maximum number of results to return. Defaults to 100 if not specified. + :param page_token: str (optional) + Pagination token to go to the next page based on a previous query. + + :returns: Iterator over :class:`OnlineStore` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of results to return. Defaults to 100 if not + specified. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to the next page based on a previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListOnlineStoresResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureStoreAPI + /api/2.0/feature-store/online-stores/{name}: + delete: + operationId: feature_store_delete_online_store + summary: Delete an Online Feature Store. + tags: + - ml + - feature_store + description: |- + Delete an Online Feature Store. + + :param name: str + Name of the online store to delete. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the online store to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureStoreAPI + get: + operationId: feature_store_get_online_store + summary: Get an Online Feature Store. + tags: + - ml + - feature_store + description: |- + Get an Online Feature Store. + + :param name: str + Name of the online store to get. + + :returns: :class:`OnlineStore` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the online store to get. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/OnlineStore' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureStoreAPI + patch: + operationId: feature_store_update_online_store + summary: Update an Online Feature Store. + tags: + - ml + - feature_store + description: |- + Update an Online Feature Store. + + :param name: str + The name of the online store. This is the unique identifier for the online store. + :param online_store: :class:`OnlineStore` + Online store to update. + :param update_mask: str + The list of fields to update. + + :returns: :class:`OnlineStore` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The name of the online store. This is the unique identifier for the + online store. + - name: update_mask + in: query + required: true + schema: + type: string + description: The list of fields to update. + requestBody: + content: + application/json: + schema: + type: object + properties: + online_store: + type: string + description: Online store to update. + required: + - online_store + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/OnlineStore' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureStoreAPI + /api/2.0/feature-store/online-tables/{online_table_name}: + delete: + operationId: feature_store_delete_online_table + summary: Delete online table. + tags: + - ml + - feature_store + description: |- + Delete online table. + + :param online_table_name: str + The full three-part (catalog, schema, table) name of the online table. + parameters: + - name: online_table_name + in: path + required: true + schema: + type: string + description: >- + The full three-part (catalog, schema, table) name of the online + table. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureStoreAPI + /api/2.0/feature-store/tables/{source_table_name}/publish: + post: + operationId: feature_store_publish_table + summary: Publish features. + tags: + - ml + - feature_store + description: |- + Publish features. + + :param source_table_name: str + The full three-part (catalog, schema, table) name of the source table. + :param publish_spec: :class:`PublishSpec` + The specification for publishing the online table from the source table. + + :returns: :class:`PublishTableResponse` + parameters: + - name: source_table_name + in: path + required: true + schema: + type: string + description: >- + The full three-part (catalog, schema, table) name of the source + table. + requestBody: + content: + application/json: + schema: + type: object + properties: + publish_spec: + type: string + description: >- + The specification for publishing the online table from the + source table. + required: + - publish_spec + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PublishTableResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: FeatureStoreAPI + /api/2.0/automl/create-forecasting-experiment: + post: + operationId: forecasting_create_experiment + summary: Creates a serverless forecasting experiment. Returns the experiment ID. + tags: + - ml + - forecasting + description: |- + Creates a serverless forecasting experiment. Returns the experiment ID. + + :param train_data_path: str + The fully qualified path of a Unity Catalog table, formatted as catalog_name.schema_name.table_name, + used as training data for the forecasting model. + :param target_column: str + The column in the input training table used as the prediction target for model training. The values + in this column are used as the ground truth for model training. + :param time_column: str + The column in the input training table that represents each row's timestamp. + :param forecast_granularity: str + The time interval between consecutive rows in the time series data. Possible values include: '1 + second', '1 minute', '5 minutes', '10 minutes', '15 minutes', '30 minutes', 'Hourly', 'Daily', + 'Weekly', 'Monthly', 'Quarterly', 'Yearly'. + :param forecast_horizon: int + The number of time steps into the future to make predictions, calculated as a multiple of + forecast_granularity. This value represents how far ahead the model should forecast. + :param custom_weights_column: str (optional) + The column in the training table used to customize weights for each time series. + :param experiment_path: str (optional) + The path in the workspace to store the created experiment. + :param future_feature_data_path: str (optional) + The fully qualified path of a Unity Catalog table, formatted as catalog_name.schema_name.table_name, + used to store future feature data for predictions. + :param holiday_regions: List[str] (optional) + The region code(s) to automatically add holiday features. Currently supports only one region. + :param include_features: List[str] (optional) + Specifies the list of feature columns to include in model training. These columns must exist in the + training data and be of type string, numerical, or boolean. If not specified, no additional features + will be included. Note: Certain columns are automatically handled: - Automatically excluded: + split_column, target_column, custom_weights_column. - Automatically included: time_column. + :param max_runtime: int (optional) + The maximum duration for the experiment in minutes. The experiment stops automatically if it exceeds + this limit. + :param prediction_data_path: str (optional) + The fully qualified path of a Unity Catalog table, formatted as catalog_name.schema_name.table_name, + used to store predictions. + :param primary_metric: str (optional) + The evaluation metric used to optimize the forecasting model. + :param register_to: str (optional) + The fully qualified path of a Unity Catalog model, formatted as catalog_name.schema_name.model_name, + used to store the best model. + :param split_column: str (optional) + // The column in the training table used for custom data splits. Values must be 'train', 'validate', + or 'test'. + :param timeseries_identifier_columns: List[str] (optional) + The column in the training table used to group the dataset for predicting individual time series. + :param training_frameworks: List[str] (optional) + List of frameworks to include for model tuning. Possible values are 'Prophet', 'ARIMA', 'DeepAR'. An + empty list includes all supported frameworks. + + :returns: + Long-running operation waiter for :class:`ForecastingExperiment`. + See :method:wait_get_experiment_forecasting_succeeded for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + train_data_path: + type: string + description: >- + The fully qualified path of a Unity Catalog table, formatted + as catalog_name.schema_name.table_name, used as training + data for the forecasting model. + target_column: + type: string + description: >- + The column in the input training table used as the + prediction target for model training. The values in this + column are used as the ground truth for model training. + time_column: + type: string + description: >- + The column in the input training table that represents each + row's timestamp. + forecast_granularity: + type: string + description: >- + The time interval between consecutive rows in the time + series data. Possible values include: '1 second', '1 + minute', '5 minutes', '10 minutes', '15 minutes', '30 + minutes', 'Hourly', 'Daily', 'Weekly', 'Monthly', + 'Quarterly', 'Yearly'. + forecast_horizon: + type: integer + description: >- + The number of time steps into the future to make + predictions, calculated as a multiple of + forecast_granularity. This value represents how far ahead + the model should forecast. + custom_weights_column: + type: string + description: >- + The column in the training table used to customize weights + for each time series. + experiment_path: + type: string + description: The path in the workspace to store the created experiment. + future_feature_data_path: + type: string + description: >- + The fully qualified path of a Unity Catalog table, formatted + as catalog_name.schema_name.table_name, used to store future + feature data for predictions. + holiday_regions: + type: string + description: >- + The region code(s) to automatically add holiday features. + Currently supports only one region. + include_features: + type: string + description: >- + Specifies the list of feature columns to include in model + training. These columns must exist in the training data and + be of type string, numerical, or boolean. If not specified, + no additional features will be included. Note: Certain + columns are automatically handled: - Automatically excluded: + split_column, target_column, custom_weights_column. - + Automatically included: time_column. + max_runtime: + type: string + description: >- + The maximum duration for the experiment in minutes. The + experiment stops automatically if it exceeds this limit. + prediction_data_path: + type: string + description: >- + The fully qualified path of a Unity Catalog table, formatted + as catalog_name.schema_name.table_name, used to store + predictions. + primary_metric: + type: string + description: >- + The evaluation metric used to optimize the forecasting + model. + register_to: + type: string + description: >- + The fully qualified path of a Unity Catalog model, formatted + as catalog_name.schema_name.model_name, used to store the + best model. + split_column: + type: string + description: >- + // The column in the training table used for custom data + splits. Values must be 'train', 'validate', or 'test'. + timeseries_identifier_columns: + type: string + description: >- + The column in the training table used to group the dataset + for predicting individual time series. + training_frameworks: + type: string + description: >- + List of frameworks to include for model tuning. Possible + values are 'Prophet', 'ARIMA', 'DeepAR'. An empty list + includes all supported frameworks. + required: + - train_data_path + - target_column + - time_column + - forecast_granularity + - forecast_horizon + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ForecastingExperiment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ForecastingAPI + /api/2.0/automl/get-forecasting-experiment/{experiment_id}: + get: + operationId: forecasting_get_experiment + summary: Public RPC to get forecasting experiment + tags: + - ml + - forecasting + description: |- + Public RPC to get forecasting experiment + + :param experiment_id: str + The unique ID of a forecasting experiment + + :returns: :class:`ForecastingExperiment` + parameters: + - name: experiment_id + in: path + required: true + schema: + type: string + description: The unique ID of a forecasting experiment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ForecastingExperiment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ForecastingAPI + /api/2.0/feature-store/feature-tables/{table_name}/features/{feature_name}/tags: + post: + operationId: materialized_features_create_feature_tag + summary: Creates a FeatureTag. + tags: + - ml + - materialized_features + description: |- + Creates a FeatureTag. + + :param table_name: str + :param feature_name: str + :param feature_tag: :class:`FeatureTag` + + :returns: :class:`FeatureTag` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: ':param feature_name: str' + - name: feature_name + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + feature_tag: + type: string + description: ':returns: :class:`FeatureTag`' + required: + - feature_tag + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FeatureTag' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MaterializedFeaturesAPI + get: + operationId: materialized_features_list_feature_tags + summary: Lists FeatureTags. + tags: + - ml + - materialized_features + description: |- + Lists FeatureTags. + + :param table_name: str + :param feature_name: str + :param page_size: int (optional) + The maximum number of results to return. + :param page_token: str (optional) + Pagination token to go to the next page based on a previous query. + + :returns: Iterator over :class:`FeatureTag` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: ':param feature_name: str' + - name: feature_name + in: path + required: true + schema: + type: string + - name: page_size + in: query + required: false + schema: + type: string + description: The maximum number of results to return. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to the next page based on a previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFeatureTagsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MaterializedFeaturesAPI + /api/2.0/feature-store/feature-tables/{table_name}/features/{feature_name}/tags/{key}: + delete: + operationId: materialized_features_delete_feature_tag + summary: Deletes a FeatureTag. + tags: + - ml + - materialized_features + description: |- + Deletes a FeatureTag. + + :param table_name: str + The name of the feature table. + :param feature_name: str + The name of the feature within the feature table. + :param key: str + The key of the tag to delete. + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: The name of the feature table. + - name: feature_name + in: path + required: true + schema: + type: string + description: The name of the feature within the feature table. + - name: key + in: path + required: true + schema: + type: string + description: The key of the tag to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MaterializedFeaturesAPI + get: + operationId: materialized_features_get_feature_tag + summary: Gets a FeatureTag. + tags: + - ml + - materialized_features + description: |- + Gets a FeatureTag. + + :param table_name: str + :param feature_name: str + :param key: str + + :returns: :class:`FeatureTag` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: ':param feature_name: str' + - name: feature_name + in: path + required: true + schema: + type: string + - name: key + in: path + required: true + schema: + type: string + description: ':returns: :class:`FeatureTag`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FeatureTag' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MaterializedFeaturesAPI + patch: + operationId: materialized_features_update_feature_tag + summary: Updates a FeatureTag. + tags: + - ml + - materialized_features + description: |- + Updates a FeatureTag. + + :param table_name: str + :param feature_name: str + :param key: str + :param feature_tag: :class:`FeatureTag` + :param update_mask: str (optional) + The list of fields to update. + + :returns: :class:`FeatureTag` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: ':param feature_name: str' + - name: feature_name + in: path + required: true + schema: + type: string + - name: key + in: path + required: true + schema: + type: string + description: ':param feature_tag: :class:`FeatureTag`' + - name: update_mask + in: query + required: false + schema: + type: string + description: The list of fields to update. + requestBody: + content: + application/json: + schema: + type: object + properties: + feature_tag: + type: string + required: + - feature_tag + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FeatureTag' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MaterializedFeaturesAPI + /api/2.0/feature-store/feature-tables/{table_name}/features/{feature_name}/lineage: + get: + operationId: materialized_features_get_feature_lineage + summary: Get Feature Lineage. + tags: + - ml + - materialized_features + description: |- + Get Feature Lineage. + + :param table_name: str + The full name of the feature table in Unity Catalog. + :param feature_name: str + The name of the feature. + + :returns: :class:`FeatureLineage` + parameters: + - name: table_name + in: path + required: true + schema: + type: string + description: The full name of the feature table in Unity Catalog. + - name: feature_name + in: path + required: true + schema: + type: string + description: The name of the feature. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FeatureLineage' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: MaterializedFeaturesAPI + /api/2.0/mlflow/transition-requests/approve: + post: + operationId: model_registry_approve_transition_request + summary: Approves a model version stage transition request. + tags: + - ml + - model_registry + description: |- + Approves a model version stage transition request. + + :param name: str + Name of the model. + :param version: str + Version of the model. + :param stage: str + Target stage of the transition. Valid values are: + + * `None`: The initial stage of a model version. + + * `Staging`: Staging or pre-production stage. + + * `Production`: Production stage. + + * `Archived`: Archived stage. + :param archive_existing_versions: bool + Specifies whether to archive all current model versions in the target stage. + :param comment: str (optional) + User-provided comment on the action. + + :returns: :class:`ApproveTransitionRequestResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the model. + version: + type: string + description: Version of the model. + stage: + type: string + description: >- + Target stage of the transition. Valid values are: * `None`: + The initial stage of a model version. * `Staging`: Staging + or pre-production stage. * `Production`: Production stage. * + `Archived`: Archived stage. + archive_existing_versions: + type: boolean + description: >- + Specifies whether to archive all current model versions in + the target stage. + comment: + type: string + description: User-provided comment on the action. + required: + - name + - version + - stage + - archive_existing_versions + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ApproveTransitionRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/comments/create: + post: + operationId: model_registry_create_comment + summary: >- + Posts a comment on a model version. A comment can be submitted either by + a user or programmatically to + tags: + - ml + - model_registry + description: >- + Posts a comment on a model version. A comment can be submitted either by + a user or programmatically to + + display relevant information about the model. For example, test results + or deployment errors. + + + :param name: str + Name of the model. + :param version: str + Version of the model. + :param comment: str + User-provided comment on the action. + + :returns: :class:`CreateCommentResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the model. + version: + type: string + description: Version of the model. + comment: + type: string + description: User-provided comment on the action. + required: + - name + - version + - comment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateCommentResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/create: + post: + operationId: model_registry_create_model + summary: >- + Creates a new registered model with the name specified in the request + body. Throws + tags: + - ml + - model_registry + description: >- + Creates a new registered model with the name specified in the request + body. Throws + + `RESOURCE_ALREADY_EXISTS` if a registered model with the given name + exists. + + + :param name: str + Register models under this name + :param description: str (optional) + Optional description for registered model. + :param tags: List[:class:`ModelTag`] (optional) + Additional metadata for registered model. + + :returns: :class:`CreateModelResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Register models under this name + description: + type: string + description: Optional description for registered model. + tags: + type: string + description: Additional metadata for registered model. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateModelResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/create: + post: + operationId: model_registry_create_model_version + summary: Creates a model version. + tags: + - ml + - model_registry + description: |- + Creates a model version. + + :param name: str + Register model under this name + :param source: str + URI indicating the location of the model artifacts. + :param description: str (optional) + Optional description for model version. + :param run_id: str (optional) + MLflow run ID for correlation, if `source` was generated by an experiment run in MLflow tracking + server + :param run_link: str (optional) + MLflow run link - this is the exact link of the run that generated this model version, potentially + hosted at another instance of MLflow. + :param tags: List[:class:`ModelVersionTag`] (optional) + Additional metadata for model version. + + :returns: :class:`CreateModelVersionResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Register model under this name + source: + type: string + description: URI indicating the location of the model artifacts. + description: + type: string + description: Optional description for model version. + run_id: + type: string + description: >- + MLflow run ID for correlation, if `source` was generated by + an experiment run in MLflow tracking server + run_link: + type: string + description: >- + MLflow run link - this is the exact link of the run that + generated this model version, potentially hosted at another + instance of MLflow. + tags: + type: string + description: Additional metadata for model version. + required: + - name + - source + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateModelVersionResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/transition-requests/create: + post: + operationId: model_registry_create_transition_request + summary: Creates a model version stage transition request. + tags: + - ml + - model_registry + description: |- + Creates a model version stage transition request. + + :param name: str + Name of the model. + :param version: str + Version of the model. + :param stage: str + Target stage of the transition. Valid values are: + + * `None`: The initial stage of a model version. + + * `Staging`: Staging or pre-production stage. + + * `Production`: Production stage. + + * `Archived`: Archived stage. + :param comment: str (optional) + User-provided comment on the action. + + :returns: :class:`CreateTransitionRequestResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the model. + version: + type: string + description: Version of the model. + stage: + type: string + description: >- + Target stage of the transition. Valid values are: * `None`: + The initial stage of a model version. * `Staging`: Staging + or pre-production stage. * `Production`: Production stage. * + `Archived`: Archived stage. + comment: + type: string + description: User-provided comment on the action. + required: + - name + - version + - stage + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateTransitionRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registry-webhooks/create: + post: + operationId: model_registry_create_webhook + summary: >- + **NOTE:** This endpoint is in Public Preview. Creates a registry + webhook. + tags: + - ml + - model_registry + description: >- + **NOTE:** This endpoint is in Public Preview. Creates a registry + webhook. + + + :param events: List[:class:`RegistryWebhookEvent`] + Events that can trigger a registry webhook: * `MODEL_VERSION_CREATED`: A new model version was + created for the associated model. + + * `MODEL_VERSION_TRANSITIONED_STAGE`: A model version’s stage was changed. + + * `TRANSITION_REQUEST_CREATED`: A user requested a model version’s stage be transitioned. + + * `COMMENT_CREATED`: A user wrote a comment on a registered model. + + * `REGISTERED_MODEL_CREATED`: A new registered model was created. This event type can only be + specified for a registry-wide webhook, which can be created by not specifying a model name in the + create request. + + * `MODEL_VERSION_TAG_SET`: A user set a tag on the model version. + + * `MODEL_VERSION_TRANSITIONED_TO_STAGING`: A model version was transitioned to staging. + + * `MODEL_VERSION_TRANSITIONED_TO_PRODUCTION`: A model version was transitioned to production. + + * `MODEL_VERSION_TRANSITIONED_TO_ARCHIVED`: A model version was archived. + + * `TRANSITION_REQUEST_TO_STAGING_CREATED`: A user requested a model version be transitioned to + staging. + + * `TRANSITION_REQUEST_TO_PRODUCTION_CREATED`: A user requested a model version be transitioned to + production. + + * `TRANSITION_REQUEST_TO_ARCHIVED_CREATED`: A user requested a model version be archived. + :param description: str (optional) + User-specified description for the webhook. + :param http_url_spec: :class:`HttpUrlSpec` (optional) + External HTTPS URL called on event trigger (by using a POST request). + :param job_spec: :class:`JobSpec` (optional) + ID of the job that the webhook runs. + :param model_name: str (optional) + If model name is not specified, a registry-wide webhook is created that listens for the specified + events across all versions of all registered models. + :param status: :class:`RegistryWebhookStatus` (optional) + Enable or disable triggering the webhook, or put the webhook into test mode. The default is + `ACTIVE`: * `ACTIVE`: Webhook is triggered when an associated event happens. + + * `DISABLED`: Webhook is not triggered. + + * `TEST_MODE`: Webhook can be triggered through the test endpoint, but is not triggered on a real + event. + + :returns: :class:`CreateWebhookResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + events: + type: string + description: >- + Events that can trigger a registry webhook: * + `MODEL_VERSION_CREATED`: A new model version was created for + the associated model. * `MODEL_VERSION_TRANSITIONED_STAGE`: + A model version’s stage was changed. * + `TRANSITION_REQUEST_CREATED`: A user requested a model + version’s stage be transitioned. * `COMMENT_CREATED`: A user + wrote a comment on a registered model. * + `REGISTERED_MODEL_CREATED`: A new registered model was + created. This event type can only be specified for a + registry-wide webhook, which can be created by not + specifying a model name in the create request. * + `MODEL_VERSION_TAG_SET`: A user set a tag on the model + version. * `MODEL_VERSION_TRANSITIONED_TO_STAGING`: A model + version was transitioned to staging. * + `MODEL_VERSION_TRANSITIONED_TO_PRODUCTION`: A model version + was transitioned to production. * + `MODEL_VERSION_TRANSITIONED_TO_ARCHIVED`: A model version + was archived. * `TRANSITION_REQUEST_TO_STAGING_CREATED`: A + user requested a model version be transitioned to staging. * + `TRANSITION_REQUEST_TO_PRODUCTION_CREATED`: A user requested + a model version be transitioned to production. * + `TRANSITION_REQUEST_TO_ARCHIVED_CREATED`: A user requested a + model version be archived. + description: + type: string + description: User-specified description for the webhook. + http_url_spec: + type: string + description: >- + External HTTPS URL called on event trigger (by using a POST + request). + job_spec: + type: string + description: ID of the job that the webhook runs. + model_name: + type: string + description: >- + If model name is not specified, a registry-wide webhook is + created that listens for the specified events across all + versions of all registered models. + status: + type: string + description: >- + Enable or disable triggering the webhook, or put the webhook + into test mode. The default is `ACTIVE`: * `ACTIVE`: Webhook + is triggered when an associated event happens. * `DISABLED`: + Webhook is not triggered. * `TEST_MODE`: Webhook can be + triggered through the test endpoint, but is not triggered on + a real event. + required: + - events + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateWebhookResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/comments/delete: + delete: + operationId: model_registry_delete_comment + summary: Deletes a comment on a model version. + tags: + - ml + - model_registry + description: |- + Deletes a comment on a model version. + + :param id: str + Unique identifier of an activity + parameters: + - name: id + in: query + required: true + schema: + type: string + description: Unique identifier of an activity + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/delete: + delete: + operationId: model_registry_delete_model + summary: Deletes a registered model. + tags: + - ml + - model_registry + description: |- + Deletes a registered model. + + :param name: str + Registered model unique name identifier. + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Registered model unique name identifier. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/delete-tag: + delete: + operationId: model_registry_delete_model_tag + summary: Deletes the tag for a registered model. + tags: + - ml + - model_registry + description: |- + Deletes the tag for a registered model. + + :param name: str + Name of the registered model that the tag was logged under. + :param key: str + Name of the tag. The name must be an exact match; wild-card deletion is not supported. Maximum size + is 250 bytes. + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Name of the registered model that the tag was logged under. + - name: key + in: query + required: true + schema: + type: string + description: >- + Name of the tag. The name must be an exact match; wild-card deletion + is not supported. Maximum size is 250 bytes. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/delete: + delete: + operationId: model_registry_delete_model_version + summary: Deletes a model version. + tags: + - ml + - model_registry + description: |- + Deletes a model version. + + :param name: str + Name of the registered model + :param version: str + Model version number + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Name of the registered model + - name: version + in: query + required: true + schema: + type: string + description: Model version number + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/delete-tag: + delete: + operationId: model_registry_delete_model_version_tag + summary: Deletes a model version tag. + tags: + - ml + - model_registry + description: |- + Deletes a model version tag. + + :param name: str + Name of the registered model that the tag was logged under. + :param version: str + Model version number that the tag was logged under. + :param key: str + Name of the tag. The name must be an exact match; wild-card deletion is not supported. Maximum size + is 250 bytes. + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Name of the registered model that the tag was logged under. + - name: version + in: query + required: true + schema: + type: string + description: Model version number that the tag was logged under. + - name: key + in: query + required: true + schema: + type: string + description: >- + Name of the tag. The name must be an exact match; wild-card deletion + is not supported. Maximum size is 250 bytes. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/transition-requests/delete: + delete: + operationId: model_registry_delete_transition_request + summary: Cancels a model version stage transition request. + tags: + - ml + - model_registry + description: |- + Cancels a model version stage transition request. + + :param name: str + Name of the model. + :param version: str + Version of the model. + :param stage: str + Target stage of the transition request. Valid values are: + + * `None`: The initial stage of a model version. + + * `Staging`: Staging or pre-production stage. + + * `Production`: Production stage. + + * `Archived`: Archived stage. + :param creator: str + Username of the user who created this request. Of the transition requests matching the specified + details, only the one transition created by this user will be deleted. + :param comment: str (optional) + User-provided comment on the action. + + :returns: :class:`DeleteTransitionRequestResponse` + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Name of the model. + - name: version + in: query + required: true + schema: + type: string + description: Version of the model. + - name: stage + in: query + required: true + schema: + type: string + description: >- + Target stage of the transition request. Valid values are: * `None`: + The initial stage of a model version. * `Staging`: Staging or + pre-production stage. * `Production`: Production stage. * + `Archived`: Archived stage. + - name: creator + in: query + required: true + schema: + type: string + description: >- + Username of the user who created this request. Of the transition + requests matching the specified details, only the one transition + created by this user will be deleted. + - name: comment + in: query + required: false + schema: + type: string + description: User-provided comment on the action. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteTransitionRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registry-webhooks/delete: + delete: + operationId: model_registry_delete_webhook + summary: >- + **NOTE:** This endpoint is in Public Preview. Deletes a registry + webhook. + tags: + - ml + - model_registry + description: >- + **NOTE:** This endpoint is in Public Preview. Deletes a registry + webhook. + + + :param id: str + Webhook ID required to delete a registry webhook. + parameters: + - name: id + in: query + required: true + schema: + type: string + description: Webhook ID required to delete a registry webhook. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/get-latest-versions: + post: + operationId: model_registry_get_latest_versions + summary: Gets the latest version of a registered model. + tags: + - ml + - model_registry + description: |- + Gets the latest version of a registered model. + + :param name: str + Registered model unique name identifier. + :param stages: List[str] (optional) + List of stages. + + :returns: Iterator over :class:`ModelVersion` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Registered model unique name identifier. + stages: + type: string + description: List of stages. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetLatestVersionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/databricks/registered-models/get: + get: + operationId: model_registry_get_model + summary: >- + Get the details of a model. This is a Databricks workspace version of + the [MLflow endpoint] that also + tags: + - ml + - model_registry + description: >- + Get the details of a model. This is a Databricks workspace version of + the [MLflow endpoint] that also + + returns the model's Databricks workspace ID and the permission level of + the requesting user on the + + model. + + + [MLflow endpoint]: + https://www.mlflow.org/docs/latest/rest-api.html#get-registeredmodel + + + :param name: str + Registered model unique name identifier. + + :returns: :class:`GetModelResponse` + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Registered model unique name identifier. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetModelResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/get: + get: + operationId: model_registry_get_model_version + summary: Get a model version. + tags: + - ml + - model_registry + description: |- + Get a model version. + + :param name: str + Name of the registered model + :param version: str + Model version number + + :returns: :class:`GetModelVersionResponse` + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Name of the registered model + - name: version + in: query + required: true + schema: + type: string + description: Model version number + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetModelVersionResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/get-download-uri: + get: + operationId: model_registry_get_model_version_download_uri + summary: Gets a URI to download the model version. + tags: + - ml + - model_registry + description: |- + Gets a URI to download the model version. + + :param name: str + Name of the registered model + :param version: str + Model version number + + :returns: :class:`GetModelVersionDownloadUriResponse` + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Name of the registered model + - name: version + in: query + required: true + schema: + type: string + description: Model version number + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetModelVersionDownloadUriResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/permissions/registered-models/{registered_model_id}/permissionLevels: + get: + operationId: model_registry_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - ml + - model_registry + description: |- + Gets the permission levels that a user can have on an object. + + :param registered_model_id: str + The registered model for which to get or manage permissions. + + :returns: :class:`GetRegisteredModelPermissionLevelsResponse` + parameters: + - name: registered_model_id + in: path + required: true + schema: + type: string + description: The registered model for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/GetRegisteredModelPermissionLevelsResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/permissions/registered-models/{registered_model_id}: + get: + operationId: model_registry_get_permissions + summary: >- + Gets the permissions of a registered model. Registered models can + inherit permissions from their root + tags: + - ml + - model_registry + description: >- + Gets the permissions of a registered model. Registered models can + inherit permissions from their root + + object. + + + :param registered_model_id: str + The registered model for which to get or manage permissions. + + :returns: :class:`RegisteredModelPermissions` + parameters: + - name: registered_model_id + in: path + required: true + schema: + type: string + description: The registered model for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegisteredModelPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + put: + operationId: model_registry_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - ml + - model_registry + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param registered_model_id: str + The registered model for which to get or manage permissions. + :param access_control_list: + List[:class:`RegisteredModelAccessControlRequest`] (optional) + + + :returns: :class:`RegisteredModelPermissions` + parameters: + - name: registered_model_id + in: path + required: true + schema: + type: string + description: The registered model for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`RegisteredModelPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegisteredModelPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + patch: + operationId: model_registry_update_permissions + summary: >- + Updates the permissions on a registered model. Registered models can + inherit permissions from their + tags: + - ml + - model_registry + description: >- + Updates the permissions on a registered model. Registered models can + inherit permissions from their + + root object. + + + :param registered_model_id: str + The registered model for which to get or manage permissions. + :param access_control_list: + List[:class:`RegisteredModelAccessControlRequest`] (optional) + + + :returns: :class:`RegisteredModelPermissions` + parameters: + - name: registered_model_id + in: path + required: true + schema: + type: string + description: The registered model for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`RegisteredModelPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RegisteredModelPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/list: + get: + operationId: model_registry_list_models + summary: >- + Lists all available registered models, up to the limit specified in + __max_results__. + tags: + - ml + - model_registry + description: >- + Lists all available registered models, up to the limit specified in + __max_results__. + + + :param max_results: int (optional) + Maximum number of registered models desired. Max threshold is 1000. + :param page_token: str (optional) + Pagination token to go to the next page based on a previous query. + + :returns: Iterator over :class:`Model` + parameters: + - name: max_results + in: query + required: false + schema: + type: string + description: Maximum number of registered models desired. Max threshold is 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to the next page based on a previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListModelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/transition-requests/list: + get: + operationId: model_registry_list_transition_requests + summary: Gets a list of all open stage transition requests for the model version. + tags: + - ml + - model_registry + description: |- + Gets a list of all open stage transition requests for the model version. + + :param name: str + Name of the registered model. + :param version: str + Version of the model. + + :returns: Iterator over :class:`Activity` + parameters: + - name: name + in: query + required: true + schema: + type: string + description: Name of the registered model. + - name: version + in: query + required: true + schema: + type: string + description: Version of the model. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListTransitionRequestsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registry-webhooks/list: + get: + operationId: model_registry_list_webhooks + summary: >- + **NOTE:** This endpoint is in Public Preview. Lists all registry + webhooks. + tags: + - ml + - model_registry + description: >- + **NOTE:** This endpoint is in Public Preview. Lists all registry + webhooks. + + + :param events: List[:class:`RegistryWebhookEvent`] (optional) + Events that trigger the webhook. * `MODEL_VERSION_CREATED`: A new model version was created for the + associated model. + + * `MODEL_VERSION_TRANSITIONED_STAGE`: A model version’s stage was changed. + + * `TRANSITION_REQUEST_CREATED`: A user requested a model version’s stage be transitioned. + + * `COMMENT_CREATED`: A user wrote a comment on a registered model. + + * `REGISTERED_MODEL_CREATED`: A new registered model was created. This event type can only be + specified for a registry-wide webhook, which can be created by not specifying a model name in the + create request. + + * `MODEL_VERSION_TAG_SET`: A user set a tag on the model version. + + * `MODEL_VERSION_TRANSITIONED_TO_STAGING`: A model version was transitioned to staging. + + * `MODEL_VERSION_TRANSITIONED_TO_PRODUCTION`: A model version was transitioned to production. + + * `MODEL_VERSION_TRANSITIONED_TO_ARCHIVED`: A model version was archived. + + * `TRANSITION_REQUEST_TO_STAGING_CREATED`: A user requested a model version be transitioned to + staging. + + * `TRANSITION_REQUEST_TO_PRODUCTION_CREATED`: A user requested a model version be transitioned to + production. + + * `TRANSITION_REQUEST_TO_ARCHIVED_CREATED`: A user requested a model version be archived. + + If `events` is specified, any webhook with one or more of the specified trigger events is included + in the output. If `events` is not specified, webhooks of all event types are included in the output. + :param max_results: int (optional) + + :param model_name: str (optional) + Registered model name If not specified, all webhooks associated with the specified events are + listed, regardless of their associated model. + :param page_token: str (optional) + Token indicating the page of artifact results to fetch + + :returns: Iterator over :class:`RegistryWebhook` + parameters: + - name: events + in: query + required: false + schema: + type: string + description: >- + Events that trigger the webhook. * `MODEL_VERSION_CREATED`: A new + model version was created for the associated model. * + `MODEL_VERSION_TRANSITIONED_STAGE`: A model version’s stage was + changed. * `TRANSITION_REQUEST_CREATED`: A user requested a model + version’s stage be transitioned. * `COMMENT_CREATED`: A user wrote a + comment on a registered model. * `REGISTERED_MODEL_CREATED`: A new + registered model was created. This event type can only be specified + for a registry-wide webhook, which can be created by not specifying + a model name in the create request. * `MODEL_VERSION_TAG_SET`: A + user set a tag on the model version. * + `MODEL_VERSION_TRANSITIONED_TO_STAGING`: A model version was + transitioned to staging. * + `MODEL_VERSION_TRANSITIONED_TO_PRODUCTION`: A model version was + transitioned to production. * + `MODEL_VERSION_TRANSITIONED_TO_ARCHIVED`: A model version was + archived. * `TRANSITION_REQUEST_TO_STAGING_CREATED`: A user + requested a model version be transitioned to staging. * + `TRANSITION_REQUEST_TO_PRODUCTION_CREATED`: A user requested a model + version be transitioned to production. * + `TRANSITION_REQUEST_TO_ARCHIVED_CREATED`: A user requested a model + version be archived. If `events` is specified, any webhook with one + or more of the specified trigger events is included in the output. + If `events` is not specified, webhooks of all event types are + included in the output. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + :param model_name: str (optional) Registered model name If not + specified, all webhooks associated with the specified events are + listed, regardless of their associated model. + - name: model_name + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: Token indicating the page of artifact results to fetch + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListRegistryWebhooks' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/transition-requests/reject: + post: + operationId: model_registry_reject_transition_request + summary: Rejects a model version stage transition request. + tags: + - ml + - model_registry + description: |- + Rejects a model version stage transition request. + + :param name: str + Name of the model. + :param version: str + Version of the model. + :param stage: str + Target stage of the transition. Valid values are: + + * `None`: The initial stage of a model version. + + * `Staging`: Staging or pre-production stage. + + * `Production`: Production stage. + + * `Archived`: Archived stage. + :param comment: str (optional) + User-provided comment on the action. + + :returns: :class:`RejectTransitionRequestResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the model. + version: + type: string + description: Version of the model. + stage: + type: string + description: >- + Target stage of the transition. Valid values are: * `None`: + The initial stage of a model version. * `Staging`: Staging + or pre-production stage. * `Production`: Production stage. * + `Archived`: Archived stage. + comment: + type: string + description: User-provided comment on the action. + required: + - name + - version + - stage + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RejectTransitionRequestResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/rename: + post: + operationId: model_registry_rename_model + summary: Renames a registered model. + tags: + - ml + - model_registry + description: |- + Renames a registered model. + + :param name: str + Registered model unique name identifier. + :param new_name: str (optional) + If provided, updates the name for this `registered_model`. + + :returns: :class:`RenameModelResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Registered model unique name identifier. + new_name: + type: string + description: If provided, updates the name for this `registered_model`. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RenameModelResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/search: + get: + operationId: model_registry_search_model_versions + summary: Searches for specific model versions based on the supplied __filter__. + tags: + - ml + - model_registry + description: |- + Searches for specific model versions based on the supplied __filter__. + + :param filter: str (optional) + String filter condition, like "name='my-model-name'". Must be a single boolean condition, with + string values wrapped in single quotes. + :param max_results: int (optional) + Maximum number of models desired. Max threshold is 10K. + :param order_by: List[str] (optional) + List of columns to be ordered by including model name, version, stage with an optional "DESC" or + "ASC" annotation, where "ASC" is the default. Tiebreaks are done by latest stage transition + timestamp, followed by name ASC, followed by version DESC. + :param page_token: str (optional) + Pagination token to go to next page based on previous search query. + + :returns: Iterator over :class:`ModelVersion` + parameters: + - name: filter + in: query + required: false + schema: + type: string + description: >- + String filter condition, like "name='my-model-name'". Must be a + single boolean condition, with string values wrapped in single + quotes. + - name: max_results + in: query + required: false + schema: + type: string + description: Maximum number of models desired. Max threshold is 10K. + - name: order_by + in: query + required: false + schema: + type: string + description: >- + List of columns to be ordered by including model name, version, + stage with an optional "DESC" or "ASC" annotation, where "ASC" is + the default. Tiebreaks are done by latest stage transition + timestamp, followed by name ASC, followed by version DESC. + - name: page_token + in: query + required: false + schema: + type: string + description: Pagination token to go to next page based on previous search query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetLatestVersionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/search: + get: + operationId: model_registry_search_models + summary: Search for registered models based on the specified __filter__. + tags: + - ml + - model_registry + description: |- + Search for registered models based on the specified __filter__. + + :param filter: str (optional) + String filter condition, like "name LIKE 'my-model-name'". Interpreted in the backend automatically + as "name LIKE '%my-model-name%'". Single boolean condition, with string values wrapped in single + quotes. + :param max_results: int (optional) + Maximum number of models desired. Default is 100. Max threshold is 1000. + :param order_by: List[str] (optional) + List of columns for ordering search results, which can include model name and last updated timestamp + with an optional "DESC" or "ASC" annotation, where "ASC" is the default. Tiebreaks are done by model + name ASC. + :param page_token: str (optional) + Pagination token to go to the next page based on a previous search query. + + :returns: Iterator over :class:`Model` + parameters: + - name: filter + in: query + required: true + schema: + type: string + description: >- + String filter condition, like "name LIKE 'my-model-name'". + Interpreted in the backend automatically as "name LIKE + '%my-model-name%'". Single boolean condition, with string values + wrapped in single quotes. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of models desired. Default is 100. Max threshold is + 1000. + - name: order_by + in: query + required: false + schema: + type: string + description: >- + List of columns for ordering search results, which can include model + name and last updated timestamp with an optional "DESC" or "ASC" + annotation, where "ASC" is the default. Tiebreaks are done by model + name ASC. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page based on a previous search + query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListModelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/set-tag: + post: + operationId: model_registry_set_model_tag + summary: Sets a tag on a registered model. + tags: + - ml + - model_registry + description: |- + Sets a tag on a registered model. + + :param name: str + Unique name of the model. + :param key: str + Name of the tag. Maximum size depends on storage backend. If a tag with this name already exists, + its preexisting value will be replaced by the specified `value`. All storage backends are guaranteed + to support key values up to 250 bytes in size. + :param value: str + String value of the tag being logged. Maximum size depends on storage backend. All storage backends + are guaranteed to support key values up to 5000 bytes in size. + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Unique name of the model. + key: + type: string + description: >- + Name of the tag. Maximum size depends on storage backend. If + a tag with this name already exists, its preexisting value + will be replaced by the specified `value`. All storage + backends are guaranteed to support key values up to 250 + bytes in size. + value: + type: string + description: >- + String value of the tag being logged. Maximum size depends + on storage backend. All storage backends are guaranteed to + support key values up to 5000 bytes in size. + required: + - name + - key + - value + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/set-tag: + post: + operationId: model_registry_set_model_version_tag + summary: Sets a model version tag. + tags: + - ml + - model_registry + description: |- + Sets a model version tag. + + :param name: str + Unique name of the model. + :param version: str + Model version number. + :param key: str + Name of the tag. Maximum size depends on storage backend. If a tag with this name already exists, + its preexisting value will be replaced by the specified `value`. All storage backends are guaranteed + to support key values up to 250 bytes in size. + :param value: str + String value of the tag being logged. Maximum size depends on storage backend. All storage backends + are guaranteed to support key values up to 5000 bytes in size. + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Unique name of the model. + version: + type: string + description: Model version number. + key: + type: string + description: >- + Name of the tag. Maximum size depends on storage backend. If + a tag with this name already exists, its preexisting value + will be replaced by the specified `value`. All storage + backends are guaranteed to support key values up to 250 + bytes in size. + value: + type: string + description: >- + String value of the tag being logged. Maximum size depends + on storage backend. All storage backends are guaranteed to + support key values up to 5000 bytes in size. + required: + - name + - version + - key + - value + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registry-webhooks/test: + post: + operationId: model_registry_test_registry_webhook + summary: '**NOTE:** This endpoint is in Public Preview. Tests a registry webhook.' + tags: + - ml + - model_registry + description: |- + **NOTE:** This endpoint is in Public Preview. Tests a registry webhook. + + :param id: str + Webhook ID + :param event: :class:`RegistryWebhookEvent` (optional) + If `event` is specified, the test trigger uses the specified event. If `event` is not specified, the + test trigger uses a randomly chosen event associated with the webhook. + + :returns: :class:`TestRegistryWebhookResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + id: + type: string + description: Webhook ID + event: + type: string + description: >- + If `event` is specified, the test trigger uses the specified + event. If `event` is not specified, the test trigger uses a + randomly chosen event associated with the webhook. + required: + - id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TestRegistryWebhookResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/databricks/model-versions/transition-stage: + post: + operationId: model_registry_transition_stage + summary: >- + Transition a model version's stage. This is a Databricks workspace + version of the [MLflow endpoint] + tags: + - ml + - model_registry + description: >- + Transition a model version's stage. This is a Databricks workspace + version of the [MLflow endpoint] + + that also accepts a comment associated with the transition to be + recorded. + + + [MLflow endpoint]: + https://www.mlflow.org/docs/latest/rest-api.html#transition-modelversion-stage + + + :param name: str + Name of the model. + :param version: str + Version of the model. + :param stage: str + Target stage of the transition. Valid values are: + + * `None`: The initial stage of a model version. + + * `Staging`: Staging or pre-production stage. + + * `Production`: Production stage. + + * `Archived`: Archived stage. + :param archive_existing_versions: bool + Specifies whether to archive all current model versions in the target stage. + :param comment: str (optional) + User-provided comment on the action. + + :returns: :class:`TransitionStageResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the model. + version: + type: string + description: Version of the model. + stage: + type: string + description: >- + Target stage of the transition. Valid values are: * `None`: + The initial stage of a model version. * `Staging`: Staging + or pre-production stage. * `Production`: Production stage. * + `Archived`: Archived stage. + archive_existing_versions: + type: boolean + description: >- + Specifies whether to archive all current model versions in + the target stage. + comment: + type: string + description: User-provided comment on the action. + required: + - name + - version + - stage + - archive_existing_versions + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TransitionStageResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/comments/update: + patch: + operationId: model_registry_update_comment + summary: Post an edit to a comment on a model version. + tags: + - ml + - model_registry + description: |- + Post an edit to a comment on a model version. + + :param id: str + Unique identifier of an activity + :param comment: str + User-provided comment on the action. + + :returns: :class:`UpdateCommentResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + id: + type: string + description: Unique identifier of an activity + comment: + type: string + description: User-provided comment on the action. + required: + - id + - comment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateCommentResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registered-models/update: + patch: + operationId: model_registry_update_model + summary: Updates a registered model. + tags: + - ml + - model_registry + description: |- + Updates a registered model. + + :param name: str + Registered model unique name identifier. + :param description: str (optional) + If provided, updates the description for this `registered_model`. + + :returns: :class:`UpdateModelResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Registered model unique name identifier. + description: + type: string + description: >- + If provided, updates the description for this + `registered_model`. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateModelResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/model-versions/update: + patch: + operationId: model_registry_update_model_version + summary: Updates the model version. + tags: + - ml + - model_registry + description: |- + Updates the model version. + + :param name: str + Name of the registered model + :param version: str + Model version number + :param description: str (optional) + If provided, updates the description for this `registered_model`. + + :returns: :class:`UpdateModelVersionResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the registered model + version: + type: string + description: Model version number + description: + type: string + description: >- + If provided, updates the description for this + `registered_model`. + required: + - name + - version + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateModelVersionResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI + /api/2.0/mlflow/registry-webhooks/update: + patch: + operationId: model_registry_update_webhook + summary: >- + **NOTE:** This endpoint is in Public Preview. Updates a registry + webhook. + tags: + - ml + - model_registry + description: >- + **NOTE:** This endpoint is in Public Preview. Updates a registry + webhook. + + + :param id: str + Webhook ID + :param description: str (optional) + User-specified description for the webhook. + :param events: List[:class:`RegistryWebhookEvent`] (optional) + Events that can trigger a registry webhook: * `MODEL_VERSION_CREATED`: A new model version was + created for the associated model. + + * `MODEL_VERSION_TRANSITIONED_STAGE`: A model version’s stage was changed. + + * `TRANSITION_REQUEST_CREATED`: A user requested a model version’s stage be transitioned. + + * `COMMENT_CREATED`: A user wrote a comment on a registered model. + + * `REGISTERED_MODEL_CREATED`: A new registered model was created. This event type can only be + specified for a registry-wide webhook, which can be created by not specifying a model name in the + create request. + + * `MODEL_VERSION_TAG_SET`: A user set a tag on the model version. + + * `MODEL_VERSION_TRANSITIONED_TO_STAGING`: A model version was transitioned to staging. + + * `MODEL_VERSION_TRANSITIONED_TO_PRODUCTION`: A model version was transitioned to production. + + * `MODEL_VERSION_TRANSITIONED_TO_ARCHIVED`: A model version was archived. + + * `TRANSITION_REQUEST_TO_STAGING_CREATED`: A user requested a model version be transitioned to + staging. + + * `TRANSITION_REQUEST_TO_PRODUCTION_CREATED`: A user requested a model version be transitioned to + production. + + * `TRANSITION_REQUEST_TO_ARCHIVED_CREATED`: A user requested a model version be archived. + :param http_url_spec: :class:`HttpUrlSpec` (optional) + + :param job_spec: :class:`JobSpec` (optional) + + :param status: :class:`RegistryWebhookStatus` (optional) + + + :returns: :class:`UpdateWebhookResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + id: + type: string + description: Webhook ID + description: + type: string + description: User-specified description for the webhook. + events: + type: string + description: >- + Events that can trigger a registry webhook: * + `MODEL_VERSION_CREATED`: A new model version was created for + the associated model. * `MODEL_VERSION_TRANSITIONED_STAGE`: + A model version’s stage was changed. * + `TRANSITION_REQUEST_CREATED`: A user requested a model + version’s stage be transitioned. * `COMMENT_CREATED`: A user + wrote a comment on a registered model. * + `REGISTERED_MODEL_CREATED`: A new registered model was + created. This event type can only be specified for a + registry-wide webhook, which can be created by not + specifying a model name in the create request. * + `MODEL_VERSION_TAG_SET`: A user set a tag on the model + version. * `MODEL_VERSION_TRANSITIONED_TO_STAGING`: A model + version was transitioned to staging. * + `MODEL_VERSION_TRANSITIONED_TO_PRODUCTION`: A model version + was transitioned to production. * + `MODEL_VERSION_TRANSITIONED_TO_ARCHIVED`: A model version + was archived. * `TRANSITION_REQUEST_TO_STAGING_CREATED`: A + user requested a model version be transitioned to staging. * + `TRANSITION_REQUEST_TO_PRODUCTION_CREATED`: A user requested + a model version be transitioned to production. * + `TRANSITION_REQUEST_TO_ARCHIVED_CREATED`: A user requested a + model version be archived. + http_url_spec: + type: string + description: ':param job_spec: :class:`JobSpec` (optional)' + job_spec: + type: string + status: + type: string + description: ':returns: :class:`UpdateWebhookResponse`' + required: + - id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateWebhookResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ModelRegistryAPI +components: + schemas: + Activity: + type: object + properties: + activity_type: + $ref: '#/components/schemas/ActivityType' + comment: + type: string + description: >- + User-provided comment associated with the activity, comment, or + transition request. + creation_timestamp: + type: integer + description: Creation time of the object, as a Unix timestamp in milliseconds. + from_stage: + type: string + description: >- + Source stage of the transition (if the activity is stage transition + related). Valid values are: * `None`: The initial stage of a model + version. * `Staging`: Staging or pre-production stage. * + `Production`: Production stage. * `Archived`: Archived stage. + id: + type: string + description: Unique identifier for the object. + last_updated_timestamp: + type: integer + description: >- + Time of the object at last update, as a Unix timestamp in + milliseconds. + system_comment: + type: string + description: >- + Comment made by system, for example explaining an activity of type + `SYSTEM_TRANSITION`. It usually describes a side effect, such as a + version being archived as part of another version's stage + transition, and may not be returned for some activity types. + to_stage: + type: string + description: >- + Target stage of the transition (if the activity is stage transition + related). Valid values are: * `None`: The initial stage of a model + version. * `Staging`: Staging or pre-production stage. * + `Production`: Production stage. * `Archived`: Archived stage. + user_id: + type: string + description: The username of the user that created the object. + description: >- + For activities, this contains the activity recorded for the action. For + comments, this contains + the comment details. For transition requests, this contains the transition request details. + ApproveTransitionRequestResponse: + type: object + properties: + activity: + $ref: '#/components/schemas/Activity' + AuthConfig: + type: object + properties: + uc_service_credential_name: + type: string + BackfillSource: + type: object + properties: + delta_table_source: + $ref: '#/components/schemas/DeltaTableSource' + BatchCreateMaterializedFeaturesResponse: + type: object + properties: + materialized_features: + type: array + items: + $ref: '#/components/schemas/MaterializedFeature' + ColumnIdentifier: + type: object + properties: + variant_expr_path: + type: string + required: + - variant_expr_path + CommentObject: + type: object + properties: + available_actions: + type: array + items: + $ref: '#/components/schemas/CommentActivityAction' + description: Array of actions on the activity allowed for the current viewer. + comment: + type: string + description: >- + User-provided comment associated with the activity, comment, or + transition request. + creation_timestamp: + type: integer + description: Creation time of the object, as a Unix timestamp in milliseconds. + id: + type: string + description: Unique identifier for the object. + last_updated_timestamp: + type: integer + description: >- + Time of the object at last update, as a Unix timestamp in + milliseconds. + user_id: + type: string + description: The username of the user that created the object. + description: >- + For activities, this contains the activity recorded for the action. For + comments, this contains + the comment details. For transition requests, this contains the transition request details. + ContinuousWindow: + type: object + properties: + window_duration: + type: string + offset: + type: string + description: The offset of the continuous window (must be non-positive). + required: + - window_duration + CreateCommentResponse: + type: object + properties: + comment: + $ref: '#/components/schemas/CommentObject' + CreateExperimentResponse: + type: object + properties: + experiment_id: + type: string + CreateForecastingExperimentResponse: + type: object + properties: + experiment_id: + type: string + CreateLoggedModelResponse: + type: object + properties: + model: + $ref: '#/components/schemas/LoggedModel' + CreateMaterializedFeatureRequest: + type: object + properties: + materialized_feature: + $ref: '#/components/schemas/MaterializedFeature' + required: + - materialized_feature + CreateModelResponse: + type: object + properties: + registered_model: + $ref: '#/components/schemas/Model' + CreateModelVersionResponse: + type: object + properties: + model_version: + $ref: '#/components/schemas/ModelVersion' + CreateRunResponse: + type: object + properties: + run: + $ref: '#/components/schemas/Run' + CreateTransitionRequestResponse: + type: object + properties: + request: + $ref: '#/components/schemas/TransitionRequest' + CreateWebhookResponse: + type: object + properties: + webhook: + $ref: '#/components/schemas/RegistryWebhook' + DataSource: + type: object + properties: + delta_table_source: + $ref: '#/components/schemas/DeltaTableSource' + kafka_source: + $ref: '#/components/schemas/KafkaSource' + Dataset: + type: object + properties: + name: + type: string + description: >- + The name of the dataset. E.g. “my.uc.table@2” “nyc-taxi-dataset”, + “fantastic-elk-3” + digest: + type: string + description: >- + Dataset digest, e.g. an md5 hash of the dataset that uniquely + identifies it within datasets of the same name. + source_type: + type: string + description: >- + The type of the dataset source, e.g. ‘databricks-uc-table’, ‘DBFS’, + ‘S3’, ... + source: + type: string + description: >- + Source information for the dataset. Note that the source may not + exactly reproduce the dataset if it was transformed / modified + before use with MLflow. + profile: + type: string + description: >- + The profile of the dataset. Summary statistics for the dataset, such + as the number of rows in a table, the mean / std / mode of each + column in a table, or the number of elements in an array. + schema: + type: string + description: >- + The schema of the dataset. E.g., MLflow ColSpec JSON for a + dataframe, MLflow TensorSpec JSON for an ndarray, or another schema + format. + required: + - name + - digest + - source_type + - source + description: >- + Dataset. Represents a reference to data used for training, testing, or + evaluation during the + model development process. + DatasetInput: + type: object + properties: + dataset: + $ref: '#/components/schemas/Dataset' + description: The dataset being used as a Run input. + tags: + type: array + items: + $ref: '#/components/schemas/InputTag' + description: >- + A list of tags for the dataset input, e.g. a “context” tag with + value “training” + required: + - dataset + description: DatasetInput. Represents a dataset and input tags. + DeleteCommentResponse: + type: object + properties: {} + DeleteExperimentResponse: + type: object + properties: {} + DeleteLoggedModelResponse: + type: object + properties: {} + DeleteLoggedModelTagResponse: + type: object + properties: {} + DeleteModelResponse: + type: object + properties: {} + DeleteModelTagResponse: + type: object + properties: {} + DeleteModelVersionResponse: + type: object + properties: {} + DeleteModelVersionTagResponse: + type: object + properties: {} + DeleteRunResponse: + type: object + properties: {} + DeleteRunsResponse: + type: object + properties: + runs_deleted: + type: integer + DeleteTagResponse: + type: object + properties: {} + DeleteTransitionRequestResponse: + type: object + properties: + activity: + $ref: '#/components/schemas/Activity' + DeleteWebhookResponse: + type: object + properties: {} + DeltaTableSource: + type: object + properties: + full_name: + type: string + entity_columns: + type: array + items: + type: string + description: The entity columns of the Delta table. + timeseries_column: + type: string + description: The timeseries column of the Delta table. + required: + - full_name + - entity_columns + - timeseries_column + Experiment: + type: object + properties: + artifact_location: + type: string + description: Location where artifacts for the experiment are stored. + creation_time: + type: integer + description: Creation time + experiment_id: + type: string + description: Unique identifier for the experiment. + last_update_time: + type: integer + description: Last update time + lifecycle_stage: + type: string + description: >- + Current life cycle stage of the experiment: "active" or "deleted". + Deleted experiments are not returned by APIs. + name: + type: string + description: Human readable name that identifies the experiment. + tags: + type: array + items: + $ref: '#/components/schemas/ExperimentTag' + description: 'Tags: Additional metadata key-value pairs.' + description: An experiment and its metadata. + ExperimentAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/ExperimentPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + ExperimentAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/ExperimentPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + ExperimentPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/ExperimentPermissionLevel' + ExperimentPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/ExperimentAccessControlResponse' + object_id: + type: string + object_type: + type: string + ExperimentPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/ExperimentPermissionLevel' + ExperimentTag: + type: object + properties: + key: + type: string + description: The tag key. + value: + type: string + description: The tag value. + description: A tag for an experiment. + Feature: + type: object + properties: + full_name: + type: string + source: + $ref: '#/components/schemas/DataSource' + description: The data source of the feature. + inputs: + type: array + items: + type: string + description: The input columns from which the feature is computed. + function: + $ref: '#/components/schemas/Function' + description: The function by which the feature is computed. + description: + type: string + description: The description of the feature. + filter_condition: + type: string + description: The filter condition applied to the source data before aggregation. + lineage_context: + $ref: '#/components/schemas/LineageContext' + description: >- + WARNING: This field is primarily intended for internal use by + Databricks systems and is automatically populated when features are + created through Databricks notebooks or jobs. Users should not + manually set this field as incorrect values may lead to inaccurate + lineage tracking or unexpected behavior. This field will be set by + feature-engineering client and should be left unset by SDK and + terraform users. + time_window: + $ref: '#/components/schemas/TimeWindow' + description: The time window in which the feature is computed. + required: + - full_name + - source + - inputs + - function + FeatureLineage: + type: object + properties: + feature_specs: + type: array + items: + $ref: '#/components/schemas/FeatureLineageFeatureSpec' + models: + type: array + items: + $ref: '#/components/schemas/FeatureLineageModel' + description: List of Unity Catalog models that were trained on this feature. + online_features: + type: array + items: + $ref: '#/components/schemas/FeatureLineageOnlineFeature' + description: List of online features that use this feature as source. + FeatureLineageFeatureSpec: + type: object + properties: + name: + type: string + FeatureLineageModel: + type: object + properties: + name: + type: string + version: + type: integer + description: The version of the model. + FeatureLineageOnlineFeature: + type: object + properties: + feature_name: + type: string + table_name: + type: string + description: The full name of the online table in Unity Catalog. + FeatureList: + type: object + properties: + features: + type: array + items: + $ref: '#/components/schemas/LinkedFeature' + description: Feature list wrap all the features for a model version + FeatureTag: + type: object + properties: + key: + type: string + value: + type: string + required: + - key + description: Represents a tag on a feature in a feature table. + FileInfo: + type: object + properties: + file_size: + type: integer + description: The size in bytes of the file. Unset for directories. + is_dir: + type: boolean + description: Whether the path is a directory. + path: + type: string + description: The path relative to the root artifact directory run. + description: Metadata of a single artifact file or directory. + FinalizeLoggedModelResponse: + type: object + properties: + model: + $ref: '#/components/schemas/LoggedModel' + ForecastingExperiment: + type: object + properties: + experiment_id: + type: string + description: The unique ID for the forecasting experiment. + experiment_page_url: + type: string + description: The URL to the forecasting experiment page. + state: + $ref: '#/components/schemas/ForecastingExperimentState' + description: The current state of the forecasting experiment. + description: >- + Represents a forecasting experiment with its unique identifier, URL, and + state. + Function: + type: object + properties: + function_type: + $ref: '#/components/schemas/FunctionFunctionType' + extra_parameters: + type: array + items: + $ref: '#/components/schemas/FunctionExtraParameter' + description: Extra parameters for parameterized functions. + required: + - function_type + FunctionExtraParameter: + type: object + properties: + key: + type: string + value: + type: string + description: The value of the parameter. + required: + - key + - value + GetExperimentByNameResponse: + type: object + properties: + experiment: + $ref: '#/components/schemas/Experiment' + GetExperimentPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/ExperimentPermissionsDescription' + GetExperimentResponse: + type: object + properties: + experiment: + $ref: '#/components/schemas/Experiment' + GetLatestVersionsResponse: + type: object + properties: + model_versions: + type: array + items: + $ref: '#/components/schemas/ModelVersion' + GetLoggedModelResponse: + type: object + properties: + model: + $ref: '#/components/schemas/LoggedModel' + GetMetricHistoryResponse: + type: object + properties: + metrics: + type: array + items: + $ref: '#/components/schemas/Metric' + next_page_token: + type: string + description: >- + A token that can be used to issue a query for the next page of + metric history values. A missing token indicates that no additional + metrics are available to fetch. + GetModelResponse: + type: object + properties: + registered_model_databricks: + $ref: '#/components/schemas/ModelDatabricks' + GetModelVersionDownloadUriResponse: + type: object + properties: + artifact_uri: + type: string + GetModelVersionResponse: + type: object + properties: + model_version: + $ref: '#/components/schemas/ModelVersion' + GetRegisteredModelPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/RegisteredModelPermissionsDescription' + GetRunResponse: + type: object + properties: + run: + $ref: '#/components/schemas/Run' + HttpUrlSpec: + type: object + properties: + url: + type: string + authorization: + type: string + description: >- + Value of the authorization header that should be sent in the request + sent by the wehbook. It should be of the form `" + "`. If set to an empty string, no authorization header + will be included in the request. + enable_ssl_verification: + type: boolean + description: >- + Enable/disable SSL certificate validation. Default is true. For + self-signed certificates, this field must be false AND the + destination server must disable certificate validation as well. For + security purposes, it is encouraged to perform secret validation + with the HMAC-encoded portion of the payload and acknowledge the + risk associated with disabling hostname validation whereby it + becomes more likely that requests can be maliciously routed to an + unintended host. + secret: + type: string + description: >- + Shared secret required for HMAC encoding payload. The HMAC-encoded + payload will be sent in the header as: { "X-Databricks-Signature": + $encoded_payload }. + required: + - url + HttpUrlSpecWithoutSecret: + type: object + properties: + enable_ssl_verification: + type: boolean + url: + type: string + description: >- + External HTTPS URL called on event trigger (by using a POST + request). + InputTag: + type: object + properties: + key: + type: string + description: The tag key. + value: + type: string + description: The tag value. + required: + - key + - value + description: Tag for a dataset input. + JobContext: + type: object + properties: + job_id: + type: integer + job_run_id: + type: integer + description: The job run ID where this API was invoked. + JobSpec: + type: object + properties: + job_id: + type: string + access_token: + type: string + description: The personal access token used to authorize webhook's job runs. + workspace_url: + type: string + description: >- + URL of the workspace containing the job that this webhook runs. If + not specified, the job’s workspace URL is assumed to be the same as + the workspace where the webhook is created. + required: + - job_id + - access_token + JobSpecWithoutSecret: + type: object + properties: + job_id: + type: string + workspace_url: + type: string + description: >- + URL of the workspace containing the job that this webhook runs. If + not specified, the job’s workspace URL is assumed to be the same as + the workspace where the webhook is created. + KafkaConfig: + type: object + properties: + name: + type: string + bootstrap_servers: + type: string + description: A comma-separated list of host/port pairs pointing to Kafka cluster. + subscription_mode: + $ref: '#/components/schemas/SubscriptionMode' + description: Options to configure which Kafka topics to pull data from. + auth_config: + $ref: '#/components/schemas/AuthConfig' + description: Authentication configuration for connection to topics. + backfill_source: + $ref: '#/components/schemas/BackfillSource' + description: >- + A user-provided and managed source for backfilling data. Historical + data is used when creating a training set from streaming features + linked to this Kafka config. In the future, a separate table will be + maintained by Databricks for forward filling data. The schema for + this source must match exactly that of the key and value schemas + specified for this Kafka config. + extra_options: + type: object + description: >- + Catch-all for miscellaneous options. Keys should be source options + or Kafka consumer options (kafka.*) + key_schema: + $ref: '#/components/schemas/SchemaConfig' + description: >- + Schema configuration for extracting message keys from topics. At + least one of key_schema and value_schema must be provided. + value_schema: + $ref: '#/components/schemas/SchemaConfig' + description: >- + Schema configuration for extracting message values from topics. At + least one of key_schema and value_schema must be provided. + required: + - name + - bootstrap_servers + - subscription_mode + - auth_config + KafkaSource: + type: object + properties: + name: + type: string + entity_column_identifiers: + type: array + items: + $ref: '#/components/schemas/ColumnIdentifier' + description: The entity column identifiers of the Kafka source. + timeseries_column_identifier: + $ref: '#/components/schemas/ColumnIdentifier' + description: The timeseries column identifier of the Kafka source. + required: + - name + - entity_column_identifiers + - timeseries_column_identifier + LineageContext: + type: object + properties: + job_context: + $ref: '#/components/schemas/JobContext' + description: Job context information including job ID and run ID. + notebook_id: + type: integer + description: The notebook ID where this API was invoked. + description: >- + Lineage context information for tracking where an API was invoked. This + will allow us to track + lineage, which currently uses caller entity information for use across the Lineage Client and + Observability in Lumberjack. + LinkedFeature: + type: object + properties: + feature_name: + type: string + description: Feature name + feature_table_id: + type: string + description: Feature table id + feature_table_name: + type: string + description: Feature table name + description: >- + Feature for model version. ([ML-57150] Renamed from Feature to + LinkedFeature) + ListArtifactsResponse: + type: object + properties: + files: + type: array + items: + $ref: '#/components/schemas/FileInfo' + next_page_token: + type: string + description: >- + The token that can be used to retrieve the next page of artifact + results. + root_uri: + type: string + description: The root artifact directory for the run. + ListExperimentsResponse: + type: object + properties: + experiments: + type: array + items: + $ref: '#/components/schemas/Experiment' + next_page_token: + type: string + description: >- + Token that can be used to retrieve the next page of experiments. + Empty token means no more experiment is available for retrieval. + ListFeatureTagsResponse: + type: object + properties: + feature_tags: + type: array + items: + $ref: '#/components/schemas/FeatureTag' + next_page_token: + type: string + description: Pagination token to request the next page of results for this query. + description: Response message for ListFeatureTag. + ListFeaturesResponse: + type: object + properties: + features: + type: array + items: + $ref: '#/components/schemas/Feature' + next_page_token: + type: string + description: Pagination token to request the next page of results for this query. + ListKafkaConfigsResponse: + type: object + properties: + kafka_configs: + type: array + items: + $ref: '#/components/schemas/KafkaConfig' + next_page_token: + type: string + description: Pagination token to request the next page of results for this query. + required: + - kafka_configs + ListMaterializedFeaturesResponse: + type: object + properties: + materialized_features: + type: array + items: + $ref: '#/components/schemas/MaterializedFeature' + next_page_token: + type: string + description: Pagination token to request the next page of results for this query. + ListModelsResponse: + type: object + properties: + next_page_token: + type: string + registered_models: + type: array + items: + $ref: '#/components/schemas/Model' + ListOnlineStoresResponse: + type: object + properties: + next_page_token: + type: string + online_stores: + type: array + items: + $ref: '#/components/schemas/OnlineStore' + description: List of online stores. + ListRegistryWebhooks: + type: object + properties: + next_page_token: + type: string + webhooks: + type: array + items: + $ref: '#/components/schemas/RegistryWebhook' + description: Array of registry webhooks. + ListTransitionRequestsResponse: + type: object + properties: + requests: + type: array + items: + $ref: '#/components/schemas/Activity' + LogBatchResponse: + type: object + properties: {} + LogInputsResponse: + type: object + properties: {} + LogLoggedModelParamsRequestResponse: + type: object + properties: {} + LogMetricResponse: + type: object + properties: {} + LogModelResponse: + type: object + properties: {} + LogOutputsResponse: + type: object + properties: {} + LogParamResponse: + type: object + properties: {} + LoggedModel: + type: object + properties: + data: + $ref: '#/components/schemas/LoggedModelData' + description: The params and metrics attached to the logged model. + info: + $ref: '#/components/schemas/LoggedModelInfo' + description: The logged model attributes such as model ID, status, tags, etc. + description: >- + A logged model message includes logged model attributes, tags, + registration info, params, and + linked run metrics. + LoggedModelData: + type: object + properties: + metrics: + type: array + items: + $ref: '#/components/schemas/Metric' + description: Performance metrics linked to the model. + params: + type: array + items: + $ref: '#/components/schemas/LoggedModelParameter' + description: Immutable string key-value pairs of the model. + description: >- + A LoggedModelData message includes logged model params and linked + metrics. + LoggedModelInfo: + type: object + properties: + artifact_uri: + type: string + description: The URI of the directory where model artifacts are stored. + creation_timestamp_ms: + type: integer + description: >- + The timestamp when the model was created in milliseconds since the + UNIX epoch. + creator_id: + type: integer + description: The ID of the user or principal that created the model. + experiment_id: + type: string + description: The ID of the experiment that owns the model. + last_updated_timestamp_ms: + type: integer + description: >- + The timestamp when the model was last updated in milliseconds since + the UNIX epoch. + model_id: + type: string + description: The unique identifier for the logged model. + model_type: + type: string + description: The type of model, such as ``"Agent"``, ``"Classifier"``, ``"LLM"``. + name: + type: string + description: The name of the model. + source_run_id: + type: string + description: The ID of the run that created the model. + status: + $ref: '#/components/schemas/LoggedModelStatus' + description: The status of whether or not the model is ready for use. + status_message: + type: string + description: Details on the current model status. + tags: + type: array + items: + $ref: '#/components/schemas/LoggedModelTag' + description: Mutable string key-value pairs set on the model. + description: >- + A LoggedModelInfo includes logged model attributes, tags, and + registration info. + LoggedModelParameter: + type: object + properties: + key: + type: string + description: The key identifying this param. + value: + type: string + description: The value of this param. + description: Parameter associated with a LoggedModel. + LoggedModelTag: + type: object + properties: + key: + type: string + description: The tag key. + value: + type: string + description: The tag value. + description: Tag for a LoggedModel. + MaterializedFeature: + type: object + properties: + feature_name: + type: string + description: The full name of the feature in Unity Catalog. + cron_schedule: + type: string + description: >- + The quartz cron expression that defines the schedule of the + materialization pipeline. The schedule is evaluated in the UTC + timezone. + last_materialization_time: + type: string + description: >- + The timestamp when the pipeline last ran and updated the + materialized feature values. If the pipeline has not run yet, this + field will be null. + materialized_feature_id: + type: string + description: Unique identifier for the materialized feature. + offline_store_config: + $ref: '#/components/schemas/OfflineStoreConfig' + online_store_config: + $ref: '#/components/schemas/OnlineStoreConfig' + pipeline_schedule_state: + $ref: '#/components/schemas/MaterializedFeaturePipelineScheduleState' + description: The schedule state of the materialization pipeline. + table_name: + type: string + description: >- + The fully qualified Unity Catalog path to the table containing the + materialized feature (Delta table or Lakebase table). Output only. + required: + - feature_name + description: >- + A materialized feature represents a feature that is continuously + computed and stored. + Metric: + type: object + properties: + dataset_digest: + type: string + description: >- + The dataset digest of the dataset associated with the metric, e.g. + an md5 hash of the dataset that uniquely identifies it within + datasets of the same name. + dataset_name: + type: string + description: >- + The name of the dataset associated with the metric. E.g. + “my.uc.table@2” “nyc-taxi-dataset”, “fantastic-elk-3” + key: + type: string + description: The key identifying the metric. + model_id: + type: string + description: >- + The ID of the logged model or registered model version associated + with the metric, if applicable. + run_id: + type: string + description: The ID of the run containing the metric. + step: + type: integer + description: The step at which the metric was logged. + timestamp: + type: integer + description: The timestamp at which the metric was recorded. + value: + type: number + description: The value of the metric. + description: Metric associated with a run, represented as a key-value pair. + Model: + type: object + properties: + creation_timestamp: + type: integer + description: + type: string + description: Description of this `registered_model`. + last_updated_timestamp: + type: integer + description: >- + Timestamp recorded when metadata for this `registered_model` was + last updated. + latest_versions: + type: array + items: + $ref: '#/components/schemas/ModelVersion' + description: >- + Collection of latest model versions for each stage. Only contains + models with current `READY` status. + name: + type: string + description: Unique name for the model. + tags: + type: array + items: + $ref: '#/components/schemas/ModelTag' + description: >- + Tags: Additional metadata key-value pairs for this + `registered_model`. + user_id: + type: string + description: User that created this `registered_model` + ModelDatabricks: + type: object + properties: + creation_timestamp: + type: integer + description: + type: string + description: User-specified description for the object. + id: + type: string + description: Unique identifier for the object. + last_updated_timestamp: + type: integer + description: Last update time of the object, as a Unix timestamp in milliseconds. + latest_versions: + type: array + items: + $ref: '#/components/schemas/ModelVersion' + description: Array of model versions, each the latest version for its stage. + name: + type: string + description: Name of the model. + permission_level: + $ref: '#/components/schemas/PermissionLevel' + description: >- + Permission level granted for the requesting user on this registered + model + tags: + type: array + items: + $ref: '#/components/schemas/ModelTag' + description: Array of tags associated with the model. + user_id: + type: string + description: The username of the user that created the object. + ModelInput: + type: object + properties: + model_id: + type: string + description: The unique identifier of the model. + required: + - model_id + description: Represents a LoggedModel or Registered Model Version input to a Run. + ModelOutput: + type: object + properties: + model_id: + type: string + description: The unique identifier of the model. + step: + type: integer + description: The step at which the model was produced. + required: + - model_id + - step + description: Represents a LoggedModel output of a Run. + ModelTag: + type: object + properties: + key: + type: string + description: The tag key. + value: + type: string + description: The tag value. + description: Tag for a registered model + ModelVersion: + type: object + properties: + creation_timestamp: + type: integer + current_stage: + type: string + description: Current stage for this `model_version`. + description: + type: string + description: Description of this `model_version`. + last_updated_timestamp: + type: integer + description: >- + Timestamp recorded when metadata for this `model_version` was last + updated. + name: + type: string + description: Unique name of the model + run_id: + type: string + description: >- + MLflow run ID used when creating `model_version`, if `source` was + generated by an experiment run stored in MLflow tracking server. + run_link: + type: string + description: 'Run Link: Direct link to the run that generated this version' + source: + type: string + description: >- + URI indicating the location of the source model artifacts, used when + creating `model_version` + status: + $ref: '#/components/schemas/ModelVersionStatus' + description: Current status of `model_version` + status_message: + type: string + description: Details on current `status`, if it is pending or failed. + tags: + type: array + items: + $ref: '#/components/schemas/ModelVersionTag' + description: 'Tags: Additional metadata key-value pairs for this `model_version`.' + user_id: + type: string + description: User that created this `model_version`. + version: + type: string + description: Model's version number. + ModelVersionDatabricks: + type: object + properties: + creation_timestamp: + type: integer + current_stage: + type: string + description: + type: string + description: User-specified description for the object. + email_subscription_status: + $ref: '#/components/schemas/RegistryEmailSubscriptionType' + description: >- + Email Subscription Status: This is the subscription status of the + user to the model version Users get subscribed by interacting with + the model version. + feature_list: + $ref: '#/components/schemas/FeatureList' + description: Feature lineage of `model_version`. + last_updated_timestamp: + type: integer + description: >- + Time of the object at last update, as a Unix timestamp in + milliseconds. + name: + type: string + description: Name of the model. + open_requests: + type: array + items: + $ref: '#/components/schemas/Activity' + description: >- + Open requests for this `model_versions`. Gap in sequence number is + intentional and is done in order to match field sequence numbers of + `ModelVersion` proto message + permission_level: + $ref: '#/components/schemas/PermissionLevel' + run_id: + type: string + description: >- + Unique identifier for the MLflow tracking run associated with the + source model artifacts. + run_link: + type: string + description: >- + URL of the run associated with the model artifacts. This field is + set at model version creation time only for model versions whose + source run is from a tracking server that is different from the + registry server. + source: + type: string + description: >- + URI that indicates the location of the source model artifacts. This + is used when creating the model version. + status: + $ref: '#/components/schemas/Status' + status_message: + type: string + description: Details on the current status, for example why registration failed. + tags: + type: array + items: + $ref: '#/components/schemas/ModelVersionTag' + description: Array of tags that are associated with the model version. + user_id: + type: string + description: The username of the user that created the object. + version: + type: string + description: Version of the model. + ModelVersionTag: + type: object + properties: + key: + type: string + value: + type: string + description: The tag value. + OfflineStoreConfig: + type: object + properties: + catalog_name: + type: string + description: The Unity Catalog catalog name. + schema_name: + type: string + description: The Unity Catalog schema name. + table_name_prefix: + type: string + description: >- + Prefix for Unity Catalog table name. The materialized feature will + be stored in a table with this prefix and a generated postfix. + required: + - catalog_name + - schema_name + - table_name_prefix + description: Configuration for offline store destination. + OnlineStore: + type: object + properties: + name: + type: string + description: >- + The name of the online store. This is the unique identifier for the + online store. + capacity: + type: string + description: >- + The capacity of the online store. Valid values are "CU_1", "CU_2", + "CU_4", "CU_8". + creation_time: + type: string + description: The timestamp when the online store was created. + creator: + type: string + description: The email of the creator of the online store. + read_replica_count: + type: integer + description: The number of read replicas for the online store. Defaults to 0. + state: + $ref: '#/components/schemas/OnlineStoreState' + description: The current state of the online store. + usage_policy_id: + type: string + description: The usage policy applied to the online store to track billing. + required: + - name + - capacity + description: >- + An OnlineStore is a logical database instance that stores and serves + features online. + OnlineStoreConfig: + type: object + properties: + catalog_name: + type: string + description: >- + The Unity Catalog catalog name. This name is also used as the + Lakebase logical database name. + schema_name: + type: string + description: The Unity Catalog schema name. + table_name_prefix: + type: string + description: >- + Prefix for Unity Catalog table name. The materialized feature will + be stored in a Lakebase table with this prefix and a generated + postfix. + online_store_name: + type: string + description: The name of the target online store. + required: + - catalog_name + - schema_name + - table_name_prefix + - online_store_name + description: Configuration for online store destination. + Param: + type: object + properties: + key: + type: string + description: Key identifying this param. + value: + type: string + description: Value associated with this param. + description: Param associated with a run. + PublishSpec: + type: object + properties: + online_store: + type: string + online_table_name: + type: string + description: >- + The full three-part (catalog, schema, table) name of the online + table. + publish_mode: + $ref: '#/components/schemas/PublishSpecPublishMode' + description: >- + The publish mode of the pipeline that syncs the online table with + the source table. + required: + - online_store + - online_table_name + - publish_mode + PublishTableResponse: + type: object + properties: + online_table_name: + type: string + pipeline_id: + type: string + description: >- + The ID of the pipeline that syncs the online table with the source + table. + RegisteredModelAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/RegisteredModelPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + RegisteredModelAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/RegisteredModelPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + RegisteredModelPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/RegisteredModelPermissionLevel' + RegisteredModelPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/RegisteredModelAccessControlResponse' + object_id: + type: string + object_type: + type: string + RegisteredModelPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/RegisteredModelPermissionLevel' + RegistryWebhook: + type: object + properties: + creation_timestamp: + type: integer + description: + type: string + description: User-specified description for the webhook. + events: + type: array + items: + $ref: '#/components/schemas/RegistryWebhookEvent' + description: >- + Events that can trigger a registry webhook: * + `MODEL_VERSION_CREATED`: A new model version was created for the + associated model. * `MODEL_VERSION_TRANSITIONED_STAGE`: A model + version’s stage was changed. * `TRANSITION_REQUEST_CREATED`: A user + requested a model version’s stage be transitioned. * + `COMMENT_CREATED`: A user wrote a comment on a registered model. * + `REGISTERED_MODEL_CREATED`: A new registered model was created. This + event type can only be specified for a registry-wide webhook, which + can be created by not specifying a model name in the create request. + * `MODEL_VERSION_TAG_SET`: A user set a tag on the model version. * + `MODEL_VERSION_TRANSITIONED_TO_STAGING`: A model version was + transitioned to staging. * + `MODEL_VERSION_TRANSITIONED_TO_PRODUCTION`: A model version was + transitioned to production. * + `MODEL_VERSION_TRANSITIONED_TO_ARCHIVED`: A model version was + archived. * `TRANSITION_REQUEST_TO_STAGING_CREATED`: A user + requested a model version be transitioned to staging. * + `TRANSITION_REQUEST_TO_PRODUCTION_CREATED`: A user requested a model + version be transitioned to production. * + `TRANSITION_REQUEST_TO_ARCHIVED_CREATED`: A user requested a model + version be archived. + http_url_spec: + $ref: '#/components/schemas/HttpUrlSpecWithoutSecret' + id: + type: string + description: Webhook ID + job_spec: + $ref: '#/components/schemas/JobSpecWithoutSecret' + last_updated_timestamp: + type: integer + description: >- + Time of the object at last update, as a Unix timestamp in + milliseconds. + model_name: + type: string + description: Name of the model whose events would trigger this webhook. + status: + $ref: '#/components/schemas/RegistryWebhookStatus' + RejectTransitionRequestResponse: + type: object + properties: + activity: + $ref: '#/components/schemas/Activity' + RenameModelResponse: + type: object + properties: + registered_model: + $ref: '#/components/schemas/Model' + RestoreExperimentResponse: + type: object + properties: {} + RestoreRunResponse: + type: object + properties: {} + RestoreRunsResponse: + type: object + properties: + runs_restored: + type: integer + Run: + type: object + properties: + data: + $ref: '#/components/schemas/RunData' + description: Run data. + info: + $ref: '#/components/schemas/RunInfo' + description: Run metadata. + inputs: + $ref: '#/components/schemas/RunInputs' + description: Run inputs. + description: A single run. + RunData: + type: object + properties: + metrics: + type: array + items: + $ref: '#/components/schemas/Metric' + description: Run metrics. + params: + type: array + items: + $ref: '#/components/schemas/Param' + description: Run parameters. + tags: + type: array + items: + $ref: '#/components/schemas/RunTag' + description: Additional metadata key-value pairs. + description: Run data (metrics, params, and tags). + RunInfo: + type: object + properties: + artifact_uri: + type: string + description: >- + URI of the directory where artifacts should be uploaded. This can be + a local path (starting with "/"), or a distributed file system (DFS) + path, like ``s3://bucket/directory`` or ``dbfs:/my/directory``. If + not set, the local ``./mlruns`` directory is chosen. + end_time: + type: integer + description: Unix timestamp of when the run ended in milliseconds. + experiment_id: + type: string + description: The experiment ID. + lifecycle_stage: + type: string + description: >- + Current life cycle stage of the experiment : OneOf("active", + "deleted") + run_id: + type: string + description: Unique identifier for the run. + run_name: + type: string + description: The name of the run. + run_uuid: + type: string + description: >- + [Deprecated, use run_id instead] Unique identifier for the run. This + field will be removed in a future MLflow version. + start_time: + type: integer + description: Unix timestamp of when the run started in milliseconds. + status: + $ref: '#/components/schemas/RunInfoStatus' + description: Current status of the run. + user_id: + type: string + description: >- + User who initiated the run. This field is deprecated as of MLflow + 1.0, and will be removed in a future MLflow release. Use + 'mlflow.user' tag instead. + description: Metadata of a single run. + RunInputs: + type: object + properties: + dataset_inputs: + type: array + items: + $ref: '#/components/schemas/DatasetInput' + description: Run metrics. + model_inputs: + type: array + items: + $ref: '#/components/schemas/ModelInput' + description: Model inputs to the Run. + description: Run inputs. + RunTag: + type: object + properties: + key: + type: string + description: The tag key. + value: + type: string + description: The tag value. + description: Tag for a run. + SchemaConfig: + type: object + properties: + json_schema: + type: string + SearchExperimentsResponse: + type: object + properties: + experiments: + type: array + items: + $ref: '#/components/schemas/Experiment' + next_page_token: + type: string + description: >- + Token that can be used to retrieve the next page of experiments. An + empty token means that no more experiments are available for + retrieval. + SearchLoggedModelsDataset: + type: object + properties: + dataset_name: + type: string + dataset_digest: + type: string + description: The digest of the dataset. + required: + - dataset_name + SearchLoggedModelsOrderBy: + type: object + properties: + field_name: + type: string + ascending: + type: boolean + description: Whether the search results order is ascending or not. + dataset_digest: + type: string + description: >- + If ``field_name`` refers to a metric, this field specifies the + digest of the dataset associated with the metric. Only metrics + associated with the specified dataset name and digest will be + considered for ordering. This field may only be set if + ``dataset_name`` is also set. + dataset_name: + type: string + description: >- + If ``field_name`` refers to a metric, this field specifies the name + of the dataset associated with the metric. Only metrics associated + with the specified dataset name will be considered for ordering. + This field may only be set if ``field_name`` refers to a metric. + required: + - field_name + SearchLoggedModelsResponse: + type: object + properties: + models: + type: array + items: + $ref: '#/components/schemas/LoggedModel' + next_page_token: + type: string + description: >- + The token that can be used to retrieve the next page of logged + models. + SearchModelVersionsResponse: + type: object + properties: + model_versions: + type: array + items: + $ref: '#/components/schemas/ModelVersion' + next_page_token: + type: string + description: >- + Pagination token to request next page of models for the same search + query. + SearchModelsResponse: + type: object + properties: + next_page_token: + type: string + registered_models: + type: array + items: + $ref: '#/components/schemas/Model' + description: Registered Models that match the search criteria. + SearchRunsResponse: + type: object + properties: + next_page_token: + type: string + runs: + type: array + items: + $ref: '#/components/schemas/Run' + description: Runs that match the search criteria. + SetExperimentTagResponse: + type: object + properties: {} + SetLoggedModelTagsResponse: + type: object + properties: {} + SetModelTagResponse: + type: object + properties: {} + SetModelVersionTagResponse: + type: object + properties: {} + SetTagResponse: + type: object + properties: {} + SlidingWindow: + type: object + properties: + window_duration: + type: string + slide_duration: + type: string + description: >- + The slide duration (interval by which windows advance, must be + positive and less than duration). + required: + - window_duration + - slide_duration + SubscriptionMode: + type: object + properties: + assign: + type: string + subscribe: + type: string + description: >- + A comma-separated list of Kafka topics to read from. For example, + 'topicA,topicB,topicC'. + subscribe_pattern: + type: string + description: >- + A regular expression matching topics to subscribe to. For example, + 'topic.*' will subscribe to all topics starting with 'topic'. + TestRegistryWebhookResponse: + type: object + properties: + body: + type: string + status_code: + type: integer + description: Status code returned by the webhook URL + TimeWindow: + type: object + properties: + continuous: + $ref: '#/components/schemas/ContinuousWindow' + sliding: + $ref: '#/components/schemas/SlidingWindow' + tumbling: + $ref: '#/components/schemas/TumblingWindow' + TransitionRequest: + type: object + properties: + available_actions: + type: array + items: + $ref: '#/components/schemas/ActivityAction' + description: Array of actions on the activity allowed for the current viewer. + comment: + type: string + description: >- + User-provided comment associated with the activity, comment, or + transition request. + creation_timestamp: + type: integer + description: Creation time of the object, as a Unix timestamp in milliseconds. + to_stage: + type: string + description: >- + Target stage of the transition (if the activity is stage transition + related). Valid values are: * `None`: The initial stage of a model + version. * `Staging`: Staging or pre-production stage. * + `Production`: Production stage. * `Archived`: Archived stage. + user_id: + type: string + description: The username of the user that created the object. + description: >- + For activities, this contains the activity recorded for the action. For + comments, this contains + the comment details. For transition requests, this contains the transition request details. + TransitionStageResponse: + type: object + properties: + model_version_databricks: + $ref: '#/components/schemas/ModelVersionDatabricks' + TumblingWindow: + type: object + properties: + window_duration: + type: string + required: + - window_duration + UpdateCommentResponse: + type: object + properties: + comment: + $ref: '#/components/schemas/CommentObject' + UpdateExperimentResponse: + type: object + properties: {} + UpdateModelResponse: + type: object + properties: + registered_model: + $ref: '#/components/schemas/Model' + UpdateModelVersionResponse: + type: object + properties: + model_version: + $ref: '#/components/schemas/ModelVersion' + UpdateRunResponse: + type: object + properties: + run_info: + $ref: '#/components/schemas/RunInfo' + UpdateWebhookResponse: + type: object + properties: + webhook: + $ref: '#/components/schemas/RegistryWebhook' + ActivityAction: + type: string + enum: + - APPROVE_TRANSITION_REQUEST + - CANCEL_TRANSITION_REQUEST + - DELETE_COMMENT + - EDIT_COMMENT + - REJECT_TRANSITION_REQUEST + description: >- + An action that a user (with sufficient permissions) could take on an + activity or comment. + + + For activities, valid values are: * `APPROVE_TRANSITION_REQUEST`: + Approve a transition request + + + * `REJECT_TRANSITION_REQUEST`: Reject a transition request + + + * `CANCEL_TRANSITION_REQUEST`: Cancel (delete) a transition request + + + For comments, valid values are: * `EDIT_COMMENT`: Edit the comment + + + * `DELETE_COMMENT`: Delete the comment + ActivityType: + type: string + enum: + - APPLIED_TRANSITION + - APPROVED_REQUEST + - CANCELLED_REQUEST + - NEW_COMMENT + - REJECTED_REQUEST + - REQUESTED_TRANSITION + - SYSTEM_TRANSITION + description: >- + Type of activity. Valid values are: * `APPLIED_TRANSITION`: User applied + the corresponding stage + + transition. + + + * `REQUESTED_TRANSITION`: User requested the corresponding stage + transition. + + + * `CANCELLED_REQUEST`: User cancelled an existing transition request. + + + * `APPROVED_REQUEST`: User approved the corresponding stage transition. + + + * `REJECTED_REQUEST`: User rejected the coressponding stage transition. + + + * `SYSTEM_TRANSITION`: For events performed as a side effect, such as + archiving existing model + + versions in a stage. + CommentActivityAction: + type: string + enum: + - APPROVE_TRANSITION_REQUEST + - CANCEL_TRANSITION_REQUEST + - DELETE_COMMENT + - EDIT_COMMENT + - REJECT_TRANSITION_REQUEST + description: >- + An action that a user (with sufficient permissions) could take on an + activity or comment. + + + For activities, valid values are: * `APPROVE_TRANSITION_REQUEST`: + Approve a transition request + + + * `REJECT_TRANSITION_REQUEST`: Reject a transition request + + + * `CANCEL_TRANSITION_REQUEST`: Cancel (delete) a transition request + + + For comments, valid values are: * `EDIT_COMMENT`: Edit the comment + + + * `DELETE_COMMENT`: Delete the comment + ExperimentPermissionLevel: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_READ + description: Permission level + ForecastingExperimentState: + type: string + enum: + - CANCELLED + - FAILED + - PENDING + - RUNNING + - SUCCEEDED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionFunctionType: + type: string + enum: + - APPROX_COUNT_DISTINCT + - APPROX_PERCENTILE + - AVG + - COUNT + - FIRST + - LAST + - MAX + - MIN + - STDDEV_POP + - STDDEV_SAMP + - SUM + - VAR_POP + - VAR_SAMP + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + LoggedModelStatus: + type: string + enum: + - LOGGED_MODEL_PENDING + - LOGGED_MODEL_READY + - LOGGED_MODEL_UPLOAD_FAILED + description: A LoggedModelStatus enum value represents the status of a logged model. + MaterializedFeaturePipelineScheduleState: + type: string + enum: + - ACTIVE + - PAUSED + - SNAPSHOT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ModelVersionStatus: + type: string + enum: + - FAILED_REGISTRATION + - PENDING_REGISTRATION + - READY + description: >- + The status of the model version. Valid values are: * + `PENDING_REGISTRATION`: Request to register + + a new model version is pending as server performs background tasks. + + + * `FAILED_REGISTRATION`: Request to register a new model version has + failed. + + + * `READY`: Model version is ready for use. + OnlineStoreState: + type: string + enum: + - AVAILABLE + - DELETING + - FAILING_OVER + - STARTING + - STOPPED + - UPDATING + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PermissionLevel: + type: string + enum: + - CAN_CREATE_REGISTERED_MODEL + - CAN_EDIT + - CAN_MANAGE + - CAN_MANAGE_PRODUCTION_VERSIONS + - CAN_MANAGE_STAGING_VERSIONS + - CAN_READ + description: >- + Permission level of the requesting user on the object. For what is + allowed at each level, see + + [MLflow Model permissions](..). + PublishSpecPublishMode: + type: string + enum: + - CONTINUOUS + - SNAPSHOT + - TRIGGERED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + RegisteredModelPermissionLevel: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_MANAGE_PRODUCTION_VERSIONS + - CAN_MANAGE_STAGING_VERSIONS + - CAN_READ + description: Permission level + RegistryEmailSubscriptionType: + type: string + enum: + - ALL_EVENTS + - DEFAULT + - SUBSCRIBED + - UNSUBSCRIBED + description: >- + .. note:: Experimental: This entity may change or be removed in a future + release without + + warning. Email subscription types for registry notifications: - + `ALL_EVENTS`: Subscribed to all + + events. - `DEFAULT`: Default subscription type. - `SUBSCRIBED`: + Subscribed to notifications. - + + `UNSUBSCRIBED`: Not subscribed to notifications. + RegistryWebhookEvent: + type: string + enum: + - COMMENT_CREATED + - MODEL_VERSION_CREATED + - MODEL_VERSION_TAG_SET + - MODEL_VERSION_TRANSITIONED_STAGE + - MODEL_VERSION_TRANSITIONED_TO_ARCHIVED + - MODEL_VERSION_TRANSITIONED_TO_PRODUCTION + - MODEL_VERSION_TRANSITIONED_TO_STAGING + - REGISTERED_MODEL_CREATED + - TRANSITION_REQUEST_CREATED + - TRANSITION_REQUEST_TO_ARCHIVED_CREATED + - TRANSITION_REQUEST_TO_PRODUCTION_CREATED + - TRANSITION_REQUEST_TO_STAGING_CREATED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + RegistryWebhookStatus: + type: string + enum: + - ACTIVE + - DISABLED + - TEST_MODE + description: >- + Enable or disable triggering the webhook, or put the webhook into test + mode. The default is + + `ACTIVE`: * `ACTIVE`: Webhook is triggered when an associated event + happens. + + + * `DISABLED`: Webhook is not triggered. + + + * `TEST_MODE`: Webhook can be triggered through the test endpoint, but + is not triggered on a + + real event. + RunInfoStatus: + type: string + enum: + - FAILED + - FINISHED + - KILLED + - RUNNING + - SCHEDULED + description: Status of a run. + Status: + type: string + enum: + - FAILED_REGISTRATION + - PENDING_REGISTRATION + - READY + description: >- + The status of the model version. Valid values are: * + `PENDING_REGISTRATION`: Request to register + + a new model version is pending as server performs background tasks. + + + * `FAILED_REGISTRATION`: Request to register a new model version has + failed. + + + * `READY`: Model version is ready for use. + UpdateRunStatus: + type: string + enum: + - FAILED + - FINISHED + - KILLED + - RUNNING + - SCHEDULED + description: Status of a run. + ViewType: + type: string + enum: + - ACTIVE_ONLY + - ALL + - DELETED_ONLY + description: Qualifier for the view type. + x-stackQL-resources: + experiments: + id: databricks_workspace.ml.experiments + name: experiments + title: Experiments + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get_by_name: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1get-by-name/get' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.experiments + restore: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1restore/post' + response: + mediaType: application/json + openAPIDocKey: '200' + search: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1search/post' + response: + mediaType: application/json + openAPIDocKey: '200' + set_tag: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1set-experiment-tag/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1experiments~1update/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/experiments/methods/get_by_name' + - $ref: '#/components/x-stackQL-resources/experiments/methods/get' + - $ref: '#/components/x-stackQL-resources/experiments/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/experiments/methods/create' + update: [] + delete: [] + replace: [] + logged_models: + id: databricks_workspace.ml.logged_models + name: logged_models + title: Logged Models + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1logged-models/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1logged-models~1{model_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + finalize: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1logged-models~1{model_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1logged-models~1{model_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + delete_tag: + operation: + $ref: >- + #/paths/~1api~12.0~1mlflow~1logged-models~1{model_id}~1tags~1{tag_key}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + log_params: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1logged-models~1{model_id}~1params/post' + response: + mediaType: application/json + openAPIDocKey: '200' + search: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1logged-models~1search/post' + response: + mediaType: application/json + openAPIDocKey: '200' + set_tags: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1logged-models~1{model_id}~1tags/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/logged_models/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/logged_models/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/logged_models/methods/delete_tag' + - $ref: '#/components/x-stackQL-resources/logged_models/methods/delete' + replace: [] + experiment_runs: + id: databricks_workspace.ml.experiment_runs + name: experiment_runs + title: Experiment Runs + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete_bulk: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1databricks~1runs~1delete-runs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete_tag: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1delete-tag/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + log_batch: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-batch/post' + response: + mediaType: application/json + openAPIDocKey: '200' + log_inputs: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-inputs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + log_metric: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-metric/post' + response: + mediaType: application/json + openAPIDocKey: '200' + log_model: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-model/post' + response: + mediaType: application/json + openAPIDocKey: '200' + log_outputs: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1outputs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + log_param: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1log-parameter/post' + response: + mediaType: application/json + openAPIDocKey: '200' + restore: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1restore/post' + response: + mediaType: application/json + openAPIDocKey: '200' + restore_bulk: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1databricks~1runs~1restore-runs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + search: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1search/post' + response: + mediaType: application/json + openAPIDocKey: '200' + set_tag: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1set-tag/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1runs~1update/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/experiment_runs/methods/create' + update: [] + delete: [] + replace: [] + experiment_run_metrics: + id: databricks_workspace.ml.experiment_run_metrics + name: experiment_run_metrics + title: Experiment Run Metrics + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1metrics~1get-history/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.metrics + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/experiment_run_metrics/methods/get + insert: [] + update: [] + delete: [] + replace: [] + experiment_permission_levels: + id: databricks_workspace.ml.experiment_permission_levels + name: experiment_permission_levels + title: Experiment Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/experiment_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + experiment_permissions: + id: databricks_workspace.ml.experiment_permissions + name: experiment_permissions + title: Experiment Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1experiments~1{experiment_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/experiment_permissions/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/experiment_permissions/methods/update + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/experiment_permissions/methods/set + experiment_run_artifacts: + id: databricks_workspace.ml.experiment_run_artifacts + name: experiment_run_artifacts + title: Experiment Run Artifacts + methods: + list: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1artifacts~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.files + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/experiment_run_artifacts/methods/list + insert: [] + update: [] + delete: [] + replace: [] + feature_materialized: + id: databricks_workspace.ml.feature_materialized + name: feature_materialized + title: Feature Materialized + methods: + batch_create: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1materialized-features:batchCreate/post + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1materialized-features/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1feature-engineering~1materialized-features/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.materialized_features + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1materialized-features~1{materialized_feature_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1materialized-features~1{materialized_feature_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1materialized-features~1{materialized_feature_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/feature_materialized/methods/get' + - $ref: '#/components/x-stackQL-resources/feature_materialized/methods/list' + insert: + - $ref: >- + #/components/x-stackQL-resources/feature_materialized/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/feature_materialized/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/feature_materialized/methods/delete + replace: [] + feature_engineering: + id: databricks_workspace.ml.feature_engineering + name: feature_engineering + title: Feature Engineering + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1feature-engineering~1features/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1feature-engineering~1features/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.features + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1features~1{full_name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1feature-engineering~1features~1{full_name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1features~1{full_name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/feature_engineering/methods/get' + - $ref: '#/components/x-stackQL-resources/feature_engineering/methods/list' + insert: + - $ref: >- + #/components/x-stackQL-resources/feature_engineering/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/feature_engineering/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/feature_engineering/methods/delete + replace: [] + feature_kafka_configs: + id: databricks_workspace.ml.feature_kafka_configs + name: feature_kafka_configs + title: Feature Kafka Configs + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1features~1kafka-configs/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1features~1kafka-configs/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.kafka_configs + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1features~1kafka-configs~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1features~1kafka-configs~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-engineering~1features~1kafka-configs~1{name}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/feature_kafka_configs/methods/get' + - $ref: >- + #/components/x-stackQL-resources/feature_kafka_configs/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/feature_kafka_configs/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/feature_kafka_configs/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/feature_kafka_configs/methods/delete + replace: [] + feature_store: + id: databricks_workspace.ml.feature_store + name: feature_store + title: Feature Store + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1feature-store~1online-stores/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1feature-store~1online-stores/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.online_stores + delete: + operation: + $ref: '#/paths/~1api~12.0~1feature-store~1online-stores~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1feature-store~1online-stores~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1feature-store~1online-stores~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + delete_table: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1online-tables~1{online_table_name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + publish_table: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1tables~1{source_table_name}~1publish/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/feature_store/methods/get' + - $ref: '#/components/x-stackQL-resources/feature_store/methods/list' + insert: + - $ref: >- + #/components/x-stackQL-resources/feature_store/methods/publish_table + - $ref: '#/components/x-stackQL-resources/feature_store/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/feature_store/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/feature_store/methods/delete' + - $ref: >- + #/components/x-stackQL-resources/feature_store/methods/delete_table + replace: [] + forecasting: + id: databricks_workspace.ml.forecasting + name: forecasting + title: Forecasting + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1automl~1create-forecasting-experiment/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1automl~1get-forecasting-experiment~1{experiment_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/forecasting/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/forecasting/methods/create' + update: [] + delete: [] + replace: [] + materialized_features: + id: databricks_workspace.ml.materialized_features + name: materialized_features + title: Materialized Features + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1feature-tables~1{table_name}~1features~1{feature_name}~1tags/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1feature-tables~1{table_name}~1features~1{feature_name}~1tags/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.feature_tags + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1feature-tables~1{table_name}~1features~1{feature_name}~1tags~1{key}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1feature-tables~1{table_name}~1features~1{feature_name}~1tags~1{key}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1feature-tables~1{table_name}~1features~1{feature_name}~1tags~1{key}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/materialized_features/methods/get' + - $ref: >- + #/components/x-stackQL-resources/materialized_features/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/materialized_features/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/materialized_features/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/materialized_features/methods/delete + replace: [] + feature_lineage: + id: databricks_workspace.ml.feature_lineage + name: feature_lineage + title: Feature Lineage + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1feature-store~1feature-tables~1{table_name}~1features~1{feature_name}~1lineage/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/feature_lineage/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + model_registry_transitions: + id: databricks_workspace.ml.model_registry_transitions + name: model_registry_transitions + title: Model Registry Transitions + methods: + approve: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1approve/post' + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1delete/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.requests + reject: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1transition-requests~1reject/post' + response: + mediaType: application/json + openAPIDocKey: '200' + transition: + operation: + $ref: >- + #/paths/~1api~12.0~1mlflow~1databricks~1model-versions~1transition-stage/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/model_registry_transitions/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/model_registry_transitions/methods/create + update: [] + delete: [] + replace: [] + model_registry_comments: + id: databricks_workspace.ml.model_registry_comments + name: model_registry_comments + title: Model Registry Comments + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1comments~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1comments~1delete/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1comments~1update/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: >- + #/components/x-stackQL-resources/model_registry_comments/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/model_registry_comments/methods/update + delete: [] + replace: [] + model_registry: + id: databricks_workspace.ml.model_registry + name: model_registry + title: Model Registry + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1delete/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + delete_tag: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1delete-tag/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1databricks~1registered-models~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.registered_models + rename: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1rename/post' + response: + mediaType: application/json + openAPIDocKey: '200' + search: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1search/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.registered_models + set_tag: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1set-tag/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registered-models~1update/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/model_registry/methods/get' + - $ref: '#/components/x-stackQL-resources/model_registry/methods/list' + - $ref: '#/components/x-stackQL-resources/model_registry/methods/search' + insert: + - $ref: '#/components/x-stackQL-resources/model_registry/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/model_registry/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/model_registry/methods/delete' + replace: [] + model_registry_model_versions: + id: databricks_workspace.ml.model_registry_model_versions + name: model_registry_model_versions + title: Model Registry Model Versions + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1delete/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + delete_tag: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1delete-tag/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get_latest: + operation: + $ref: >- + #/paths/~1api~12.0~1mlflow~1registered-models~1get-latest-versions/post + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + search: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1search/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.model_versions + set_tag: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1set-tag/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1update/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/model_registry_model_versions/methods/get + - $ref: >- + #/components/x-stackQL-resources/model_registry_model_versions/methods/search + insert: + - $ref: >- + #/components/x-stackQL-resources/model_registry_model_versions/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/model_registry_model_versions/methods/update + delete: [] + replace: [] + model_registry_webhooks: + id: databricks_workspace.ml.model_registry_webhooks + name: model_registry_webhooks + title: Model Registry Webhooks + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1delete/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.webhooks + test: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1test/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1registry-webhooks~1update/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/model_registry_webhooks/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/model_registry_webhooks/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/model_registry_webhooks/methods/update + delete: [] + replace: [] + model_registry_model_version_uri: + id: databricks_workspace.ml.model_registry_model_version_uri + name: model_registry_model_version_uri + title: Model Registry Model Version Uri + methods: + get_download_uri: + operation: + $ref: '#/paths/~1api~12.0~1mlflow~1model-versions~1get-download-uri/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/model_registry_model_version_uri/methods/get_download_uri + insert: [] + update: [] + delete: [] + replace: [] + model_registry_permission_levels: + id: databricks_workspace.ml.model_registry_permission_levels + name: model_registry_permission_levels + title: Model Registry Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/model_registry_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + model_registry_permissions: + id: databricks_workspace.ml.model_registry_permissions + name: model_registry_permissions + title: Model Registry Permissions + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1registered-models~1{registered_model_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/model_registry_permissions/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/model_registry_permissions/methods/update + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/model_registry_permissions/methods/set diff --git a/providers/src/databricks_workspace/v00.00.00000/services/oauth2.yaml b/providers/src/databricks_workspace/v00.00.00000/services/oauth2.yaml new file mode 100644 index 00000000..e64be86d --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/oauth2.yaml @@ -0,0 +1,559 @@ +openapi: 3.0.0 +info: + title: Databricks Oauth2 API (workspace) + description: >- + OpenAPI specification for the Databricks oauth2 service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.oauth2 +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/accounts/servicePrincipals/{service_principal_id}/credentials/secrets: + post: + operationId: service_principal_secrets_proxy_create + summary: Create a secret for the given service principal. + tags: + - oauth2 + - service_principal_secrets_proxy + description: |- + Create a secret for the given service principal. + + :param service_principal_id: str + The service principal ID. + :param lifetime: str (optional) + The lifetime of the secret in seconds. If this parameter is not provided, the secret will have a + default lifetime of 730 days (63072000s). + + :returns: :class:`CreateServicePrincipalSecretResponse` + parameters: + - name: service_principal_id + in: path + required: true + schema: + type: string + description: The service principal ID. + requestBody: + content: + application/json: + schema: + type: object + properties: + lifetime: + type: string + description: >- + The lifetime of the secret in seconds. If this parameter is + not provided, the secret will have a default lifetime of 730 + days (63072000s). + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateServicePrincipalSecretResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalSecretsProxyAPI + get: + operationId: service_principal_secrets_proxy_list + summary: >- + List all secrets associated with the given service principal. This + operation only returns information + tags: + - oauth2 + - service_principal_secrets_proxy + description: >- + List all secrets associated with the given service principal. This + operation only returns information + + about the secrets themselves and does not include the secret values. + + + :param service_principal_id: str + The service principal ID. + :param page_size: int (optional) + + :param page_token: str (optional) + An opaque page token which was the `next_page_token` in the response of the previous request to list + the secrets for this service principal. Provide this token to retrieve the next page of secret + entries. When providing a `page_token`, all other parameters provided to the request must match the + previous request. To list all of the secrets for a service principal, it is necessary to continue + requesting pages of entries until the response contains no `next_page_token`. Note that the number + of entries returned must not be used to determine when the listing is complete. + + :returns: Iterator over :class:`SecretInfo` + parameters: + - name: service_principal_id + in: path + required: true + schema: + type: string + description: The service principal ID. + - name: page_size + in: query + required: false + schema: + type: string + description: >- + :param page_token: str (optional) An opaque page token which was the + `next_page_token` in the response of the previous request to list + the secrets for this service principal. Provide this token to + retrieve the next page of secret entries. When providing a + `page_token`, all other parameters provided to the request must + match the previous request. To list all of the secrets for a service + principal, it is necessary to continue requesting pages of entries + until the response contains no `next_page_token`. Note that the + number of entries returned must not be used to determine when the + listing is complete. + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListServicePrincipalSecretsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalSecretsProxyAPI + /api/2.0/accounts/servicePrincipals/{service_principal_id}/credentials/secrets/{secret_id}: + delete: + operationId: service_principal_secrets_proxy_delete + summary: Delete a secret from the given service principal. + tags: + - oauth2 + - service_principal_secrets_proxy + description: |- + Delete a secret from the given service principal. + + :param service_principal_id: str + The service principal ID. + :param secret_id: str + The secret ID. + parameters: + - name: service_principal_id + in: path + required: true + schema: + type: string + description: The service principal ID. + - name: secret_id + in: path + required: true + schema: + type: string + description: The secret ID. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServicePrincipalSecretsProxyAPI +components: + schemas: + CreateCustomAppIntegrationOutput: + type: object + properties: + client_id: + type: string + client_secret: + type: string + description: >- + OAuth client-secret generated by the Databricks. If this is a + confidential OAuth app client-secret will be generated. + integration_id: + type: string + description: Unique integration id for the custom OAuth app + CreatePublishedAppIntegrationOutput: + type: object + properties: + integration_id: + type: string + CreateServicePrincipalSecretResponse: + type: object + properties: + create_time: + type: string + expire_time: + type: string + description: >- + UTC time when the secret will expire. If the field is not present, + the secret does not expire. + id: + type: string + description: ID of the secret + secret: + type: string + description: Secret Value + secret_hash: + type: string + description: Secret Hash + status: + type: string + description: Status of the secret + update_time: + type: string + description: UTC time when the secret was updated + DeleteCustomAppIntegrationOutput: + type: object + properties: {} + DeletePublishedAppIntegrationOutput: + type: object + properties: {} + FederationPolicy: + type: object + properties: + create_time: + type: string + description: + type: string + description: Description of the federation policy. + name: + type: string + description: >- + Resource name for the federation policy. Example values include + `accounts//federationPolicies/my-federation-policy` for + Account Federation Policies, and + `accounts//servicePrincipals//federationPolicies/my-federation-policy` + for Service Principal Federation Policies. Typically an output + parameter, which does not need to be specified in create or update + requests. If specified in a request, must match the value in the + request URL. + oidc_policy: + $ref: '#/components/schemas/OidcFederationPolicy' + policy_id: + type: string + description: The ID of the federation policy. Output only. + service_principal_id: + type: integer + description: >- + The service principal ID that this federation policy applies to. + Output only. Only set for service principal federation policies. + uid: + type: string + description: Unique, immutable id of the federation policy. + update_time: + type: string + description: Last update time of the federation policy. + GetCustomAppIntegrationOutput: + type: object + properties: + client_id: + type: string + confidential: + type: boolean + description: >- + This field indicates whether an OAuth client secret is required to + authenticate this client. + create_time: + type: string + created_by: + type: integer + creator_username: + type: string + integration_id: + type: string + description: ID of this custom app + name: + type: string + description: The display name of the custom OAuth app + redirect_urls: + type: array + items: + type: string + description: List of OAuth redirect urls + scopes: + type: array + items: + type: string + token_access_policy: + $ref: '#/components/schemas/TokenAccessPolicy' + description: Token access policy + user_authorized_scopes: + type: array + items: + type: string + description: >- + Scopes that will need to be consented by end user to mint the access + token. If the user does not authorize the access token will not be + minted. Must be a subset of scopes. + GetCustomAppIntegrationsOutput: + type: object + properties: + apps: + type: array + items: + $ref: '#/components/schemas/GetCustomAppIntegrationOutput' + next_page_token: + type: string + GetPublishedAppIntegrationOutput: + type: object + properties: + app_id: + type: string + create_time: + type: string + created_by: + type: integer + integration_id: + type: string + description: Unique integration id for the published OAuth app + name: + type: string + description: Display name of the published OAuth app + token_access_policy: + $ref: '#/components/schemas/TokenAccessPolicy' + description: Token access policy + GetPublishedAppIntegrationsOutput: + type: object + properties: + apps: + type: array + items: + $ref: '#/components/schemas/GetPublishedAppIntegrationOutput' + next_page_token: + type: string + GetPublishedAppsOutput: + type: object + properties: + apps: + type: array + items: + $ref: '#/components/schemas/PublishedAppOutput' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If not + present, there are no more results to show. + ListFederationPoliciesResponse: + type: object + properties: + next_page_token: + type: string + policies: + type: array + items: + $ref: '#/components/schemas/FederationPolicy' + ListServicePrincipalSecretsResponse: + type: object + properties: + next_page_token: + type: string + secrets: + type: array + items: + $ref: '#/components/schemas/SecretInfo' + description: List of the secrets + OidcFederationPolicy: + type: object + properties: + audiences: + type: array + items: + type: string + description: >- + The allowed token audiences, as specified in the 'aud' claim of + federated tokens. The audience identifier is intended to represent + the recipient of the token. Can be any non-empty string value. As + long as the audience in the token matches at least one audience in + the policy, the token is considered a match. If audiences is + unspecified, defaults to your Databricks account id. + issuer: + type: string + description: >- + The required token issuer, as specified in the 'iss' claim of + federated tokens. + jwks_json: + type: string + description: >- + The public keys used to validate the signature of federated tokens, + in JWKS format. Most use cases should not need to specify this + field. If jwks_uri and jwks_json are both unspecified (recommended), + Databricks automatically fetches the public keys from your issuer’s + well known endpoint. Databricks strongly recommends relying on your + issuer’s well known endpoint for discovering public keys. + jwks_uri: + type: string + description: >- + URL of the public keys used to validate the signature of federated + tokens, in JWKS format. Most use cases should not need to specify + this field. If jwks_uri and jwks_json are both unspecified + (recommended), Databricks automatically fetches the public keys from + your issuer’s well known endpoint. Databricks strongly recommends + relying on your issuer’s well known endpoint for discovering public + keys. + subject: + type: string + description: >- + The required token subject, as specified in the subject claim of + federated tokens. Must be specified for service principal federation + policies. Must not be specified for account federation policies. + subject_claim: + type: string + description: >- + The claim that contains the subject of the token. If unspecified, + the default value is 'sub'. + description: >- + Specifies the policy to use for validating OIDC claims in your federated + tokens. + PublishedAppOutput: + type: object + properties: + app_id: + type: string + client_id: + type: string + description: >- + Client ID of the published OAuth app. It is the client_id in the + OAuth flow + description: + type: string + description: Description of the published OAuth app. + is_confidential_client: + type: boolean + description: >- + Whether the published OAuth app is a confidential client. It is + always false for published OAuth apps. + name: + type: string + description: The display name of the published OAuth app. + redirect_urls: + type: array + items: + type: string + description: Redirect URLs of the published OAuth app. + scopes: + type: array + items: + type: string + description: Required scopes for the published OAuth app. + SecretInfo: + type: object + properties: + create_time: + type: string + expire_time: + type: string + description: >- + UTC time when the secret will expire. If the field is not present, + the secret does not expire. + id: + type: string + description: ID of the secret + secret_hash: + type: string + description: Secret Hash + status: + type: string + description: Status of the secret + update_time: + type: string + description: UTC time when the secret was updated + TokenAccessPolicy: + type: object + properties: + absolute_session_lifetime_in_minutes: + type: integer + access_token_ttl_in_minutes: + type: integer + description: access token time to live in minutes + enable_single_use_refresh_tokens: + type: boolean + description: >- + Whether to enable single-use refresh tokens (refresh token + rotation). If this feature is enabled, upon successfully getting a + new access token using a refresh token, Databricks will issue a new + refresh token along with the access token in the response and + invalidate the old refresh token. The client should use the new + refresh token to get access tokens in future requests. + refresh_token_ttl_in_minutes: + type: integer + description: >- + Refresh token time to live in minutes. When single-use refresh + tokens are enabled, this represents the TTL of an individual refresh + token. If the refresh token is used before it expires, a new one is + issued with a renewed individual TTL. + UpdateCustomAppIntegrationOutput: + type: object + properties: {} + UpdatePublishedAppIntegrationOutput: + type: object + properties: {} + x-stackQL-resources: + service_principal_secrets: + id: databricks_workspace.oauth2.service_principal_secrets + name: service_principal_secrets + title: Service Principal Secrets + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1servicePrincipals~1{service_principal_id}~1credentials~1secrets/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1servicePrincipals~1{service_principal_id}~1credentials~1secrets/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.secrets + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1servicePrincipals~1{service_principal_id}~1credentials~1secrets~1{secret_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/service_principal_secrets/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/service_principal_secrets/methods/create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/service_principal_secrets/methods/delete + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/pipelines.yaml b/providers/src/databricks_workspace/v00.00.00000/services/pipelines.yaml new file mode 100644 index 00000000..e9d39136 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/pipelines.yaml @@ -0,0 +1,3207 @@ +openapi: 3.0.0 +info: + title: Databricks Pipelines API (workspace) + description: >- + OpenAPI specification for the Databricks pipelines service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.pipelines +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/pipelines/{pipeline_id}/clone: + post: + operationId: pipelines_clone + summary: >- + Creates a new pipeline using Unity Catalog from a pipeline using Hive + Metastore. This method returns + tags: + - pipelines + description: >- + Creates a new pipeline using Unity Catalog from a pipeline using Hive + Metastore. This method returns + + the ID of the newly created clone. Additionally, this method starts an + update for the newly created + + pipeline. + + + :param pipeline_id: str + Source pipeline to clone from + :param allow_duplicate_names: bool (optional) + If false, deployment will fail if name conflicts with that of another pipeline. + :param budget_policy_id: str (optional) + Budget policy of this pipeline. + :param catalog: str (optional) + A catalog in Unity Catalog to publish data from this pipeline to. If `target` is specified, tables + in this pipeline are published to a `target` schema inside `catalog` (for example, + `catalog`.`target`.`table`). If `target` is not specified, no data is published to Unity Catalog. + :param channel: str (optional) + DLT Release Channel that specifies which version to use. + :param clone_mode: :class:`CloneMode` (optional) + The type of clone to perform. Currently, only deep copies are supported + :param clusters: List[:class:`PipelineCluster`] (optional) + Cluster settings for this pipeline deployment. + :param configuration: Dict[str,str] (optional) + String-String configuration for this pipeline execution. + :param continuous: bool (optional) + Whether the pipeline is continuous or triggered. This replaces `trigger`. + :param deployment: :class:`PipelineDeployment` (optional) + Deployment type of this pipeline. + :param development: bool (optional) + Whether the pipeline is in Development mode. Defaults to false. + :param edition: str (optional) + Pipeline product edition. + :param environment: :class:`PipelinesEnvironment` (optional) + Environment specification for this pipeline used to install dependencies. + :param event_log: :class:`EventLogSpec` (optional) + Event log configuration for this pipeline + :param expected_last_modified: int (optional) + If present, the last-modified time of the pipeline settings before the clone. If the settings were + modified after that time, then the request will fail with a conflict. + :param filters: :class:`Filters` (optional) + Filters on which Pipeline packages to include in the deployed graph. + :param gateway_definition: :class:`IngestionGatewayPipelineDefinition` + (optional) + The definition of a gateway pipeline to support change data capture. + :param id: str (optional) + Unique identifier for this pipeline. + :param ingestion_definition: :class:`IngestionPipelineDefinition` + (optional) + The configuration for a managed ingestion pipeline. These settings cannot be used with the + 'libraries', 'schema', 'target', or 'catalog' settings. + :param libraries: List[:class:`PipelineLibrary`] (optional) + Libraries or code needed by this deployment. + :param name: str (optional) + Friendly identifier for this pipeline. + :param notifications: List[:class:`Notifications`] (optional) + List of notification settings for this pipeline. + :param photon: bool (optional) + Whether Photon is enabled for this pipeline. + :param restart_window: :class:`RestartWindow` (optional) + Restart window of this pipeline. + :param root_path: str (optional) + Root path for this pipeline. This is used as the root directory when editing the pipeline in the + Databricks user interface and it is added to sys.path when executing Python sources during pipeline + execution. + :param schema: str (optional) + The default schema (database) where tables are read from or published to. + :param serverless: bool (optional) + Whether serverless compute is enabled for this pipeline. + :param storage: str (optional) + DBFS root directory for storing checkpoints and tables. + :param tags: Dict[str,str] (optional) + A map of tags associated with the pipeline. These are forwarded to the cluster as cluster tags, and + are therefore subject to the same limitations. A maximum of 25 tags can be added to the pipeline. + :param target: str (optional) + Target schema (database) to add tables in this pipeline to. Exactly one of `schema` or `target` must + be specified. To publish to Unity Catalog, also specify `catalog`. This legacy field is deprecated + for pipeline creation in favor of the `schema` field. + :param trigger: :class:`PipelineTrigger` (optional) + Which pipeline trigger to use. Deprecated: Use `continuous` instead. + :param usage_policy_id: str (optional) + Usage policy of this pipeline. + + :returns: :class:`ClonePipelineResponse` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: Source pipeline to clone from + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_duplicate_names: + type: string + description: >- + If false, deployment will fail if name conflicts with that + of another pipeline. + budget_policy_id: + type: string + description: Budget policy of this pipeline. + catalog: + type: string + description: >- + A catalog in Unity Catalog to publish data from this + pipeline to. If `target` is specified, tables in this + pipeline are published to a `target` schema inside `catalog` + (for example, `catalog`.`target`.`table`). If `target` is + not specified, no data is published to Unity Catalog. + channel: + type: string + description: DLT Release Channel that specifies which version to use. + clone_mode: + type: string + description: >- + The type of clone to perform. Currently, only deep copies + are supported + clusters: + type: string + description: Cluster settings for this pipeline deployment. + configuration: + type: string + description: String-String configuration for this pipeline execution. + continuous: + type: string + description: >- + Whether the pipeline is continuous or triggered. This + replaces `trigger`. + deployment: + type: string + description: Deployment type of this pipeline. + development: + type: string + description: >- + Whether the pipeline is in Development mode. Defaults to + false. + edition: + type: string + description: Pipeline product edition. + environment: + type: string + description: >- + Environment specification for this pipeline used to install + dependencies. + event_log: + type: string + description: Event log configuration for this pipeline + expected_last_modified: + type: string + description: >- + If present, the last-modified time of the pipeline settings + before the clone. If the settings were modified after that + time, then the request will fail with a conflict. + filters: + type: string + description: >- + Filters on which Pipeline packages to include in the + deployed graph. + gateway_definition: + type: string + description: >- + The definition of a gateway pipeline to support change data + capture. + id: + type: string + description: Unique identifier for this pipeline. + ingestion_definition: + type: string + description: >- + The configuration for a managed ingestion pipeline. These + settings cannot be used with the 'libraries', 'schema', + 'target', or 'catalog' settings. + libraries: + type: string + description: Libraries or code needed by this deployment. + name: + type: string + description: Friendly identifier for this pipeline. + notifications: + type: string + description: List of notification settings for this pipeline. + photon: + type: string + description: Whether Photon is enabled for this pipeline. + restart_window: + type: string + description: Restart window of this pipeline. + root_path: + type: string + description: >- + Root path for this pipeline. This is used as the root + directory when editing the pipeline in the Databricks user + interface and it is added to sys.path when executing Python + sources during pipeline execution. + schema: + type: string + description: >- + The default schema (database) where tables are read from or + published to. + serverless: + type: string + description: Whether serverless compute is enabled for this pipeline. + storage: + type: string + description: DBFS root directory for storing checkpoints and tables. + tags: + type: string + description: >- + A map of tags associated with the pipeline. These are + forwarded to the cluster as cluster tags, and are therefore + subject to the same limitations. A maximum of 25 tags can be + added to the pipeline. + target: + type: string + description: >- + Target schema (database) to add tables in this pipeline to. + Exactly one of `schema` or `target` must be specified. To + publish to Unity Catalog, also specify `catalog`. This + legacy field is deprecated for pipeline creation in favor of + the `schema` field. + trigger: + type: string + description: >- + Which pipeline trigger to use. Deprecated: Use `continuous` + instead. + usage_policy_id: + type: string + description: Usage policy of this pipeline. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClonePipelineResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/pipelines: + post: + operationId: pipelines_create + summary: >- + Creates a new data processing pipeline based on the requested + configuration. If successful, this + tags: + - pipelines + description: >- + Creates a new data processing pipeline based on the requested + configuration. If successful, this + + method returns the ID of the new pipeline. + + + :param allow_duplicate_names: bool (optional) + If false, deployment will fail if name conflicts with that of another pipeline. + :param budget_policy_id: str (optional) + Budget policy of this pipeline. + :param catalog: str (optional) + A catalog in Unity Catalog to publish data from this pipeline to. If `target` is specified, tables + in this pipeline are published to a `target` schema inside `catalog` (for example, + `catalog`.`target`.`table`). If `target` is not specified, no data is published to Unity Catalog. + :param channel: str (optional) + DLT Release Channel that specifies which version to use. + :param clusters: List[:class:`PipelineCluster`] (optional) + Cluster settings for this pipeline deployment. + :param configuration: Dict[str,str] (optional) + String-String configuration for this pipeline execution. + :param continuous: bool (optional) + Whether the pipeline is continuous or triggered. This replaces `trigger`. + :param deployment: :class:`PipelineDeployment` (optional) + Deployment type of this pipeline. + :param development: bool (optional) + Whether the pipeline is in Development mode. Defaults to false. + :param dry_run: bool (optional) + + :param edition: str (optional) + Pipeline product edition. + :param environment: :class:`PipelinesEnvironment` (optional) + Environment specification for this pipeline used to install dependencies. + :param event_log: :class:`EventLogSpec` (optional) + Event log configuration for this pipeline + :param filters: :class:`Filters` (optional) + Filters on which Pipeline packages to include in the deployed graph. + :param gateway_definition: :class:`IngestionGatewayPipelineDefinition` + (optional) + The definition of a gateway pipeline to support change data capture. + :param id: str (optional) + Unique identifier for this pipeline. + :param ingestion_definition: :class:`IngestionPipelineDefinition` + (optional) + The configuration for a managed ingestion pipeline. These settings cannot be used with the + 'libraries', 'schema', 'target', or 'catalog' settings. + :param libraries: List[:class:`PipelineLibrary`] (optional) + Libraries or code needed by this deployment. + :param name: str (optional) + Friendly identifier for this pipeline. + :param notifications: List[:class:`Notifications`] (optional) + List of notification settings for this pipeline. + :param photon: bool (optional) + Whether Photon is enabled for this pipeline. + :param restart_window: :class:`RestartWindow` (optional) + Restart window of this pipeline. + :param root_path: str (optional) + Root path for this pipeline. This is used as the root directory when editing the pipeline in the + Databricks user interface and it is added to sys.path when executing Python sources during pipeline + execution. + :param run_as: :class:`RunAs` (optional) + + :param schema: str (optional) + The default schema (database) where tables are read from or published to. + :param serverless: bool (optional) + Whether serverless compute is enabled for this pipeline. + :param storage: str (optional) + DBFS root directory for storing checkpoints and tables. + :param tags: Dict[str,str] (optional) + A map of tags associated with the pipeline. These are forwarded to the cluster as cluster tags, and + are therefore subject to the same limitations. A maximum of 25 tags can be added to the pipeline. + :param target: str (optional) + Target schema (database) to add tables in this pipeline to. Exactly one of `schema` or `target` must + be specified. To publish to Unity Catalog, also specify `catalog`. This legacy field is deprecated + for pipeline creation in favor of the `schema` field. + :param trigger: :class:`PipelineTrigger` (optional) + Which pipeline trigger to use. Deprecated: Use `continuous` instead. + :param usage_policy_id: str (optional) + Usage policy of this pipeline. + + :returns: :class:`CreatePipelineResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_duplicate_names: + type: string + description: >- + If false, deployment will fail if name conflicts with that + of another pipeline. + budget_policy_id: + type: string + description: Budget policy of this pipeline. + catalog: + type: string + description: >- + A catalog in Unity Catalog to publish data from this + pipeline to. If `target` is specified, tables in this + pipeline are published to a `target` schema inside `catalog` + (for example, `catalog`.`target`.`table`). If `target` is + not specified, no data is published to Unity Catalog. + channel: + type: string + description: DLT Release Channel that specifies which version to use. + clusters: + type: string + description: Cluster settings for this pipeline deployment. + configuration: + type: string + description: String-String configuration for this pipeline execution. + continuous: + type: string + description: >- + Whether the pipeline is continuous or triggered. This + replaces `trigger`. + deployment: + type: string + description: Deployment type of this pipeline. + development: + type: string + description: >- + Whether the pipeline is in Development mode. Defaults to + false. + dry_run: + type: string + description: ':param edition: str (optional) Pipeline product edition.' + edition: + type: string + environment: + type: string + description: >- + Environment specification for this pipeline used to install + dependencies. + event_log: + type: string + description: Event log configuration for this pipeline + filters: + type: string + description: >- + Filters on which Pipeline packages to include in the + deployed graph. + gateway_definition: + type: string + description: >- + The definition of a gateway pipeline to support change data + capture. + id: + type: string + description: Unique identifier for this pipeline. + ingestion_definition: + type: string + description: >- + The configuration for a managed ingestion pipeline. These + settings cannot be used with the 'libraries', 'schema', + 'target', or 'catalog' settings. + libraries: + type: string + description: Libraries or code needed by this deployment. + name: + type: string + description: Friendly identifier for this pipeline. + notifications: + type: string + description: List of notification settings for this pipeline. + photon: + type: string + description: Whether Photon is enabled for this pipeline. + restart_window: + type: string + description: Restart window of this pipeline. + root_path: + type: string + description: >- + Root path for this pipeline. This is used as the root + directory when editing the pipeline in the Databricks user + interface and it is added to sys.path when executing Python + sources during pipeline execution. + run_as: + type: string + description: >- + :param schema: str (optional) The default schema (database) + where tables are read from or published to. + schema: + type: string + serverless: + type: string + description: Whether serverless compute is enabled for this pipeline. + storage: + type: string + description: DBFS root directory for storing checkpoints and tables. + tags: + type: string + description: >- + A map of tags associated with the pipeline. These are + forwarded to the cluster as cluster tags, and are therefore + subject to the same limitations. A maximum of 25 tags can be + added to the pipeline. + target: + type: string + description: >- + Target schema (database) to add tables in this pipeline to. + Exactly one of `schema` or `target` must be specified. To + publish to Unity Catalog, also specify `catalog`. This + legacy field is deprecated for pipeline creation in favor of + the `schema` field. + trigger: + type: string + description: >- + Which pipeline trigger to use. Deprecated: Use `continuous` + instead. + usage_policy_id: + type: string + description: Usage policy of this pipeline. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreatePipelineResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + get: + operationId: pipelines_list_pipelines + summary: Lists pipelines defined in the Spark Declarative Pipelines system. + tags: + - pipelines + description: |- + Lists pipelines defined in the Spark Declarative Pipelines system. + + :param filter: str (optional) + Select a subset of results based on the specified criteria. The supported filters are: + + * `notebook=''` to select pipelines that reference the provided notebook path. * `name LIKE + '[pattern]'` to select pipelines with a name that matches pattern. Wildcards are supported, for + example: `name LIKE '%shopping%'` + + Composite filters are not supported. This field is optional. + :param max_results: int (optional) + The maximum number of entries to return in a single page. The system may return fewer than + max_results events in a response, even if there are more events available. This field is optional. + The default value is 25. The maximum value is 100. An error is returned if the value of max_results + is greater than 100. + :param order_by: List[str] (optional) + A list of strings specifying the order of results. Supported order_by fields are id and name. The + default is id asc. This field is optional. + :param page_token: str (optional) + Page token returned by previous call + + :returns: Iterator over :class:`PipelineStateInfo` + parameters: + - name: filter + in: query + required: false + schema: + type: string + description: >- + Select a subset of results based on the specified criteria. The + supported filters are: * `notebook=''` to select pipelines + that reference the provided notebook path. * `name LIKE '[pattern]'` + to select pipelines with a name that matches pattern. Wildcards are + supported, for example: `name LIKE '%shopping%'` Composite filters + are not supported. This field is optional. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + The maximum number of entries to return in a single page. The system + may return fewer than max_results events in a response, even if + there are more events available. This field is optional. The default + value is 25. The maximum value is 100. An error is returned if the + value of max_results is greater than 100. + - name: order_by + in: query + required: false + schema: + type: string + description: >- + A list of strings specifying the order of results. Supported + order_by fields are id and name. The default is id asc. This field + is optional. + - name: page_token + in: query + required: false + schema: + type: string + description: Page token returned by previous call + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListPipelinesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/pipelines/{pipeline_id}: + delete: + operationId: pipelines_delete + summary: >- + Deletes a pipeline. If the pipeline publishes to Unity Catalog, pipeline + deletion will cascade to all + tags: + - pipelines + description: >- + Deletes a pipeline. If the pipeline publishes to Unity Catalog, pipeline + deletion will cascade to all + + pipeline tables. Please reach out to Databricks support for assistance + to undo this action. + + + :param pipeline_id: str + + :param force: bool (optional) + If true, deletion will proceed even if resource cleanup fails. By default, deletion will fail if + resources cleanup is required but fails. + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: >- + :param force: bool (optional) If true, deletion will proceed even if + resource cleanup fails. By default, deletion will fail if resources + cleanup is required but fails. + - name: force + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + get: + operationId: pipelines_get + summary: Get a pipeline. + tags: + - pipelines + description: |- + Get a pipeline. + + :param pipeline_id: str + + :returns: :class:`GetPipelineResponse` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`GetPipelineResponse`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPipelineResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + put: + operationId: pipelines_update + summary: Updates a pipeline with the supplied configuration. + tags: + - pipelines + description: >- + Updates a pipeline with the supplied configuration. + + + :param pipeline_id: str + Unique identifier for this pipeline. + :param allow_duplicate_names: bool (optional) + If false, deployment will fail if name has changed and conflicts the name of another pipeline. + :param budget_policy_id: str (optional) + Budget policy of this pipeline. + :param catalog: str (optional) + A catalog in Unity Catalog to publish data from this pipeline to. If `target` is specified, tables + in this pipeline are published to a `target` schema inside `catalog` (for example, + `catalog`.`target`.`table`). If `target` is not specified, no data is published to Unity Catalog. + :param channel: str (optional) + DLT Release Channel that specifies which version to use. + :param clusters: List[:class:`PipelineCluster`] (optional) + Cluster settings for this pipeline deployment. + :param configuration: Dict[str,str] (optional) + String-String configuration for this pipeline execution. + :param continuous: bool (optional) + Whether the pipeline is continuous or triggered. This replaces `trigger`. + :param deployment: :class:`PipelineDeployment` (optional) + Deployment type of this pipeline. + :param development: bool (optional) + Whether the pipeline is in Development mode. Defaults to false. + :param edition: str (optional) + Pipeline product edition. + :param environment: :class:`PipelinesEnvironment` (optional) + Environment specification for this pipeline used to install dependencies. + :param event_log: :class:`EventLogSpec` (optional) + Event log configuration for this pipeline + :param expected_last_modified: int (optional) + If present, the last-modified time of the pipeline settings before the edit. If the settings were + modified after that time, then the request will fail with a conflict. + :param filters: :class:`Filters` (optional) + Filters on which Pipeline packages to include in the deployed graph. + :param gateway_definition: :class:`IngestionGatewayPipelineDefinition` + (optional) + The definition of a gateway pipeline to support change data capture. + :param id: str (optional) + Unique identifier for this pipeline. + :param ingestion_definition: :class:`IngestionPipelineDefinition` + (optional) + The configuration for a managed ingestion pipeline. These settings cannot be used with the + 'libraries', 'schema', 'target', or 'catalog' settings. + :param libraries: List[:class:`PipelineLibrary`] (optional) + Libraries or code needed by this deployment. + :param name: str (optional) + Friendly identifier for this pipeline. + :param notifications: List[:class:`Notifications`] (optional) + List of notification settings for this pipeline. + :param photon: bool (optional) + Whether Photon is enabled for this pipeline. + :param restart_window: :class:`RestartWindow` (optional) + Restart window of this pipeline. + :param root_path: str (optional) + Root path for this pipeline. This is used as the root directory when editing the pipeline in the + Databricks user interface and it is added to sys.path when executing Python sources during pipeline + execution. + :param run_as: :class:`RunAs` (optional) + + :param schema: str (optional) + The default schema (database) where tables are read from or published to. + :param serverless: bool (optional) + Whether serverless compute is enabled for this pipeline. + :param storage: str (optional) + DBFS root directory for storing checkpoints and tables. + :param tags: Dict[str,str] (optional) + A map of tags associated with the pipeline. These are forwarded to the cluster as cluster tags, and + are therefore subject to the same limitations. A maximum of 25 tags can be added to the pipeline. + :param target: str (optional) + Target schema (database) to add tables in this pipeline to. Exactly one of `schema` or `target` must + be specified. To publish to Unity Catalog, also specify `catalog`. This legacy field is deprecated + for pipeline creation in favor of the `schema` field. + :param trigger: :class:`PipelineTrigger` (optional) + Which pipeline trigger to use. Deprecated: Use `continuous` instead. + :param usage_policy_id: str (optional) + Usage policy of this pipeline. + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: Unique identifier for this pipeline. + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_duplicate_names: + type: string + description: >- + If false, deployment will fail if name has changed and + conflicts the name of another pipeline. + budget_policy_id: + type: string + description: Budget policy of this pipeline. + catalog: + type: string + description: >- + A catalog in Unity Catalog to publish data from this + pipeline to. If `target` is specified, tables in this + pipeline are published to a `target` schema inside `catalog` + (for example, `catalog`.`target`.`table`). If `target` is + not specified, no data is published to Unity Catalog. + channel: + type: string + description: DLT Release Channel that specifies which version to use. + clusters: + type: string + description: Cluster settings for this pipeline deployment. + configuration: + type: string + description: String-String configuration for this pipeline execution. + continuous: + type: string + description: >- + Whether the pipeline is continuous or triggered. This + replaces `trigger`. + deployment: + type: string + description: Deployment type of this pipeline. + development: + type: string + description: >- + Whether the pipeline is in Development mode. Defaults to + false. + edition: + type: string + description: Pipeline product edition. + environment: + type: string + description: >- + Environment specification for this pipeline used to install + dependencies. + event_log: + type: string + description: Event log configuration for this pipeline + expected_last_modified: + type: string + description: >- + If present, the last-modified time of the pipeline settings + before the edit. If the settings were modified after that + time, then the request will fail with a conflict. + filters: + type: string + description: >- + Filters on which Pipeline packages to include in the + deployed graph. + gateway_definition: + type: string + description: >- + The definition of a gateway pipeline to support change data + capture. + id: + type: string + description: Unique identifier for this pipeline. + ingestion_definition: + type: string + description: >- + The configuration for a managed ingestion pipeline. These + settings cannot be used with the 'libraries', 'schema', + 'target', or 'catalog' settings. + libraries: + type: string + description: Libraries or code needed by this deployment. + name: + type: string + description: Friendly identifier for this pipeline. + notifications: + type: string + description: List of notification settings for this pipeline. + photon: + type: string + description: Whether Photon is enabled for this pipeline. + restart_window: + type: string + description: Restart window of this pipeline. + root_path: + type: string + description: >- + Root path for this pipeline. This is used as the root + directory when editing the pipeline in the Databricks user + interface and it is added to sys.path when executing Python + sources during pipeline execution. + run_as: + type: string + description: >- + :param schema: str (optional) The default schema (database) + where tables are read from or published to. + schema: + type: string + serverless: + type: string + description: Whether serverless compute is enabled for this pipeline. + storage: + type: string + description: DBFS root directory for storing checkpoints and tables. + tags: + type: string + description: >- + A map of tags associated with the pipeline. These are + forwarded to the cluster as cluster tags, and are therefore + subject to the same limitations. A maximum of 25 tags can be + added to the pipeline. + target: + type: string + description: >- + Target schema (database) to add tables in this pipeline to. + Exactly one of `schema` or `target` must be specified. To + publish to Unity Catalog, also specify `catalog`. This + legacy field is deprecated for pipeline creation in favor of + the `schema` field. + trigger: + type: string + description: >- + Which pipeline trigger to use. Deprecated: Use `continuous` + instead. + usage_policy_id: + type: string + description: Usage policy of this pipeline. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/permissions/pipelines/{pipeline_id}/permissionLevels: + get: + operationId: pipelines_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - pipelines + description: |- + Gets the permission levels that a user can have on an object. + + :param pipeline_id: str + The pipeline for which to get or manage permissions. + + :returns: :class:`GetPipelinePermissionLevelsResponse` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: The pipeline for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPipelinePermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/permissions/pipelines/{pipeline_id}: + get: + operationId: pipelines_get_permissions + summary: >- + Gets the permissions of a pipeline. Pipelines can inherit permissions + from their root object. + tags: + - pipelines + description: >- + Gets the permissions of a pipeline. Pipelines can inherit permissions + from their root object. + + + :param pipeline_id: str + The pipeline for which to get or manage permissions. + + :returns: :class:`PipelinePermissions` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: The pipeline for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PipelinePermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + put: + operationId: pipelines_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - pipelines + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param pipeline_id: str + The pipeline for which to get or manage permissions. + :param access_control_list: List[:class:`PipelineAccessControlRequest`] + (optional) + + + :returns: :class:`PipelinePermissions` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: The pipeline for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`PipelinePermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PipelinePermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + patch: + operationId: pipelines_update_permissions + summary: >- + Updates the permissions on a pipeline. Pipelines can inherit permissions + from their root object. + tags: + - pipelines + description: >- + Updates the permissions on a pipeline. Pipelines can inherit permissions + from their root object. + + + :param pipeline_id: str + The pipeline for which to get or manage permissions. + :param access_control_list: List[:class:`PipelineAccessControlRequest`] + (optional) + + + :returns: :class:`PipelinePermissions` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: The pipeline for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`PipelinePermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PipelinePermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/pipelines/{pipeline_id}/updates/{update_id}: + get: + operationId: pipelines_get_update + summary: Gets an update from an active pipeline. + tags: + - pipelines + description: |- + Gets an update from an active pipeline. + + :param pipeline_id: str + The ID of the pipeline. + :param update_id: str + The ID of the update. + + :returns: :class:`GetUpdateResponse` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: The ID of the pipeline. + - name: update_id + in: path + required: true + schema: + type: string + description: The ID of the update. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetUpdateResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/pipelines/{pipeline_id}/events: + get: + operationId: pipelines_list_pipeline_events + summary: Retrieves events for a pipeline. + tags: + - pipelines + description: |- + Retrieves events for a pipeline. + + :param pipeline_id: str + The pipeline to return events for. + :param filter: str (optional) + Criteria to select a subset of results, expressed using a SQL-like syntax. The supported filters + are: 1. level='INFO' (or WARN or ERROR) 2. level in ('INFO', 'WARN') 3. id='[event-id]' 4. timestamp + > 'TIMESTAMP' (or >=,<,<=,=) + + Composite expressions are supported, for example: level in ('ERROR', 'WARN') AND timestamp> + '2021-07-22T06:37:33.083Z' + :param max_results: int (optional) + Max number of entries to return in a single page. The system may return fewer than max_results + events in a response, even if there are more events available. + :param order_by: List[str] (optional) + A string indicating a sort order by timestamp for the results, for example, ["timestamp asc"]. The + sort order can be ascending or descending. By default, events are returned in descending order by + timestamp. + :param page_token: str (optional) + Page token returned by previous call. This field is mutually exclusive with all fields in this + request except max_results. An error is returned if any fields other than max_results are set when + this field is set. + + :returns: Iterator over :class:`PipelineEvent` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: The pipeline to return events for. + - name: filter + in: query + required: false + schema: + type: string + description: >- + Criteria to select a subset of results, expressed using a SQL-like + syntax. The supported filters are: 1. level='INFO' (or WARN or + ERROR) 2. level in ('INFO', 'WARN') 3. id='[event-id]' 4. timestamp + > 'TIMESTAMP' (or >=,<,<=,=) Composite expressions are supported, + for example: level in ('ERROR', 'WARN') AND timestamp> + '2021-07-22T06:37:33.083Z' + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Max number of entries to return in a single page. The system may + return fewer than max_results events in a response, even if there + are more events available. + - name: order_by + in: query + required: false + schema: + type: string + description: >- + A string indicating a sort order by timestamp for the results, for + example, ["timestamp asc"]. The sort order can be ascending or + descending. By default, events are returned in descending order by + timestamp. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Page token returned by previous call. This field is mutually + exclusive with all fields in this request except max_results. An + error is returned if any fields other than max_results are set when + this field is set. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListPipelineEventsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/pipelines/{pipeline_id}/updates: + get: + operationId: pipelines_list_updates + summary: List updates for an active pipeline. + tags: + - pipelines + description: |- + List updates for an active pipeline. + + :param pipeline_id: str + The pipeline to return updates for. + :param max_results: int (optional) + Max number of entries to return in a single page. + :param page_token: str (optional) + Page token returned by previous call + :param until_update_id: str (optional) + If present, returns updates until and including this update_id. + + :returns: :class:`ListUpdatesResponse` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: The pipeline to return updates for. + - name: max_results + in: query + required: false + schema: + type: string + description: Max number of entries to return in a single page. + - name: page_token + in: query + required: false + schema: + type: string + description: Page token returned by previous call + - name: until_update_id + in: query + required: false + schema: + type: string + description: If present, returns updates until and including this update_id. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListUpdatesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + post: + operationId: pipelines_start_update + summary: >- + Starts a new update for the pipeline. If there is already an active + update for the pipeline, the + tags: + - pipelines + description: >- + Starts a new update for the pipeline. If there is already an active + update for the pipeline, the + + request will fail and the active update will remain running. + + + :param pipeline_id: str + + :param cause: :class:`StartUpdateCause` (optional) + + :param full_refresh: bool (optional) + If true, this update will reset all tables before running. + :param full_refresh_selection: List[str] (optional) + A list of tables to update with fullRefresh. If both refresh_selection and full_refresh_selection + are empty, this is a full graph update. Full Refresh on a table means that the states of the table + will be reset before the refresh. + :param refresh_selection: List[str] (optional) + A list of tables to update without fullRefresh. If both refresh_selection and full_refresh_selection + are empty, this is a full graph update. Full Refresh on a table means that the states of the table + will be reset before the refresh. + :param rewind_spec: :class:`RewindSpec` (optional) + The information about the requested rewind operation. If specified this is a rewind mode update. + :param validate_only: bool (optional) + If true, this update only validates the correctness of pipeline source code but does not materialize + or publish any datasets. + + :returns: :class:`StartUpdateResponse` + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: ':param cause: :class:`StartUpdateCause` (optional)' + requestBody: + content: + application/json: + schema: + type: object + properties: + cause: + type: string + full_refresh: + type: string + description: If true, this update will reset all tables before running. + full_refresh_selection: + type: string + description: >- + A list of tables to update with fullRefresh. If both + refresh_selection and full_refresh_selection are empty, this + is a full graph update. Full Refresh on a table means that + the states of the table will be reset before the refresh. + refresh_selection: + type: string + description: >- + A list of tables to update without fullRefresh. If both + refresh_selection and full_refresh_selection are empty, this + is a full graph update. Full Refresh on a table means that + the states of the table will be reset before the refresh. + rewind_spec: + type: string + description: >- + The information about the requested rewind operation. If + specified this is a rewind mode update. + validate_only: + type: string + description: >- + If true, this update only validates the correctness of + pipeline source code but does not materialize or publish any + datasets. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StartUpdateResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI + /api/2.0/pipelines/{pipeline_id}/stop: + post: + operationId: pipelines_stop + summary: >- + Stops the pipeline by canceling the active update. If there is no active + update for the pipeline, this + tags: + - pipelines + description: >- + Stops the pipeline by canceling the active update. If there is no active + update for the pipeline, this + + request is a no-op. + + + :param pipeline_id: str + + + :returns: + Long-running operation waiter for :class:`GetPipelineResponse`. + See :method:wait_get_pipeline_idle for more details. + parameters: + - name: pipeline_id + in: path + required: true + schema: + type: string + description: >- + :returns: Long-running operation waiter for + :class:`GetPipelineResponse`. See :method:wait_get_pipeline_idle for + more details. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetPipelineResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PipelinesAPI +components: + schemas: + AutoFullRefreshPolicy: + type: object + properties: + enabled: + type: boolean + description: (Required, Mutable) Whether to enable auto full refresh or not. + min_interval_hours: + type: integer + description: >- + (Optional, Mutable) Specify the minimum interval in hours between + the timestamp at which a table was last full refreshed and the + current timestamp for triggering auto full If unspecified and + autoFullRefresh is enabled then by default min_interval_hours is 24 + hours. + required: + - enabled + description: Policy for auto full refresh. + ClonePipelineResponse: + type: object + properties: + pipeline_id: + type: string + ConnectionParameters: + type: object + properties: + source_catalog: + type: string + CreatePipelineResponse: + type: object + properties: + effective_settings: + $ref: '#/components/schemas/PipelineSpec' + pipeline_id: + type: string + description: >- + The unique identifier for the newly created pipeline. Only returned + when dry_run is false. + CronTrigger: + type: object + properties: + quartz_cron_schedule: + type: string + timezone_id: + type: string + DataPlaneId: + type: object + properties: + instance: + type: string + seq_no: + type: integer + description: >- + A sequence number, unique and increasing within the data plane + instance. + DeletePipelineResponse: + type: object + properties: {} + EditPipelineResponse: + type: object + properties: {} + ErrorDetail: + type: object + properties: + exceptions: + type: array + items: + $ref: '#/components/schemas/SerializedException' + fatal: + type: boolean + description: Whether this error is considered fatal, that is, unrecoverable. + EventLogSpec: + type: object + properties: + catalog: + type: string + description: The UC catalog the event log is published under. + name: + type: string + description: The name the event log is published to in UC. + schema: + type: string + description: The UC schema the event log is published under. + description: Configurable event log parameters. + FileLibrary: + type: object + properties: + path: + type: string + Filters: + type: object + properties: + exclude: + type: array + items: + type: string + include: + type: array + items: + type: string + description: Paths to include. + GetPipelinePermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/PipelinePermissionsDescription' + GetPipelineResponse: + type: object + properties: + cause: + type: string + cluster_id: + type: string + description: The ID of the cluster that the pipeline is running on. + creator_user_name: + type: string + description: The username of the pipeline creator. + effective_budget_policy_id: + type: string + description: Serverless budget policy ID of this pipeline. + health: + $ref: '#/components/schemas/GetPipelineResponseHealth' + description: The health of a pipeline. + last_modified: + type: integer + description: The last time the pipeline settings were modified or created. + latest_updates: + type: array + items: + $ref: '#/components/schemas/UpdateStateInfo' + description: >- + Status of the latest updates for the pipeline. Ordered with the + newest update first. + name: + type: string + description: A human friendly identifier for the pipeline, taken from the `spec`. + pipeline_id: + type: string + description: The ID of the pipeline. + run_as: + $ref: '#/components/schemas/RunAs' + description: >- + The user or service principal that the pipeline runs as, if + specified in the request. This field indicates the explicit + configuration of `run_as` for the pipeline. To find the value in all + cases, explicit or implicit, use `run_as_user_name`. + run_as_user_name: + type: string + description: Username of the user that the pipeline will run on behalf of. + spec: + $ref: '#/components/schemas/PipelineSpec' + description: >- + The pipeline specification. This field is not returned when called + by `ListPipelines`. + state: + $ref: '#/components/schemas/PipelineState' + description: The pipeline state. + GetUpdateResponse: + type: object + properties: + update: + $ref: '#/components/schemas/UpdateInfo' + IngestionConfig: + type: object + properties: + report: + $ref: '#/components/schemas/ReportSpec' + schema: + $ref: '#/components/schemas/SchemaSpec' + description: Select all tables from a specific source schema. + table: + $ref: '#/components/schemas/TableSpec' + description: Select a specific source table. + IngestionGatewayPipelineDefinition: + type: object + properties: + connection_name: + type: string + gateway_storage_catalog: + type: string + description: >- + Required, Immutable. The name of the catalog for the gateway + pipeline's storage location. + gateway_storage_schema: + type: string + description: >- + Required, Immutable. The name of the schema for the gateway + pipelines's storage location. + connection_id: + type: string + description: >- + [Deprecated, use connection_name instead] Immutable. The Unity + Catalog connection that this gateway pipeline uses to communicate + with the source. + connection_parameters: + $ref: '#/components/schemas/ConnectionParameters' + description: >- + Optional, Internal. Parameters required to establish an initial + connection with the source. + gateway_storage_name: + type: string + description: >- + Optional. The Unity Catalog-compatible name for the gateway storage + location. This is the destination to use for the data that is + extracted by the gateway. Spark Declarative Pipelines system will + automatically create the storage location under the catalog and + schema. + required: + - connection_name + - gateway_storage_catalog + - gateway_storage_schema + IngestionPipelineDefinition: + type: object + properties: + connection_name: + type: string + full_refresh_window: + $ref: '#/components/schemas/OperationTimeWindow' + description: >- + (Optional) A window that specifies a set of time ranges for snapshot + queries in CDC. + ingest_from_uc_foreign_catalog: + type: boolean + description: >- + Immutable. If set to true, the pipeline will ingest tables from the + UC foreign catalogs directly without the need to specify a UC + connection or ingestion gateway. The `source_catalog` fields in + objects of IngestionConfig are interpreted as the UC foreign + catalogs to ingest from. + ingestion_gateway_id: + type: string + description: >- + Identifier for the gateway that is used by this ingestion pipeline + to communicate with the source database. This is used with CDC + connectors to databases like SQL Server using a gateway pipeline + (connector_type = CDC). Under certain conditions, this can be + replaced with connection_name to change the connector to Combined + Cdc Managed Ingestion Pipeline. + netsuite_jar_path: + type: string + description: >- + Netsuite only configuration. When the field is set for a netsuite + connector, the jar stored in the field will be validated and added + to the classpath of pipeline's cluster. + objects: + type: array + items: + $ref: '#/components/schemas/IngestionConfig' + description: >- + Required. Settings specifying tables to replicate and the + destination for the replicated tables. + source_configurations: + type: array + items: + $ref: '#/components/schemas/SourceConfig' + description: Top-level source configurations + source_type: + $ref: '#/components/schemas/IngestionSourceType' + description: >- + The type of the foreign source. The source type will be inferred + from the source connection or ingestion gateway. This field is + output only and will be ignored if provided. + table_configuration: + $ref: '#/components/schemas/TableSpecificConfig' + description: >- + Configuration settings to control the ingestion of tables. These + settings are applied to all tables in the pipeline. + IngestionPipelineDefinitionTableSpecificConfigQueryBasedConnectorConfig: + type: object + properties: + cursor_columns: + type: array + items: + type: string + description: >- + The names of the monotonically increasing columns in the source + table that are used to enable the table to be read and ingested + incrementally through structured streaming. The columns are allowed + to have repeated values but have to be non-decreasing. If the source + data is merged into the destination (e.g., using SCD Type 1 or Type + 2), these columns will implicitly define the `sequence_by` behavior. + You can still explicitly set `sequence_by` to override this default. + deletion_condition: + type: string + description: >- + Specifies a SQL WHERE condition that specifies that the source row + has been deleted. This is sometimes referred to as "soft-deletes". + For example: "Operation = 'DELETE'" or "is_deleted = true". This + field is orthogonal to `hard_deletion_sync_interval_in_seconds`, one + for soft-deletes and the other for hard-deletes. See also the + hard_deletion_sync_min_interval_in_seconds field for handling of + "hard deletes" where the source rows are physically removed from the + table. + hard_deletion_sync_min_interval_in_seconds: + type: integer + description: >- + Specifies the minimum interval (in seconds) between snapshots on + primary keys for detecting and synchronizing hard deletions—i.e., + rows that have been physically removed from the source table. This + interval acts as a lower bound. If ingestion runs less frequently + than this value, hard deletion synchronization will align with the + actual ingestion frequency instead of happening more often. If not + set, hard deletion synchronization via snapshots is disabled. This + field is mutable and can be updated without triggering a full + snapshot. + description: >- + Configurations that are only applicable for query-based ingestion + connectors. + IngestionPipelineDefinitionWorkdayReportParameters: + type: object + properties: + incremental: + type: boolean + parameters: + type: object + description: >- + Parameters for the Workday report. Each key represents the parameter + name (e.g., "start_date", "end_date"), and the corresponding value + is a SQL-like expression used to compute the parameter value at + runtime. Example: { "start_date": "{ coalesce(current_offset(), + date(\"2025-02-01\")) }", "end_date": "{ current_date() - INTERVAL 1 + DAY }" } + report_parameters: + type: array + items: + $ref: >- + #/components/schemas/IngestionPipelineDefinitionWorkdayReportParametersQueryKeyValue + description: >- + (Optional) Additional custom parameters for Workday Report This + field is deprecated and should not be used. Use `parameters` + instead. + IngestionPipelineDefinitionWorkdayReportParametersQueryKeyValue: + type: object + properties: + key: + type: string + value: + type: string + description: >- + Value for the report parameter. Possible values it can take are + these sql functions: 1. coalesce(current_offset(), + date("YYYY-MM-DD")) -> if current_offset() is null, then the passed + date, else current_offset() 2. current_date() 3. + date_sub(current_date(), x) -> subtract x (some non-negative + integer) days from current date + ListPipelineEventsResponse: + type: object + properties: + events: + type: array + items: + $ref: '#/components/schemas/PipelineEvent' + next_page_token: + type: string + description: If present, a token to fetch the next page of events. + prev_page_token: + type: string + description: If present, a token to fetch the previous page of events. + ListPipelinesResponse: + type: object + properties: + next_page_token: + type: string + statuses: + type: array + items: + $ref: '#/components/schemas/PipelineStateInfo' + description: The list of events matching the request criteria. + ListUpdatesResponse: + type: object + properties: + next_page_token: + type: string + prev_page_token: + type: string + description: >- + If present, then this token can be used in a subsequent request to + fetch the previous page. + updates: + type: array + items: + $ref: '#/components/schemas/UpdateInfo' + ManualTrigger: + type: object + properties: {} + NotebookLibrary: + type: object + properties: + path: + type: string + Notifications: + type: object + properties: + alerts: + type: array + items: + type: string + email_recipients: + type: array + items: + type: string + description: >- + A list of email addresses notified when a configured alert is + triggered. + OperationTimeWindow: + type: object + properties: + start_hour: + type: integer + description: >- + An integer between 0 and 23 denoting the start hour for the window + in the 24-hour day. + days_of_week: + type: array + items: + $ref: '#/components/schemas/DayOfWeek' + description: >- + Days of week in which the window is allowed to happen If not + specified all days of the week will be used. + time_zone_id: + type: string + description: >- + Time zone id of window. See + https://docs.databricks.com/sql/language-manual/sql-ref-syntax-aux-conf-mgmt-set-timezone.html + for details. If not specified, UTC will be used. + required: + - start_hour + description: Proto representing a window + Origin: + type: object + properties: + batch_id: + type: integer + cloud: + type: string + description: The cloud provider, e.g., AWS or Azure. + cluster_id: + type: string + description: >- + The id of the cluster where an execution happens. Unique within a + region. + dataset_name: + type: string + description: The name of a dataset. Unique within a pipeline. + flow_id: + type: string + description: >- + The id of the flow. Globally unique. Incremental queries will + generally reuse the same id while complete queries will have a new + id per update. + flow_name: + type: string + description: The name of the flow. Not unique. + host: + type: string + description: The optional host name where the event was triggered + maintenance_id: + type: string + description: The id of a maintenance run. Globally unique. + materialization_name: + type: string + description: Materialization name. + org_id: + type: integer + description: The org id of the user. Unique within a cloud. + pipeline_id: + type: string + description: The id of the pipeline. Globally unique. + pipeline_name: + type: string + description: The name of the pipeline. Not unique. + region: + type: string + description: The cloud region. + request_id: + type: string + description: The id of the request that caused an update. + table_id: + type: string + description: The id of a (delta) table. Globally unique. + uc_resource_id: + type: string + description: The Unity Catalog id of the MV or ST being updated. + update_id: + type: string + description: The id of an execution. Globally unique. + PathPattern: + type: object + properties: + include: + type: string + PipelineAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/PipelinePermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + PipelineAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/PipelinePermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + PipelineCluster: + type: object + properties: + apply_policy_default_values: + type: boolean + autoscale: + $ref: '#/components/schemas/PipelineClusterAutoscale' + description: >- + Parameters needed in order to automatically scale clusters up and + down based on load. Note: autoscaling works best with DB runtime + versions 3.0 or later. + aws_attributes: + type: string + description: >- + Attributes related to clusters running on Amazon Web Services. If + not specified at cluster creation, a set of default values will be + used. + azure_attributes: + type: string + description: >- + Attributes related to clusters running on Microsoft Azure. If not + specified at cluster creation, a set of default values will be used. + cluster_log_conf: + type: string + description: >- + The configuration for delivering spark logs to a long-term storage + destination. Only dbfs destinations are supported. Only one + destination can be specified for one cluster. If the conf is given, + the logs will be delivered to the destination every `5 mins`. The + destination of driver logs is `$destination/$clusterId/driver`, + while the destination of executor logs is + `$destination/$clusterId/executor`. + custom_tags: + type: object + description: >- + Additional tags for cluster resources. Databricks will tag all + cluster resources (e.g., AWS instances and EBS volumes) with these + tags in addition to `default_tags`. Notes: - Currently, Databricks + allows at most 45 custom tags - Clusters can only reuse cloud + resources if the resources' tags are a subset of the cluster tags + driver_instance_pool_id: + type: string + description: >- + The optional ID of the instance pool for the driver of the cluster + belongs. The pool cluster uses the instance pool with id + (instance_pool_id) if the driver pool is not assigned. + driver_node_type_id: + type: string + description: >- + The node type of the Spark driver. Note that this field is optional; + if unset, the driver node type will be set as the same value as + `node_type_id` defined above. + enable_local_disk_encryption: + type: boolean + description: Whether to enable local disk encryption for the cluster. + gcp_attributes: + type: string + description: >- + Attributes related to clusters running on Google Cloud Platform. If + not specified at cluster creation, a set of default values will be + used. + init_scripts: + type: string + description: >- + The configuration for storing init scripts. Any number of + destinations can be specified. The scripts are executed sequentially + in the order provided. If `cluster_log_conf` is specified, init + script logs are sent to `//init_scripts`. + instance_pool_id: + type: string + description: The optional ID of the instance pool to which the cluster belongs. + label: + type: string + description: >- + A label for the cluster specification, either `default` to configure + the default cluster, or `maintenance` to configure the maintenance + cluster. This field is optional. The default value is `default`. + node_type_id: + type: string + description: >- + This field encodes, through a single value, the resources available + to each of the Spark nodes in this cluster. For example, the Spark + nodes can be provisioned and optimized for memory or compute + intensive workloads. A list of available node types can be retrieved + by using the :method:clusters/listNodeTypes API call. + num_workers: + type: integer + description: >- + Number of worker nodes that this cluster should have. A cluster has + one Spark Driver and `num_workers` Executors for a total of + `num_workers` + 1 Spark nodes. Note: When reading the properties of + a cluster, this field reflects the desired number of workers rather + than the actual current number of workers. For instance, if a + cluster is resized from 5 to 10 workers, this field will immediately + be updated to reflect the target size of 10 workers, whereas the + workers listed in `spark_info` will gradually increase from 5 to 10 + as the new nodes are provisioned. + policy_id: + type: string + description: >- + The ID of the cluster policy used to create the cluster if + applicable. + spark_conf: + type: object + description: >- + An object containing a set of optional, user-specified Spark + configuration key-value pairs. See :method:clusters/create for more + details. + spark_env_vars: + type: object + description: >- + An object containing a set of optional, user-specified environment + variable key-value pairs. Please note that key-value pair of the + form (X,Y) will be exported as is (i.e., `export X='Y'`) while + launching the driver and workers. In order to specify an additional + set of `SPARK_DAEMON_JAVA_OPTS`, we recommend appending them to + `$SPARK_DAEMON_JAVA_OPTS` as shown in the example below. This + ensures that all default databricks managed environmental variables + are included as well. Example Spark environment variables: + `{"SPARK_WORKER_MEMORY": "28000m", "SPARK_LOCAL_DIRS": + "/local_disk0"}` or `{"SPARK_DAEMON_JAVA_OPTS": + "$SPARK_DAEMON_JAVA_OPTS -Dspark.shuffle.service.enabled=true"}` + ssh_public_keys: + type: array + items: + type: string + description: >- + SSH public key contents that will be added to each Spark node in + this cluster. The corresponding private keys can be used to login + with the user name `ubuntu` on port `2200`. Up to 10 keys can be + specified. + PipelineClusterAutoscale: + type: object + properties: + min_workers: + type: integer + max_workers: + type: integer + description: >- + The maximum number of workers to which the cluster can scale up when + overloaded. `max_workers` must be strictly greater than + `min_workers`. + mode: + $ref: '#/components/schemas/PipelineClusterAutoscaleMode' + description: >- + Databricks Enhanced Autoscaling optimizes cluster utilization by + automatically allocating cluster resources based on workload volume, + with minimal impact to the data processing latency of your + pipelines. Enhanced Autoscaling is available for `updates` clusters + only. The legacy autoscaling feature is used for `maintenance` + clusters. + required: + - min_workers + - max_workers + PipelineDeployment: + type: object + properties: + kind: + $ref: '#/components/schemas/DeploymentKind' + metadata_file_path: + type: string + description: The path to the file containing metadata about the deployment. + required: + - kind + PipelineEvent: + type: object + properties: + error: + $ref: '#/components/schemas/ErrorDetail' + event_type: + type: string + description: The event type. Should always correspond to the details + id: + type: string + description: A time-based, globally unique id. + level: + $ref: '#/components/schemas/EventLevel' + description: The severity level of the event. + maturity_level: + $ref: '#/components/schemas/MaturityLevel' + description: Maturity level for event_type. + message: + type: string + description: The display message associated with the event. + origin: + $ref: '#/components/schemas/Origin' + description: Describes where the event originates from. + sequence: + $ref: '#/components/schemas/Sequencing' + description: A sequencing object to identify and order events. + timestamp: + type: string + description: The time of the event. + truncation: + $ref: '#/components/schemas/Truncation' + description: >- + Information about which fields were truncated from this event due to + size constraints. If empty or absent, no truncation occurred. See + https://docs.databricks.com/en/ldp/monitor-event-logs for + information on retrieving complete event data. + PipelineLibrary: + type: object + properties: + file: + $ref: '#/components/schemas/FileLibrary' + glob: + $ref: '#/components/schemas/PathPattern' + description: >- + The unified field to include source codes. Each entry can be a + notebook path, a file path, or a folder path that ends `/**`. This + field cannot be used together with `notebook` or `file`. + jar: + type: string + description: URI of the jar to be installed. Currently only DBFS is supported. + maven: + type: string + description: Specification of a maven library to be installed. + notebook: + $ref: '#/components/schemas/NotebookLibrary' + description: >- + The path to a notebook that defines a pipeline and is stored in the + Databricks workspace. + whl: + type: string + description: URI of the whl to be installed. + PipelinePermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/PipelinePermissionLevel' + PipelinePermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/PipelineAccessControlResponse' + object_id: + type: string + object_type: + type: string + PipelinePermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/PipelinePermissionLevel' + PipelineSpec: + type: object + properties: + budget_policy_id: + type: string + catalog: + type: string + description: >- + A catalog in Unity Catalog to publish data from this pipeline to. If + `target` is specified, tables in this pipeline are published to a + `target` schema inside `catalog` (for example, + `catalog`.`target`.`table`). If `target` is not specified, no data + is published to Unity Catalog. + channel: + type: string + description: DLT Release Channel that specifies which version to use. + clusters: + type: array + items: + $ref: '#/components/schemas/PipelineCluster' + description: Cluster settings for this pipeline deployment. + configuration: + type: object + description: String-String configuration for this pipeline execution. + continuous: + type: boolean + description: >- + Whether the pipeline is continuous or triggered. This replaces + `trigger`. + deployment: + $ref: '#/components/schemas/PipelineDeployment' + description: Deployment type of this pipeline. + development: + type: boolean + description: Whether the pipeline is in Development mode. Defaults to false. + edition: + type: string + description: Pipeline product edition. + environment: + $ref: '#/components/schemas/PipelinesEnvironment' + description: >- + Environment specification for this pipeline used to install + dependencies. + event_log: + $ref: '#/components/schemas/EventLogSpec' + description: Event log configuration for this pipeline + filters: + $ref: '#/components/schemas/Filters' + description: Filters on which Pipeline packages to include in the deployed graph. + gateway_definition: + $ref: '#/components/schemas/IngestionGatewayPipelineDefinition' + description: The definition of a gateway pipeline to support change data capture. + id: + type: string + description: Unique identifier for this pipeline. + ingestion_definition: + $ref: '#/components/schemas/IngestionPipelineDefinition' + description: >- + The configuration for a managed ingestion pipeline. These settings + cannot be used with the 'libraries', 'schema', 'target', or + 'catalog' settings. + libraries: + type: array + items: + $ref: '#/components/schemas/PipelineLibrary' + description: Libraries or code needed by this deployment. + name: + type: string + description: Friendly identifier for this pipeline. + notifications: + type: array + items: + $ref: '#/components/schemas/Notifications' + description: List of notification settings for this pipeline. + photon: + type: boolean + description: Whether Photon is enabled for this pipeline. + restart_window: + $ref: '#/components/schemas/RestartWindow' + description: Restart window of this pipeline. + root_path: + type: string + description: >- + Root path for this pipeline. This is used as the root directory when + editing the pipeline in the Databricks user interface and it is + added to sys.path when executing Python sources during pipeline + execution. + schema: + type: string + description: >- + The default schema (database) where tables are read from or + published to. + serverless: + type: boolean + description: Whether serverless compute is enabled for this pipeline. + storage: + type: string + description: DBFS root directory for storing checkpoints and tables. + tags: + type: object + description: >- + A map of tags associated with the pipeline. These are forwarded to + the cluster as cluster tags, and are therefore subject to the same + limitations. A maximum of 25 tags can be added to the pipeline. + target: + type: string + description: >- + Target schema (database) to add tables in this pipeline to. Exactly + one of `schema` or `target` must be specified. To publish to Unity + Catalog, also specify `catalog`. This legacy field is deprecated for + pipeline creation in favor of the `schema` field. + trigger: + $ref: '#/components/schemas/PipelineTrigger' + description: 'Which pipeline trigger to use. Deprecated: Use `continuous` instead.' + usage_policy_id: + type: string + description: Usage policy of this pipeline. + PipelineStateInfo: + type: object + properties: + cluster_id: + type: string + creator_user_name: + type: string + description: The username of the pipeline creator. + health: + $ref: '#/components/schemas/PipelineStateInfoHealth' + description: The health of a pipeline. + latest_updates: + type: array + items: + $ref: '#/components/schemas/UpdateStateInfo' + description: >- + Status of the latest updates for the pipeline. Ordered with the + newest update first. + name: + type: string + description: The user-friendly name of the pipeline. + pipeline_id: + type: string + description: The unique identifier of the pipeline. + run_as_user_name: + type: string + description: >- + The username that the pipeline runs as. This is a read only value + derived from the pipeline owner. + state: + $ref: '#/components/schemas/PipelineState' + PipelineTrigger: + type: object + properties: + cron: + $ref: '#/components/schemas/CronTrigger' + manual: + $ref: '#/components/schemas/ManualTrigger' + PipelinesEnvironment: + type: object + properties: + dependencies: + type: array + items: + type: string + description: >- + List of pip dependencies, as supported by the version of pip in this + environment. Each dependency is a pip requirement file line + https://pip.pypa.io/en/stable/reference/requirements-file-format/ + Allowed dependency could be , , (WSFS or Volumes in Databricks), + description: >- + The environment entity used to preserve serverless environment side + panel, jobs' environment for + non-notebook task, and DLT's environment for classic and serverless pipelines. In this minimal + environment spec, only pip dependencies are supported. + PostgresCatalogConfig: + type: object + properties: + slot_config: + $ref: '#/components/schemas/PostgresSlotConfig' + description: >- + Optional. The Postgres slot configuration to use for logical + replication + description: PG-specific catalog-level configuration parameters + PostgresSlotConfig: + type: object + properties: + publication_name: + type: string + description: The name of the publication to use for the Postgres source + slot_name: + type: string + description: >- + The name of the logical replication slot to use for the Postgres + source + description: >- + PostgresSlotConfig contains the configuration for a Postgres logical + replication slot + ReportSpec: + type: object + properties: + source_url: + type: string + destination_catalog: + type: string + description: Required. Destination catalog to store table. + destination_schema: + type: string + description: Required. Destination schema to store table. + destination_table: + type: string + description: >- + Required. Destination table name. The pipeline fails if a table with + that name already exists. + table_configuration: + $ref: '#/components/schemas/TableSpecificConfig' + description: >- + Configuration settings to control the ingestion of tables. These + settings override the table_configuration defined in the + IngestionPipelineDefinition object. + required: + - source_url + - destination_catalog + - destination_schema + RestartWindow: + type: object + properties: + start_hour: + type: integer + days_of_week: + type: array + items: + $ref: '#/components/schemas/DayOfWeek' + description: >- + Days of week in which the restart is allowed to happen (within a + five-hour window starting at start_hour). If not specified all days + of the week will be used. + time_zone_id: + type: string + description: >- + Time zone id of restart window. See + https://docs.databricks.com/sql/language-manual/sql-ref-syntax-aux-conf-mgmt-set-timezone.html + for details. If not specified, UTC will be used. + required: + - start_hour + RewindDatasetSpec: + type: object + properties: + cascade: + type: boolean + description: >- + Whether to cascade the rewind to dependent datasets. Must be + specified. + identifier: + type: string + description: The identifier of the dataset (e.g., "main.foo.tbl1"). + reset_checkpoints: + type: boolean + description: Whether to reset checkpoints for this dataset. + description: Configuration for rewinding a specific dataset. + RewindSpec: + type: object + properties: + datasets: + type: array + items: + $ref: '#/components/schemas/RewindDatasetSpec' + description: >- + List of datasets to rewind with specific configuration for each. + When not specified, all datasets will be rewound with cascade = true + and reset_checkpoints = true. + dry_run: + type: boolean + description: >- + If true, this is a dry run and we should emit the RewindSummary but + not perform the rewind. + rewind_timestamp: + type: string + description: The base timestamp to rewind to. Must be specified. + description: >- + Information about a rewind being requested for this pipeline or some of + the datasets in it. + RunAs: + type: object + properties: + service_principal_name: + type: string + description: >- + Application ID of an active service principal. Setting this field + requires the `servicePrincipal/user` role. + user_name: + type: string + description: >- + The email of an active workspace user. Users can only set this field + to their own email. + description: >- + Write-only setting, available only in Create/Update calls. Specifies the + user or service + principal that the pipeline runs as. If not specified, the pipeline runs as the user who created + the pipeline. + + Only `user_name` or `service_principal_name` can be specified. If both are specified, an error + is thrown. + SchemaSpec: + type: object + properties: + source_schema: + type: string + destination_catalog: + type: string + description: Required. Destination catalog to store tables. + destination_schema: + type: string + description: >- + Required. Destination schema to store tables in. Tables with the + same name as the source tables are created in this destination + schema. The pipeline fails If a table with the same name already + exists. + source_catalog: + type: string + description: >- + The source catalog name. Might be optional depending on the type of + source. + table_configuration: + $ref: '#/components/schemas/TableSpecificConfig' + description: >- + Configuration settings to control the ingestion of tables. These + settings are applied to all tables in this schema and override the + table_configuration defined in the IngestionPipelineDefinition + object. + required: + - source_schema + - destination_catalog + - destination_schema + Sequencing: + type: object + properties: + control_plane_seq_no: + type: integer + data_plane_id: + $ref: '#/components/schemas/DataPlaneId' + description: the ID assigned by the data plane. + SerializedException: + type: object + properties: + class_name: + type: string + message: + type: string + description: Exception message + stack: + type: array + items: + $ref: '#/components/schemas/StackFrame' + description: Stack trace consisting of a list of stack frames + SourceCatalogConfig: + type: object + properties: + postgres: + $ref: '#/components/schemas/PostgresCatalogConfig' + description: Postgres-specific catalog-level configuration parameters + source_catalog: + type: string + description: Source catalog name + description: >- + SourceCatalogConfig contains catalog-level custom configuration + parameters for each source + SourceConfig: + type: object + properties: + catalog: + $ref: '#/components/schemas/SourceCatalogConfig' + StackFrame: + type: object + properties: + declaring_class: + type: string + file_name: + type: string + description: File where the method is defined + line_number: + type: integer + description: Line from which the method was called + method_name: + type: string + description: Name of the method which was called + StartUpdateResponse: + type: object + properties: + update_id: + type: string + StopPipelineResponse: + type: object + properties: {} + TableSpec: + type: object + properties: + source_table: + type: string + destination_catalog: + type: string + description: Required. Destination catalog to store table. + destination_schema: + type: string + description: Required. Destination schema to store table. + destination_table: + type: string + description: >- + Optional. Destination table name. The pipeline fails if a table with + that name already exists. If not set, the source table name is used. + source_catalog: + type: string + description: >- + Source catalog name. Might be optional depending on the type of + source. + source_schema: + type: string + description: >- + Schema name in the source database. Might be optional depending on + the type of source. + table_configuration: + $ref: '#/components/schemas/TableSpecificConfig' + description: >- + Configuration settings to control the ingestion of tables. These + settings override the table_configuration defined in the + IngestionPipelineDefinition object and the SchemaSpec. + required: + - source_table + - destination_catalog + - destination_schema + TableSpecificConfig: + type: object + properties: + auto_full_refresh_policy: + $ref: '#/components/schemas/AutoFullRefreshPolicy' + exclude_columns: + type: array + items: + type: string + description: >- + A list of column names to be excluded for the ingestion. When not + specified, include_columns fully controls what columns to be + ingested. When specified, all other columns including future ones + will be automatically included for ingestion. This field in mutually + exclusive with `include_columns`. + include_columns: + type: array + items: + type: string + description: >- + A list of column names to be included for the ingestion. When not + specified, all columns except ones in exclude_columns will be + included. Future columns will be automatically included. When + specified, all other future columns will be automatically excluded + from ingestion. This field in mutually exclusive with + `exclude_columns`. + primary_keys: + type: array + items: + type: string + description: The primary key of the table used to apply changes. + query_based_connector_config: + $ref: >- + #/components/schemas/IngestionPipelineDefinitionTableSpecificConfigQueryBasedConnectorConfig + row_filter: + type: string + description: >- + (Optional, Immutable) The row filter condition to be applied to the + table. It must not contain the WHERE keyword, only the actual filter + condition. It must be in DBSQL format. + salesforce_include_formula_fields: + type: boolean + description: >- + If true, formula fields defined in the table are included in the + ingestion. This setting is only valid for the Salesforce connector + scd_type: + $ref: '#/components/schemas/TableSpecificConfigScdType' + description: The SCD type to use to ingest the table. + sequence_by: + type: array + items: + type: string + description: >- + The column names specifying the logical order of events in the + source data. Spark Declarative Pipelines uses this sequencing to + handle change events that arrive out of order. + workday_report_parameters: + $ref: >- + #/components/schemas/IngestionPipelineDefinitionWorkdayReportParameters + description: (Optional) Additional custom parameters for Workday Report + Truncation: + type: object + properties: + truncated_fields: + type: array + items: + $ref: '#/components/schemas/TruncationTruncationDetail' + description: >- + List of fields that were truncated from this event. If empty or + absent, no truncation occurred. + description: Information about truncations applied to this event. + TruncationTruncationDetail: + type: object + properties: + field_name: + type: string + description: >- + The name of the truncated field (e.g., "error"). Corresponds to + field names in PipelineEvent. + description: Details about a specific field that was truncated. + UpdateInfo: + type: object + properties: + cause: + $ref: '#/components/schemas/UpdateInfoCause' + cluster_id: + type: string + description: The ID of the cluster that the update is running on. + config: + $ref: '#/components/schemas/PipelineSpec' + description: >- + The pipeline configuration with system defaults applied where + unspecified by the user. Not returned by ListUpdates. + creation_time: + type: integer + description: The time when this update was created. + full_refresh: + type: boolean + description: If true, this update will reset all tables before running. + full_refresh_selection: + type: array + items: + type: string + description: >- + A list of tables to update with fullRefresh. If both + refresh_selection and full_refresh_selection are empty, this is a + full graph update. Full Refresh on a table means that the states of + the table will be reset before the refresh. + pipeline_id: + type: string + description: The ID of the pipeline. + refresh_selection: + type: array + items: + type: string + description: >- + A list of tables to update without fullRefresh. If both + refresh_selection and full_refresh_selection are empty, this is a + full graph update. Full Refresh on a table means that the states of + the table will be reset before the refresh. + state: + $ref: '#/components/schemas/UpdateInfoState' + description: The update state. + update_id: + type: string + description: The ID of this update. + validate_only: + type: boolean + description: >- + If true, this update only validates the correctness of pipeline + source code but does not materialize or publish any datasets. + UpdateStateInfo: + type: object + properties: + creation_time: + type: string + state: + $ref: '#/components/schemas/UpdateStateInfoState' + update_id: + type: string + CloneMode: + type: string + enum: + - MIGRATE_TO_UC + description: Enum to specify which mode of clone to execute + DayOfWeek: + type: string + enum: + - FRIDAY + - MONDAY + - SATURDAY + - SUNDAY + - THURSDAY + - TUESDAY + - WEDNESDAY + description: >- + Days of week in which the window is allowed to happen. If not specified + all days of the week + + will be used. + DeploymentKind: + type: string + enum: + - BUNDLE + description: >- + The deployment method that manages the pipeline: - BUNDLE: The pipeline + is managed by a + + Databricks Asset Bundle. + EventLevel: + type: string + enum: + - ERROR + - INFO + - METRICS + - WARN + description: The severity level of the event. + GetPipelineResponseHealth: + type: string + enum: + - HEALTHY + - UNHEALTHY + description: The health of a pipeline. + IngestionSourceType: + type: string + enum: + - BIGQUERY + - DYNAMICS365 + - FOREIGN_CATALOG + - GA4_RAW_DATA + - MANAGED_POSTGRESQL + - MYSQL + - NETSUITE + - ORACLE + - POSTGRESQL + - SALESFORCE + - SERVICENOW + - SHAREPOINT + - SQLSERVER + - TERADATA + - WORKDAY_RAAS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + MaturityLevel: + type: string + enum: + - DEPRECATED + - EVOLVING + - STABLE + description: Maturity level for EventDetails. + PipelineClusterAutoscaleMode: + type: string + enum: + - ENHANCED + - LEGACY + description: >- + Databricks Enhanced Autoscaling optimizes cluster utilization by + automatically allocating + + cluster resources based on workload volume, with minimal impact to the + data processing latency + + of your pipelines. Enhanced Autoscaling is available for `updates` + clusters only. The legacy + + autoscaling feature is used for `maintenance` clusters. + PipelinePermissionLevel: + type: string + enum: + - CAN_MANAGE + - CAN_RUN + - CAN_VIEW + - IS_OWNER + description: Permission level + PipelineState: + type: string + enum: + - DELETED + - DEPLOYING + - FAILED + - IDLE + - RECOVERING + - RESETTING + - RUNNING + - STARTING + - STOPPING + description: The pipeline state. + PipelineStateInfoHealth: + type: string + enum: + - HEALTHY + - UNHEALTHY + description: The health of a pipeline. + StartUpdateCause: + type: string + enum: + - API_CALL + - INFRASTRUCTURE_MAINTENANCE + - JOB_TASK + - RETRY_ON_FAILURE + - SCHEMA_CHANGE + - SERVICE_UPGRADE + - USER_ACTION + description: What triggered this update. + TableSpecificConfigScdType: + type: string + enum: + - APPEND_ONLY + - SCD_TYPE_1 + - SCD_TYPE_2 + description: The SCD type to use to ingest the table. + UpdateInfoCause: + type: string + enum: + - API_CALL + - INFRASTRUCTURE_MAINTENANCE + - JOB_TASK + - RETRY_ON_FAILURE + - SCHEMA_CHANGE + - SERVICE_UPGRADE + - USER_ACTION + description: What triggered this update. + UpdateInfoState: + type: string + enum: + - CANCELED + - COMPLETED + - CREATED + - FAILED + - INITIALIZING + - QUEUED + - RESETTING + - RUNNING + - SETTING_UP_TABLES + - STOPPING + - WAITING_FOR_RESOURCES + description: The update state. + UpdateStateInfoState: + type: string + enum: + - CANCELED + - COMPLETED + - CREATED + - FAILED + - INITIALIZING + - QUEUED + - RESETTING + - RUNNING + - SETTING_UP_TABLES + - STOPPING + - WAITING_FOR_RESOURCES + description: The update state. + x-stackQL-resources: + pipelines: + id: databricks_workspace.pipelines.pipelines + name: pipelines + title: Pipelines + methods: + clone: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1clone/post' + response: + mediaType: application/json + openAPIDocKey: '200' + create: + operation: + $ref: '#/paths/~1api~12.0~1pipelines/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1pipelines/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.statuses + delete: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + stop: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1stop/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/pipelines/methods/get' + - $ref: '#/components/x-stackQL-resources/pipelines/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/pipelines/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/pipelines/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/pipelines/methods/update' + pipeline_permission_levels: + id: databricks_workspace.pipelines.pipeline_permission_levels + name: pipeline_permission_levels + title: Pipeline Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/pipeline_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + pipeline_permissions: + id: databricks_workspace.pipelines.pipeline_permissions + name: pipeline_permissions + title: Pipeline Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1pipelines~1{pipeline_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/pipeline_permissions/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/pipeline_permissions/methods/update + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/pipeline_permissions/methods/set' + pipeline_updates: + id: databricks_workspace.pipelines.pipeline_updates + name: pipeline_updates + title: Pipeline Updates + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1pipelines~1{pipeline_id}~1updates~1{update_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1updates/get' + response: + mediaType: application/json + openAPIDocKey: '200' + start: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1updates/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/pipeline_updates/methods/get' + - $ref: '#/components/x-stackQL-resources/pipeline_updates/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + pipeline_events: + id: databricks_workspace.pipelines.pipeline_events + name: pipeline_events + title: Pipeline Events + methods: + list: + operation: + $ref: '#/paths/~1api~12.0~1pipelines~1{pipeline_id}~1events/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.events + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/pipeline_events/methods/list' + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/postgres.yaml b/providers/src/databricks_workspace/v00.00.00000/services/postgres.yaml new file mode 100644 index 00000000..2c15f54f --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/postgres.yaml @@ -0,0 +1,1586 @@ +openapi: 3.0.0 +info: + title: Databricks Postgres API (workspace) + description: >- + OpenAPI specification for the Databricks postgres service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.postgres +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/postgres/{parent}/branches: + post: + operationId: postgres_create_branch + summary: Creates a new database branch in the project. + tags: + - postgres + description: |- + Creates a new database branch in the project. + + :param parent: str + The Project where this Branch will be created. Format: projects/{project_id} + :param branch: :class:`Branch` + The Branch to create. + :param branch_id: str + The ID to use for the Branch. This becomes the final component of the branch's resource name. The ID + is required and must be 1-63 characters long, start with a lowercase letter, and contain only + lowercase letters, numbers, and hyphens. For example, `development` becomes + `projects/my-app/branches/development`. + + :returns: :class:`Operation` + parameters: + - name: parent + in: path + required: true + schema: + type: string + description: >- + The Project where this Branch will be created. Format: + projects/{project_id} + - name: branch_id + in: query + required: true + schema: + type: string + description: >- + The ID to use for the Branch. This becomes the final component of + the branch's resource name. The ID is required and must be 1-63 + characters long, start with a lowercase letter, and contain only + lowercase letters, numbers, and hyphens. For example, `development` + becomes `projects/my-app/branches/development`. + requestBody: + content: + application/json: + schema: + type: object + properties: + branch: + type: string + description: The Branch to create. + required: + - branch + responses: + '200': + description: Success + content: + application/json: + schema: + type: object + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + get: + operationId: postgres_list_branches + summary: Returns a paginated list of database branches in the project. + tags: + - postgres + description: |- + Returns a paginated list of database branches in the project. + + :param parent: str + The Project that owns this collection of branches. Format: projects/{project_id} + :param page_size: int (optional) + Upper bound for items returned. Cannot be negative. + :param page_token: str (optional) + Page token from a previous response. If not provided, returns the first page. + + :returns: Iterator over :class:`Branch` + parameters: + - name: parent + in: path + required: true + schema: + type: string + description: >- + The Project that owns this collection of branches. Format: + projects/{project_id} + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. Cannot be negative. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Page token from a previous response. If not provided, returns the + first page. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListBranchesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + /api/2.0/postgres/{parent}/endpoints: + post: + operationId: postgres_create_endpoint + summary: Creates a new compute endpoint in the branch. + tags: + - postgres + description: |- + Creates a new compute endpoint in the branch. + + :param parent: str + The Branch where this Endpoint will be created. Format: projects/{project_id}/branches/{branch_id} + :param endpoint: :class:`Endpoint` + The Endpoint to create. + :param endpoint_id: str + The ID to use for the Endpoint. This becomes the final component of the endpoint's resource name. + The ID is required and must be 1-63 characters long, start with a lowercase letter, and contain only + lowercase letters, numbers, and hyphens. For example, `primary` becomes + `projects/my-app/branches/development/endpoints/primary`. + + :returns: :class:`Operation` + parameters: + - name: parent + in: path + required: true + schema: + type: string + description: >- + The Branch where this Endpoint will be created. Format: + projects/{project_id}/branches/{branch_id} + - name: endpoint_id + in: query + required: true + schema: + type: string + description: >- + The ID to use for the Endpoint. This becomes the final component of + the endpoint's resource name. The ID is required and must be 1-63 + characters long, start with a lowercase letter, and contain only + lowercase letters, numbers, and hyphens. For example, `primary` + becomes `projects/my-app/branches/development/endpoints/primary`. + requestBody: + content: + application/json: + schema: + type: object + properties: + endpoint: + type: string + description: The Endpoint to create. + required: + - endpoint + responses: + '200': + description: Success + content: + application/json: + schema: + type: object + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + get: + operationId: postgres_list_endpoints + summary: Returns a paginated list of compute endpoints in the branch. + tags: + - postgres + description: |- + Returns a paginated list of compute endpoints in the branch. + + :param parent: str + The Branch that owns this collection of endpoints. Format: + projects/{project_id}/branches/{branch_id} + :param page_size: int (optional) + Upper bound for items returned. Cannot be negative. + :param page_token: str (optional) + Page token from a previous response. If not provided, returns the first page. + + :returns: Iterator over :class:`Endpoint` + parameters: + - name: parent + in: path + required: true + schema: + type: string + description: >- + The Branch that owns this collection of endpoints. Format: + projects/{project_id}/branches/{branch_id} + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. Cannot be negative. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Page token from a previous response. If not provided, returns the + first page. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListEndpointsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + /api/2.0/postgres/projects: + post: + operationId: postgres_create_project + summary: >- + Creates a new Lakebase Autoscaling Postgres database project, which + contains branches and compute + tags: + - postgres + description: >- + Creates a new Lakebase Autoscaling Postgres database project, which + contains branches and compute + + endpoints. + + + :param project: :class:`Project` + The Project to create. + :param project_id: str + The ID to use for the Project. This becomes the final component of the project's resource name. The + ID is required and must be 1-63 characters long, start with a lowercase letter, and contain only + lowercase letters, numbers, and hyphens. For example, `my-app` becomes `projects/my-app`. + + :returns: :class:`Operation` + parameters: + - name: project_id + in: query + required: true + schema: + type: string + description: >- + The ID to use for the Project. This becomes the final component of + the project's resource name. The ID is required and must be 1-63 + characters long, start with a lowercase letter, and contain only + lowercase letters, numbers, and hyphens. For example, `my-app` + becomes `projects/my-app`. + requestBody: + content: + application/json: + schema: + type: object + properties: + project: + type: string + description: The Project to create. + required: + - project + responses: + '200': + description: Success + content: + application/json: + schema: + type: object + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + get: + operationId: postgres_list_projects + summary: >- + Returns a paginated list of database projects in the workspace that the + user has permission to access. + tags: + - postgres + description: >- + Returns a paginated list of database projects in the workspace that the + user has permission to access. + + + :param page_size: int (optional) + Upper bound for items returned. Cannot be negative. + :param page_token: str (optional) + Page token from a previous response. If not provided, returns the first page. + + :returns: Iterator over :class:`Project` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. Cannot be negative. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Page token from a previous response. If not provided, returns the + first page. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListProjectsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + /api/2.0/postgres/{parent}/roles: + post: + operationId: postgres_create_role + summary: Creates a new Postgres role in the branch. + tags: + - postgres + description: |- + Creates a new Postgres role in the branch. + + :param parent: str + The Branch where this Role is created. Format: projects/{project_id}/branches/{branch_id} + :param role: :class:`Role` + The desired specification of a Role. + :param role_id: str (optional) + The ID to use for the Role, which will become the final component of the role's resource name. This + ID becomes the role in Postgres. + + This value should be 4-63 characters, and valid characters are lowercase letters, numbers, and + hyphens, as defined by RFC 1123. + + If role_id is not specified in the request, it is generated automatically. + + :returns: :class:`Operation` + parameters: + - name: parent + in: path + required: true + schema: + type: string + description: >- + The Branch where this Role is created. Format: + projects/{project_id}/branches/{branch_id} + - name: role_id + in: query + required: false + schema: + type: string + description: >- + The ID to use for the Role, which will become the final component of + the role's resource name. This ID becomes the role in Postgres. This + value should be 4-63 characters, and valid characters are lowercase + letters, numbers, and hyphens, as defined by RFC 1123. If role_id is + not specified in the request, it is generated automatically. + requestBody: + content: + application/json: + schema: + type: object + properties: + role: + type: string + description: The desired specification of a Role. + required: + - role + responses: + '200': + description: Success + content: + application/json: + schema: + type: object + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + get: + operationId: postgres_list_roles + summary: Returns a paginated list of Postgres roles in the branch. + tags: + - postgres + description: |- + Returns a paginated list of Postgres roles in the branch. + + :param parent: str + The Branch that owns this collection of roles. Format: projects/{project_id}/branches/{branch_id} + :param page_size: int (optional) + Upper bound for items returned. Cannot be negative. + :param page_token: str (optional) + Page token from a previous response. If not provided, returns the first page. + + :returns: Iterator over :class:`Role` + parameters: + - name: parent + in: path + required: true + schema: + type: string + description: >- + The Branch that owns this collection of roles. Format: + projects/{project_id}/branches/{branch_id} + - name: page_size + in: query + required: false + schema: + type: string + description: Upper bound for items returned. Cannot be negative. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Page token from a previous response. If not provided, returns the + first page. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListRolesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + /api/2.0/postgres/{name}: + delete: + operationId: postgres_delete_role + summary: Deletes the specified Postgres role. + tags: + - postgres + description: |- + Deletes the specified Postgres role. + + :param name: str + The full resource path of the role to delete. Format: + projects/{project_id}/branches/{branch_id}/roles/{role_id} + :param reassign_owned_to: str (optional) + Reassign objects. If this is set, all objects owned by the role are reassigned to the role specified + in this parameter. + + NOTE: setting this requires spinning up a compute to succeed, since it involves running SQL queries. + + TODO: #LKB-7187 implement reassign_owned_to on LBM side. This might end-up being a synchronous query + when this parameter is used. + + :returns: :class:`Operation` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The full resource path of the role to delete. Format: + projects/{project_id}/branches/{branch_id}/roles/{role_id} + - name: reassign_owned_to + in: query + required: false + schema: + type: string + description: >- + Reassign objects. If this is set, all objects owned by the role are + reassigned to the role specified in this parameter. NOTE: setting + this requires spinning up a compute to succeed, since it involves + running SQL queries. TODO: #LKB-7187 implement reassign_owned_to on + LBM side. This might end-up being a synchronous query when this + parameter is used. + responses: + '200': + description: Success + content: + application/json: + schema: + type: object + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + get: + operationId: postgres_get_role + summary: >- + Retrieves information about the specified Postgres role, including its + authentication method and + tags: + - postgres + description: >- + Retrieves information about the specified Postgres role, including its + authentication method and + + permissions. + + + :param name: str + The full resource path of the role to retrieve. Format: + projects/{project_id}/branches/{branch_id}/roles/{role_id} + + :returns: :class:`Role` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The full resource path of the role to retrieve. Format: + projects/{project_id}/branches/{branch_id}/roles/{role_id} + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Role' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + patch: + operationId: postgres_update_project + summary: Updates the specified database project. + tags: + - postgres + description: |- + Updates the specified database project. + + :param name: str + Output only. The full resource path of the project. Format: projects/{project_id} + :param project: :class:`Project` + The Project to update. + + The project's `name` field is used to identify the project to update. Format: projects/{project_id} + :param update_mask: FieldMask + The list of fields to update. If unspecified, all fields will be updated when possible. + + :returns: :class:`Operation` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + Output only. The full resource path of the project. Format: + projects/{project_id} + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + The list of fields to update. If unspecified, all fields will be + updated when possible. + requestBody: + content: + application/json: + schema: + type: object + properties: + project: + type: string + description: >- + The Project to update. The project's `name` field is used to + identify the project to update. Format: + projects/{project_id} + required: + - project + responses: + '200': + description: Success + content: + application/json: + schema: + type: object + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI + /api/2.0/postgres/credentials: + post: + operationId: postgres_generate_database_credential + summary: Generate OAuth credentials for a Postgres database. + tags: + - postgres + description: |- + Generate OAuth credentials for a Postgres database. + + :param endpoint: str + This field is not yet supported. The endpoint for which this credential will be generated. Format: + projects/{project_id}/branches/{branch_id}/endpoints/{endpoint_id} + :param claims: List[:class:`RequestedClaims`] (optional) + The returned token will be scoped to UC tables with the specified permissions. + + :returns: :class:`DatabaseCredential` + requestBody: + content: + application/json: + schema: + type: object + properties: + endpoint: + type: string + description: >- + This field is not yet supported. The endpoint for which this + credential will be generated. Format: + projects/{project_id}/branches/{branch_id}/endpoints/{endpoint_id} + claims: + type: string + description: >- + The returned token will be scoped to UC tables with the + specified permissions. + required: + - endpoint + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DatabaseCredential' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PostgresAPI +components: + schemas: + Branch: + type: object + properties: + create_time: + type: string + format: date-time + name: + type: string + description: >- + Output only. The full resource path of the branch. Format: + projects/{project_id}/branches/{branch_id} + parent: + type: string + description: >- + The project containing this branch (API resource hierarchy). Format: + projects/{project_id} Note: This field indicates where the branch + exists in the resource hierarchy. For point-in-time branching from + another branch, see `status.source_branch`. + spec: + $ref: '#/components/schemas/BranchSpec' + description: The spec contains the branch configuration. + status: + $ref: '#/components/schemas/BranchStatus' + description: The current status of a Branch. + uid: + type: string + description: System-generated unique ID for the branch. + update_time: + type: string + format: date-time + description: A timestamp indicating when the branch was last updated. + BranchOperationMetadata: + type: object + properties: {} + BranchSpec: + type: object + properties: + expire_time: + type: string + format: date-time + is_protected: + type: boolean + description: >- + When set to true, protects the branch from deletion and reset. + Associated compute endpoints and the project cannot be deleted while + the branch is protected. + no_expiry: + type: boolean + description: >- + Explicitly disable expiration. When set to true, the branch will not + expire. If set to false, the request is invalid; provide either ttl + or expire_time instead. + source_branch: + type: string + description: >- + The name of the source branch from which this branch was created + (data lineage for point-in-time recovery). If not specified, + defaults to the project's default branch. Format: + projects/{project_id}/branches/{branch_id} + source_branch_lsn: + type: string + description: >- + The Log Sequence Number (LSN) on the source branch from which this + branch was created. + source_branch_time: + type: string + format: date-time + description: >- + The point in time on the source branch from which this branch was + created. + ttl: + type: string + description: >- + Relative time-to-live duration. When set, the branch will expire at + creation_time + ttl. + BranchStatus: + type: object + properties: + current_state: + $ref: '#/components/schemas/BranchStatusState' + default: + type: boolean + description: Whether the branch is the project's default branch. + expire_time: + type: string + format: date-time + description: >- + Absolute expiration time for the branch. Empty if expiration is + disabled. + is_protected: + type: boolean + description: Whether the branch is protected. + logical_size_bytes: + type: integer + description: The logical size of the branch. + pending_state: + $ref: '#/components/schemas/BranchStatusState' + description: >- + The pending state of the branch, if a state transition is in + progress. + source_branch: + type: string + description: >- + The name of the source branch from which this branch was created. + Format: projects/{project_id}/branches/{branch_id} + source_branch_lsn: + type: string + description: >- + The Log Sequence Number (LSN) on the source branch from which this + branch was created. + source_branch_time: + type: string + format: date-time + description: >- + The point in time on the source branch from which this branch was + created. + state_change_time: + type: string + format: date-time + description: A timestamp indicating when the `current_state` began. + DatabaseCredential: + type: object + properties: + expire_time: + type: string + format: date-time + token: + type: string + description: >- + The OAuth token that can be used as a password when connecting to a + database. + DatabricksServiceExceptionWithDetailsProto: + type: object + properties: + details: + type: array + items: + type: object + error_code: + $ref: '#/components/schemas/ErrorCode' + message: + type: string + stack_trace: + type: string + description: Databricks Error that is returned by all Databricks APIs. + Endpoint: + type: object + properties: + create_time: + type: string + format: date-time + name: + type: string + description: >- + Output only. The full resource path of the endpoint. Format: + projects/{project_id}/branches/{branch_id}/endpoints/{endpoint_id} + parent: + type: string + description: >- + The branch containing this endpoint (API resource hierarchy). + Format: projects/{project_id}/branches/{branch_id} + spec: + $ref: '#/components/schemas/EndpointSpec' + description: >- + The spec contains the compute endpoint configuration, including + autoscaling limits, suspend timeout, and disabled state. + status: + $ref: '#/components/schemas/EndpointStatus' + description: Current operational status of the compute endpoint. + uid: + type: string + description: System-generated unique ID for the endpoint. + update_time: + type: string + format: date-time + description: A timestamp indicating when the compute endpoint was last updated. + EndpointHosts: + type: object + properties: + host: + type: string + description: >- + The hostname to connect to this endpoint. For read-write endpoints, + this is a read-write hostname which connects to the primary compute. + For read-only endpoints, this is a read-only hostname which allows + read-only operations. + description: >- + Encapsulates various hostnames (r/w or r/o, pooled or not) for an + endpoint. + EndpointOperationMetadata: + type: object + properties: {} + EndpointSettings: + type: object + properties: + pg_settings: + type: object + description: A raw representation of Postgres settings. + description: A collection of settings for a compute endpoint. + EndpointSpec: + type: object + properties: + endpoint_type: + $ref: '#/components/schemas/EndpointType' + autoscaling_limit_max_cu: + type: number + description: The maximum number of Compute Units. Minimum value is 0.5. + autoscaling_limit_min_cu: + type: number + description: The minimum number of Compute Units. Minimum value is 0.5. + disabled: + type: boolean + description: >- + Whether to restrict connections to the compute endpoint. Enabling + this option schedules a suspend compute operation. A disabled + compute endpoint cannot be enabled by a connection or console + action. + no_suspension: + type: boolean + description: >- + When set to true, explicitly disables automatic suspension (never + suspend). Should be set to true when provided. + settings: + $ref: '#/components/schemas/EndpointSettings' + suspend_timeout_duration: + type: string + description: >- + Duration of inactivity after which the compute endpoint is + automatically suspended. If specified should be between 60s and + 604800s (1 minute to 1 week). + required: + - endpoint_type + EndpointStatus: + type: object + properties: + autoscaling_limit_max_cu: + type: number + autoscaling_limit_min_cu: + type: number + description: The minimum number of Compute Units. + current_state: + $ref: '#/components/schemas/EndpointStatusState' + disabled: + type: boolean + description: >- + Whether to restrict connections to the compute endpoint. Enabling + this option schedules a suspend compute operation. A disabled + compute endpoint cannot be enabled by a connection or console + action. + endpoint_type: + $ref: '#/components/schemas/EndpointType' + description: The endpoint type. A branch can only have one READ_WRITE endpoint. + hosts: + $ref: '#/components/schemas/EndpointHosts' + description: Contains host information for connecting to the endpoint. + pending_state: + $ref: '#/components/schemas/EndpointStatusState' + settings: + $ref: '#/components/schemas/EndpointSettings' + suspend_timeout_duration: + type: string + description: >- + Duration of inactivity after which the compute endpoint is + automatically suspended. + ListBranchesResponse: + type: object + properties: + branches: + type: array + items: + $ref: '#/components/schemas/Branch' + next_page_token: + type: string + description: Token to request the next page of branches. + ListEndpointsResponse: + type: object + properties: + endpoints: + type: array + items: + $ref: '#/components/schemas/Endpoint' + next_page_token: + type: string + description: Token to request the next page of compute endpoints. + ListProjectsResponse: + type: object + properties: + next_page_token: + type: string + projects: + type: array + items: + $ref: '#/components/schemas/Project' + description: >- + List of all projects in the workspace that the user has permission + to access. + ListRolesResponse: + type: object + properties: + next_page_token: + type: string + roles: + type: array + items: + $ref: '#/components/schemas/Role' + description: List of Postgres roles in the branch. + Operation: + type: object + properties: + done: + type: boolean + description: >- + If the value is `false`, it means the operation is still in + progress. If `true`, the operation is completed, and either `error` + or `response` is available. + error: + $ref: '#/components/schemas/DatabricksServiceExceptionWithDetailsProto' + description: >- + The error result of the operation in case of failure or + cancellation. + metadata: + type: object + description: >- + Service-specific metadata associated with the operation. It + typically contains progress information and common metadata such as + create time. Some services might not provide such metadata. + name: + type: string + description: >- + The server-assigned name, which is only unique within the same + service that originally returns it. If you use the default HTTP + mapping, the `name` should be a resource name ending with + `operations/{unique_id}`. + response: + type: object + description: The normal, successful response of the operation. + description: >- + This resource represents a long-running operation that is the result of + a network API call. + Project: + type: object + properties: + create_time: + type: string + format: date-time + name: + type: string + description: >- + Output only. The full resource path of the project. Format: + projects/{project_id} + spec: + $ref: '#/components/schemas/ProjectSpec' + description: >- + The spec contains the project configuration, including display_name, + pg_version (Postgres version), history_retention_duration, and + default_endpoint_settings. + status: + $ref: '#/components/schemas/ProjectStatus' + description: The current status of a Project. + uid: + type: string + description: System-generated unique ID for the project. + update_time: + type: string + format: date-time + description: A timestamp indicating when the project was last updated. + ProjectDefaultEndpointSettings: + type: object + properties: + autoscaling_limit_max_cu: + type: number + description: The maximum number of Compute Units. Minimum value is 0.5. + autoscaling_limit_min_cu: + type: number + description: The minimum number of Compute Units. Minimum value is 0.5. + no_suspension: + type: boolean + description: >- + When set to true, explicitly disables automatic suspension (never + suspend). Should be set to true when provided. + pg_settings: + type: object + description: A raw representation of Postgres settings. + suspend_timeout_duration: + type: string + description: >- + Duration of inactivity after which the compute endpoint is + automatically suspended. If specified should be between 60s and + 604800s (1 minute to 1 week). + description: A collection of settings for a compute endpoint. + ProjectOperationMetadata: + type: object + properties: {} + ProjectSpec: + type: object + properties: + default_endpoint_settings: + $ref: '#/components/schemas/ProjectDefaultEndpointSettings' + display_name: + type: string + description: >- + Human-readable project name. Length should be between 1 and 256 + characters. + history_retention_duration: + type: string + description: >- + The number of seconds to retain the shared history for point in time + recovery for all branches in this project. Value should be between + 0s and 2592000s (up to 30 days). + pg_version: + type: integer + description: The major Postgres version number. Supported versions are 16 and 17. + ProjectStatus: + type: object + properties: + branch_logical_size_limit_bytes: + type: integer + default_endpoint_settings: + $ref: '#/components/schemas/ProjectDefaultEndpointSettings' + description: The effective default endpoint settings. + display_name: + type: string + description: The effective human-readable project name. + history_retention_duration: + type: string + description: >- + The effective number of seconds to retain the shared history for + point in time recovery. + owner: + type: string + description: The email of the project owner. + pg_version: + type: integer + description: The effective major Postgres version number. + synthetic_storage_size_bytes: + type: integer + description: The current space occupied by the project in storage. + RequestedClaims: + type: object + properties: + permission_set: + $ref: '#/components/schemas/RequestedClaimsPermissionSet' + resources: + type: array + items: + $ref: '#/components/schemas/RequestedResource' + RequestedResource: + type: object + properties: + table_name: + type: string + unspecified_resource_name: + type: string + Role: + type: object + properties: + create_time: + type: string + format: date-time + name: + type: string + description: >- + Output only. The full resource path of the role. Format: + projects/{project_id}/branches/{branch_id}/roles/{role_id} + parent: + type: string + description: >- + The Branch where this Role exists. Format: + projects/{project_id}/branches/{branch_id} + spec: + $ref: '#/components/schemas/RoleRoleSpec' + description: >- + The spec contains the role configuration, including identity type, + authentication method, and role attributes. + status: + $ref: '#/components/schemas/RoleRoleStatus' + description: >- + Current status of the role, including its identity type, + authentication method, and role attributes. + update_time: + type: string + format: date-time + description: Role represents a Postgres role within a Branch. + RoleOperationMetadata: + type: object + properties: {} + RoleRoleSpec: + type: object + properties: + auth_method: + $ref: '#/components/schemas/RoleAuthMethod' + identity_type: + $ref: '#/components/schemas/RoleIdentityType' + description: >- + The type of role. When specifying a managed-identity, the chosen + role_id must be a valid: * application ID for SERVICE_PRINCIPAL * + user email for USER * group name for GROUP + postgres_role: + type: string + description: >- + The name of the Postgres role. This expects a valid Postgres + identifier as specified in the link below. + https://www.postgresql.org/docs/current/sql-syntax-lexical.html#SQL-SYNTAX-IDENTIFIERS + Required when creating the Role. If you wish to create a Postgres + Role backed by a managed Databricks identity, then postgres_role + must be one of the following: 1. user email for IdentityType.USER 2. + app ID for IdentityType.SERVICE_PRINCIPAL 2. group name for + IdentityType.GROUP + RoleRoleStatus: + type: object + properties: + auth_method: + $ref: '#/components/schemas/RoleAuthMethod' + identity_type: + $ref: '#/components/schemas/RoleIdentityType' + description: The type of the role. + postgres_role: + type: string + description: The name of the Postgres role. + BranchStatusState: + type: string + enum: + - ARCHIVED + - IMPORTING + - INIT + - READY + - RESETTING + description: The state of the branch. + EndpointStatusState: + type: string + enum: + - ACTIVE + - IDLE + - INIT + description: The state of the compute endpoint. + EndpointType: + type: string + enum: + - ENDPOINT_TYPE_READ_ONLY + - ENDPOINT_TYPE_READ_WRITE + description: The compute endpoint type. Either `read_write` or `read_only`. + ErrorCode: + type: string + enum: + - ABORTED + - ALREADY_EXISTS + - BAD_REQUEST + - CANCELLED + - CATALOG_ALREADY_EXISTS + - CATALOG_DOES_NOT_EXIST + - CATALOG_NOT_EMPTY + - COULD_NOT_ACQUIRE_LOCK + - CUSTOMER_UNAUTHORIZED + - DAC_ALREADY_EXISTS + - DAC_DOES_NOT_EXIST + - DATA_LOSS + - DEADLINE_EXCEEDED + - DEPLOYMENT_TIMEOUT + - DIRECTORY_NOT_EMPTY + - DIRECTORY_PROTECTED + - DRY_RUN_FAILED + - ENDPOINT_NOT_FOUND + - EXTERNAL_LOCATION_ALREADY_EXISTS + - EXTERNAL_LOCATION_DOES_NOT_EXIST + - FEATURE_DISABLED + - GIT_CONFLICT + - GIT_REMOTE_ERROR + - GIT_SENSITIVE_TOKEN_DETECTED + - GIT_UNKNOWN_REF + - GIT_URL_NOT_ON_ALLOW_LIST + - INSECURE_PARTNER_RESPONSE + - INTERNAL_ERROR + - INVALID_PARAMETER_VALUE + - INVALID_STATE + - INVALID_STATE_TRANSITION + - IO_ERROR + - IPYNB_FILE_IN_REPO + - MALFORMED_PARTNER_RESPONSE + - MALFORMED_REQUEST + - MANAGED_RESOURCE_GROUP_DOES_NOT_EXIST + - MAX_BLOCK_SIZE_EXCEEDED + - MAX_CHILD_NODE_SIZE_EXCEEDED + - MAX_LIST_SIZE_EXCEEDED + - MAX_NOTEBOOK_SIZE_EXCEEDED + - MAX_READ_SIZE_EXCEEDED + - METASTORE_ALREADY_EXISTS + - METASTORE_DOES_NOT_EXIST + - METASTORE_NOT_EMPTY + - NOT_FOUND + - NOT_IMPLEMENTED + - PARTIAL_DELETE + - PERMISSION_DENIED + - PERMISSION_NOT_PROPAGATED + - PRINCIPAL_DOES_NOT_EXIST + - PROJECTS_OPERATION_TIMEOUT + - PROVIDER_ALREADY_EXISTS + - PROVIDER_DOES_NOT_EXIST + - PROVIDER_SHARE_NOT_ACCESSIBLE + - QUOTA_EXCEEDED + - RECIPIENT_ALREADY_EXISTS + - RECIPIENT_DOES_NOT_EXIST + - REQUEST_LIMIT_EXCEEDED + - RESOURCE_ALREADY_EXISTS + - RESOURCE_CONFLICT + - RESOURCE_DOES_NOT_EXIST + - RESOURCE_EXHAUSTED + - RESOURCE_LIMIT_EXCEEDED + - SCHEMA_ALREADY_EXISTS + - SCHEMA_DOES_NOT_EXIST + - SCHEMA_NOT_EMPTY + - SEARCH_QUERY_TOO_LONG + - SEARCH_QUERY_TOO_SHORT + - SERVICE_UNDER_MAINTENANCE + - SHARE_ALREADY_EXISTS + - SHARE_DOES_NOT_EXIST + - STORAGE_CREDENTIAL_ALREADY_EXISTS + - STORAGE_CREDENTIAL_DOES_NOT_EXIST + - TABLE_ALREADY_EXISTS + - TABLE_DOES_NOT_EXIST + - TEMPORARILY_UNAVAILABLE + - UNAUTHENTICATED + - UNAVAILABLE + - UNKNOWN + - UNPARSEABLE_HTTP_ERROR + - WORKSPACE_TEMPORARILY_UNAVAILABLE + description: >- + Error codes returned by Databricks APIs to indicate specific failure + conditions. + RequestedClaimsPermissionSet: + type: string + enum: + - READ_ONLY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + RoleAuthMethod: + type: string + enum: + - LAKEBASE_OAUTH_V1 + - NO_LOGIN + - PG_PASSWORD_SCRAM_SHA_256 + description: How the role is authenticated when connecting to Postgres. + RoleIdentityType: + type: string + enum: + - GROUP + - SERVICE_PRINCIPAL + - USER + description: >- + The type of the Databricks managed identity that this Role represents. + Leave empty if you wish + + to create a regular Postgres role not associated with a Databricks + identity. + x-stackQL-resources: + postgres_branches: + id: databricks_workspace.postgres.postgres_branches + name: postgres_branches + title: Postgres Branches + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{parent}~1branches/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{parent}~1branches/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.branches + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/postgres_branches/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/postgres_branches/methods/create' + update: [] + delete: [] + replace: [] + postgres_endpoints: + id: databricks_workspace.postgres.postgres_endpoints + name: postgres_endpoints + title: Postgres Endpoints + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{parent}~1endpoints/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{parent}~1endpoints/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.endpoints + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/postgres_endpoints/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/postgres_endpoints/methods/create' + update: [] + delete: [] + replace: [] + postgres_projects: + id: databricks_workspace.postgres.postgres_projects + name: postgres_projects + title: Postgres Projects + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1projects/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1projects/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.projects + update: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/postgres_projects/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/postgres_projects/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/postgres_projects/methods/update' + delete: [] + replace: [] + postgres_roles: + id: databricks_workspace.postgres.postgres_roles + name: postgres_roles + title: Postgres Roles + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{parent}~1roles/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{parent}~1roles/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.roles + delete: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/postgres_roles/methods/list' + - $ref: '#/components/x-stackQL-resources/postgres_roles/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/postgres_roles/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/postgres_roles/methods/delete' + replace: [] + postgres_credentials: + id: databricks_workspace.postgres.postgres_credentials + name: postgres_credentials + title: Postgres Credentials + methods: + generate: + operation: + $ref: '#/paths/~1api~12.0~1postgres~1credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/qualitymonitorv2.yaml b/providers/src/databricks_workspace/v00.00.00000/services/qualitymonitorv2.yaml new file mode 100644 index 00000000..22c9966e --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/qualitymonitorv2.yaml @@ -0,0 +1,430 @@ +openapi: 3.0.0 +info: + title: Databricks Qualitymonitorv2 API (workspace) + description: >- + OpenAPI specification for the Databricks qualitymonitorv2 service + (workspace-level APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.qualitymonitorv2 +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/quality-monitors: + post: + operationId: quality_monitor_v2_create_quality_monitor + summary: >- + [DEPRECATED] Create a quality monitor on UC object. Use Data Quality + Monitoring API instead. + tags: + - qualitymonitorv2 + - quality_monitor_v2 + description: >- + [DEPRECATED] Create a quality monitor on UC object. Use Data Quality + Monitoring API instead. + + + :param quality_monitor: :class:`QualityMonitor` + + + :returns: :class:`QualityMonitor` + requestBody: + content: + application/json: + schema: + type: object + properties: + quality_monitor: + type: string + description: ':returns: :class:`QualityMonitor`' + required: + - quality_monitor + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/QualityMonitor' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorV2API + get: + operationId: quality_monitor_v2_list_quality_monitor + summary: >- + [DEPRECATED] (Unimplemented) List quality monitors. Use Data Quality + Monitoring API instead. + tags: + - qualitymonitorv2 + - quality_monitor_v2 + description: >- + [DEPRECATED] (Unimplemented) List quality monitors. Use Data Quality + Monitoring API instead. + + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`QualityMonitor` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListQualityMonitorResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorV2API + /api/2.0/quality-monitors/{object_type}/{object_id}: + delete: + operationId: quality_monitor_v2_delete_quality_monitor + summary: >- + [DEPRECATED] Delete a quality monitor on UC object. Use Data Quality + Monitoring API instead. + tags: + - qualitymonitorv2 + - quality_monitor_v2 + description: >- + [DEPRECATED] Delete a quality monitor on UC object. Use Data Quality + Monitoring API instead. + + + :param object_type: str + The type of the monitored object. Can be one of the following: schema. + :param object_id: str + The uuid of the request object. For example, schema id. + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + schema. + - name: object_id + in: path + required: true + schema: + type: string + description: The uuid of the request object. For example, schema id. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorV2API + get: + operationId: quality_monitor_v2_get_quality_monitor + summary: >- + [DEPRECATED] Read a quality monitor on UC object. Use Data Quality + Monitoring API instead. + tags: + - qualitymonitorv2 + - quality_monitor_v2 + description: >- + [DEPRECATED] Read a quality monitor on UC object. Use Data Quality + Monitoring API instead. + + + :param object_type: str + The type of the monitored object. Can be one of the following: schema. + :param object_id: str + The uuid of the request object. For example, schema id. + + :returns: :class:`QualityMonitor` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + schema. + - name: object_id + in: path + required: true + schema: + type: string + description: The uuid of the request object. For example, schema id. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/QualityMonitor' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorV2API + put: + operationId: quality_monitor_v2_update_quality_monitor + summary: >- + [DEPRECATED] (Unimplemented) Update a quality monitor on UC object. Use + Data Quality Monitoring API + tags: + - qualitymonitorv2 + - quality_monitor_v2 + description: >- + [DEPRECATED] (Unimplemented) Update a quality monitor on UC object. Use + Data Quality Monitoring API + + instead. + + + :param object_type: str + The type of the monitored object. Can be one of the following: schema. + :param object_id: str + The uuid of the request object. For example, schema id. + :param quality_monitor: :class:`QualityMonitor` + + + :returns: :class:`QualityMonitor` + parameters: + - name: object_type + in: path + required: true + schema: + type: string + description: >- + The type of the monitored object. Can be one of the following: + schema. + - name: object_id + in: path + required: true + schema: + type: string + description: The uuid of the request object. For example, schema id. + requestBody: + content: + application/json: + schema: + type: object + properties: + quality_monitor: + type: string + description: ':returns: :class:`QualityMonitor`' + required: + - quality_monitor + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/QualityMonitor' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QualityMonitorV2API +components: + schemas: + AnomalyDetectionConfig: + type: object + properties: + excluded_table_full_names: + type: array + items: + type: string + last_run_id: + type: string + description: Run id of the last run of the workflow + latest_run_status: + $ref: '#/components/schemas/AnomalyDetectionRunStatus' + description: The status of the last run of the workflow. + ListQualityMonitorResponse: + type: object + properties: + next_page_token: + type: string + quality_monitors: + type: array + items: + $ref: '#/components/schemas/QualityMonitor' + PercentNullValidityCheck: + type: object + properties: + column_names: + type: array + items: + type: string + upper_bound: + type: number + description: Optional upper bound; we should use auto determined bounds for now + QualityMonitor: + type: object + properties: + object_type: + type: string + object_id: + type: string + description: The uuid of the request object. For example, schema id. + anomaly_detection_config: + $ref: '#/components/schemas/AnomalyDetectionConfig' + validity_check_configurations: + type: array + items: + $ref: '#/components/schemas/ValidityCheckConfiguration' + description: Validity check configurations for anomaly detection. + required: + - object_type + - object_id + RangeValidityCheck: + type: object + properties: + column_names: + type: array + items: + type: string + lower_bound: + type: number + description: Lower bound for the range + upper_bound: + type: number + description: Upper bound for the range + UniquenessValidityCheck: + type: object + properties: + column_names: + type: array + items: + type: string + ValidityCheckConfiguration: + type: object + properties: + name: + type: string + percent_null_validity_check: + $ref: '#/components/schemas/PercentNullValidityCheck' + range_validity_check: + $ref: '#/components/schemas/RangeValidityCheck' + uniqueness_validity_check: + $ref: '#/components/schemas/UniquenessValidityCheck' + AnomalyDetectionRunStatus: + type: string + enum: + - ANOMALY_DETECTION_RUN_STATUS_CANCELED + - ANOMALY_DETECTION_RUN_STATUS_FAILED + - ANOMALY_DETECTION_RUN_STATUS_JOB_DELETED + - ANOMALY_DETECTION_RUN_STATUS_PENDING + - ANOMALY_DETECTION_RUN_STATUS_RUNNING + - ANOMALY_DETECTION_RUN_STATUS_SUCCESS + - ANOMALY_DETECTION_RUN_STATUS_UNKNOWN + - ANOMALY_DETECTION_RUN_STATUS_WORKSPACE_MISMATCH_ERROR + description: Status of Anomaly Detection Job Run + x-stackQL-resources: + quality_monitor_v2: + id: databricks_workspace.qualitymonitorv2.quality_monitor_v2 + name: quality_monitor_v2 + title: Quality Monitor V2 + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1quality-monitors/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1quality-monitors/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.quality_monitors + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1quality-monitors~1{object_type}~1{object_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1quality-monitors~1{object_type}~1{object_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1quality-monitors~1{object_type}~1{object_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/quality_monitor_v2/methods/get' + - $ref: '#/components/x-stackQL-resources/quality_monitor_v2/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/quality_monitor_v2/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/quality_monitor_v2/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/quality_monitor_v2/methods/update' diff --git a/providers/src/databricks_workspace/v00.00.00000/services/realtimeserving.yaml b/providers/src/databricks_workspace/v00.00.00000/services/realtimeserving.yaml deleted file mode 100644 index 779fca1e..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/realtimeserving.yaml +++ /dev/null @@ -1,1882 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Realtimeserving API - description: Realtimeserving -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/serving-endpoints/{name}/ai-gateway: - put: - operationId: ai-gateways-putaigateway - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/putaigateway - x-stackQL-resource: ai_gateways - x-stackQL-method: putaigateway - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Serving endpoint AI Gateway was updated successfully. - content: - application/json: - schema: - type: object - properties: - usage_tracking_config: - type: object - properties: - enabled: - type: boolean - inference_table_config: - type: object - properties: - catalog_name: - type: string - schema_name: - type: string - table_name_prefix: - type: string - enabled: - type: boolean - rate_limits: - type: array - items: - type: object - properties: - calls: - type: integer - key: - type: string - renewal_period: - type: string - guardrails: - type: object - properties: - input: - type: object - properties: - safety: - type: boolean - pii: - type: object - properties: - behavior: - type: string - valid_topics: - type: array - items: - type: string - invalid_keywords: - type: array - items: - type: string - output: - type: object - properties: - safety: - type: boolean - pii: - type: object - properties: - behavior: - type: string - valid_topics: - type: array - items: - type: string - invalid_keywords: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - usage_tracking_config: - type: object - inference_table_config: - type: object - rate_limits: - type: Array of object - guardrails: - type: object - example: - usage_tracking_config: - enabled: true - inference_table_config: - catalog_name: my-catalog - schema_name: my-schema - table_name_prefix: my-prefix - enabled: true - rate_limits: - - calls: 15 - key: user - renewal_period: minute - guardrails: - input: - safety: true - pii: - behavior: BLOCK - valid_topics: - - topic1 - invalid_keywords: - - keyword1 - output: - safety: true - pii: - behavior: BLOCK - valid_topics: - - topic1 - invalid_keywords: - - keyword1 - description: 'Used to update the AI Gateway of a serving endpoint. NOTE: Only - external model endpoints are currently supported.' - /api/2.0/serving-endpoints/{name}/served-models/{served_model_name}/build-logs: - get: - operationId: build-logs-buildlogs - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/buildlogs - x-stackQL-resource: build_logs - x-stackQL-method: buildlogs - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: path - - name: served_model_name - required: true - description: string - in: path - responses: - '200': - description: Build logs were fetched successfully. - content: - application/json: - schema: - type: object - properties: - logs: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves the build logs associated with the provided served model. - /api/2.0/serving-endpoints/{name}/served-models/{served_model_name}/logs: - get: - operationId: logs-logs - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/logs - x-stackQL-resource: logs - x-stackQL-method: logs - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: name - required: true - description: string - in: path - - name: served_model_name - required: true - description: string - in: path - responses: - '200': - description: Logs were fetched successfully. - content: - application/json: - schema: - type: object - properties: - logs: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves the service logs associated with the provided served - model. - /api/2.0/serving-endpoints/{name}/metrics: - get: - operationId: metrics-exportmetrics - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/exportmetrics - x-stackQL-resource: metrics - x-stackQL-method: exportmetrics - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Metrics for the serving endpoint in Prometheus exposition or - OpenMetrics exposition format were fetched successfully. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves the metrics associated with the provided serving endpoint - in either Prometheus or OpenMetrics exposition format. - /api/2.0/serving-endpoints/{name}/openapi: - get: - operationId: schemas-getopenapi - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/getopenapi - x-stackQL-resource: schemas - x-stackQL-method: getopenapi - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The OpenAPI spec for the invocation schema was fetched successfully - content: - application/json: - schema: - type: object - properties: - openapi: - type: string - info: - type: object - properties: - title: - type: string - version: - type: string - servers: - type: array - items: - type: object - properties: - url: - type: string - paths: - type: object - properties: - /served-models/model1/invocations: - type: object - properties: - post: - type: object - properties: - requestBody: - type: object - properties: - content: - type: object - properties: - application/json: - type: object - properties: - schema: - type: object - properties: - oneOf: - type: array - items: - type: object - properties: - properties: - type: object - properties: - dataframe_split: - type: object - properties: - properties: - type: object - properties: - columns: - type: object - properties: - type: - type: string - description: - type: string - items: - type: object - properties: - type: - type: string - enum: - type: array - items: - type: string - data: - type: object - properties: - type: - type: string - items: - type: object - properties: - type: - type: string - prefixItems: - type: array - items: - type: object - properties: - type: - type: string - format: - type: string - responses: - type: object - properties: - '200': - type: object - properties: - description: - type: string - content: - type: object - properties: - application/json: - type: object - properties: - schema: - type: object - properties: - properties: - type: object - properties: - predictions: - type: object - properties: - type: - type: string - items: - type: object - properties: - type: - type: string - format: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get the query schema of the serving endpoint in OpenAPI format. - The schema contains information for the supported paths, input and output - format and datatypes. - /api/2.0/serving-endpoints: - post: - operationId: serving-endpoints-create - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/create - x-stackQL-resource: serving_endpoints - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Serving endpoint was created successfully - content: - application/json: - schema: - type: object - properties: - name: - type: string - creator: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - state: - type: object - properties: - ready: - type: string - config: - type: object - properties: - served_entities: - type: array - items: - type: object - properties: - name: - type: string - external_model: - type: object - properties: - provider: - type: string - name: - type: string - task: - type: string - openai_config: - type: object - properties: - openai_api_key: - type: string - state: - type: object - properties: - deployment: - type: string - deployment_state_message: - type: string - creator: - type: string - creation_timestamp: - type: integer - traffic_config: - type: object - properties: - routes: - type: array - items: - type: object - properties: - served_model_name: - type: string - traffic_percentage: - type: integer - config_version: - type: integer - ai_gateway: - type: object - properties: - usage_tracking_config: - type: object - properties: - enabled: - type: boolean - inference_table_config: - type: object - properties: - catalog_name: - type: string - schema_name: - type: string - table_name_prefix: - type: string - enabled: - type: boolean - rate_limits: - type: array - items: - type: object - properties: - calls: - type: integer - key: - type: string - renewal_period: - type: string - guardrails: - type: object - properties: - input: - type: object - properties: - safety: - type: boolean - pii: - type: object - properties: - behavior: - type: string - valid_topics: - type: array - items: - type: string - invalid_keywords: - type: array - items: - type: string - output: - type: object - properties: - safety: - type: boolean - pii: - type: object - properties: - behavior: - type: string - valid_topics: - type: array - items: - type: string - invalid_keywords: - type: array - items: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - id: - type: string - permission_level: - type: string - route_optimized: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - route_optimized: - type: string - config: - type: required - tags: - type: object - rate_limits: - type: Array of object - ai_gateway: - type: Array of object - example: - name: openai_endpoint - config: - served_entities: - - name: openai_embeddings - external_model: - name: text-embedding-ada-002 - provider: openai - task: llm/v1/embeddings - openai_config: - openai_api_key: '{{secrets/my_scope/my_openai_api_key}}' - ai_gateway: - usage_tracking_config: - enabled: true - inference_table_config: - catalog_name: my-catalog - schema_name: my-schema - table_name_prefix: my-prefix - enabled: true - rate_limits: - - calls: 100 - key: user - renewal_period: minute - guardrails: - input: - safety: true - pii: - behavior: BLOCK - valid_topics: - - topic1 - - topic2 - invalid_keywords: - - keyword1 - - keyword2 - output: - safety: true - pii: - behavior: BLOCK - valid_topics: - - topic1 - - topic2 - invalid_keywords: - - keyword1 - - keyword2 - tags: - - key: team - value: gen-ai - get: - operationId: serving-endpoints-list - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/list - x-stackQL-resource: serving_endpoints - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: List of serving endpoints was retrieved successfully. - content: - application/json: - schema: - type: object - properties: - endpoints: - type: array - items: - type: object - properties: - name: - type: string - creator: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - state: - type: object - properties: - ready: - type: string - config_update: - type: string - config: - type: object - properties: - served_entities: - type: array - items: - type: object - properties: - name: - type: string - external_model: - type: object - properties: - provider: - type: string - name: - type: string - task: - type: string - ai21labs_config: - type: object - properties: - ai21labs_api_key: - type: string - ai21labs_api_key_plaintext: - type: string - anthropic_config: - type: object - properties: - anthropic_api_key: - type: string - anthropic_api_key_plaintext: - type: string - amazon_bedrock_config: - type: object - properties: - aws_region: - type: string - aws_access_key_id: - type: string - aws_access_key_id_plaintext: - type: string - aws_secret_access_key: - type: string - aws_secret_access_key_plaintext: - type: string - bedrock_provider: - type: string - cohere_config: - type: object - properties: - cohere_api_key: - type: string - cohere_api_key_plaintext: - type: string - cohere_api_base: - type: string - google_cloud_vertex_ai_config: - type: object - properties: - private_key: - type: string - private_key_plaintext: - type: string - region: - type: string - project_id: - type: string - databricks_model_serving_config: - type: object - properties: - databricks_api_token: - type: string - databricks_api_token_plaintext: - type: string - databricks_workspace_url: - type: string - openai_config: - type: object - properties: - openai_api_key: - type: string - openai_api_key_plaintext: - type: string - openai_api_type: - type: string - openai_api_base: - type: string - openai_api_version: - type: string - openai_organization: - type: string - openai_deployment_name: - type: string - microsoft_entra_tenant_id: - type: string - microsoft_entra_client_id: - type: string - microsoft_entra_client_secret: - type: string - microsoft_entra_client_secret_plaintext: - type: string - palm_config: - type: object - properties: - palm_api_key: - type: string - palm_api_key_plaintext: - type: string - foundation_model: - type: object - properties: - name: - type: string - display_name: - type: string - docs: - type: string - description: - type: string - entity_name: - type: string - entity_version: - type: string - served_models: - type: array - items: - type: object - properties: - name: - type: string - model_name: - type: string - model_version: - type: string - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - id: - type: string - task: - type: string - ai_gateway: - type: object - properties: - usage_tracking_config: - type: object - properties: - enabled: - type: boolean - inference_table_config: - type: object - properties: - catalog_name: - type: string - schema_name: - type: string - table_name_prefix: - type: string - enabled: - type: boolean - rate_limits: - type: array - items: - type: object - properties: - calls: - type: integer - key: - type: string - renewal_period: - type: string - guardrails: - type: object - properties: - input: - type: object - properties: - safety: - type: boolean - pii: - type: object - properties: - behavior: - type: string - valid_topics: - type: array - items: - type: string - invalid_keywords: - type: array - items: - type: string - output: - type: object - properties: - safety: - type: boolean - pii: - type: object - properties: - behavior: - type: string - valid_topics: - type: array - items: - type: string - invalid_keywords: - type: array - items: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.endpoints - /api/2.0/serving-endpoints/{name}: - delete: - operationId: serving-endpoints-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/delete - x-stackQL-resource: serving_endpoints - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Serving endpoint was deleted successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - get: - operationId: serving-endpoints-get - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/get - x-stackQL-resource: serving_endpoints - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Serving endpoint was retrieved successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - creator: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - state: - type: object - properties: - ready: - type: string - config_update: - type: string - config: - type: object - properties: - served_entities: - type: array - items: - type: object - properties: - name: - type: string - entity_name: - type: string - workload_size: - type: string - scale_to_zero_enabled: - type: boolean - workload_type: - type: string - state: - type: object - properties: - deployment: - type: string - deployment_state_message: - type: string - creator: - type: string - creation_timestamp: - type: integer - traffic_config: - type: object - properties: - routes: - type: array - items: - type: object - properties: - traffic_percentage: - type: integer - served_model_name: - type: string - config_version: - type: integer - tags: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - id: - type: string - permission_level: - type: string - route_optimized: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves the details for a single serving endpoint. - /api/2.0/serving-endpoints/{name}/tags: - patch: - operationId: serving-endpoints-patch - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/patch - x-stackQL-resource: serving_endpoints - x-stackQL-method: patch - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Serving endpoint tags were patched successfully. - content: - application/json: - schema: - type: array - items: - type: object - properties: - key: - type: string - value: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - delete_tags: - type: Array of string - add_tags: - type: Array of object - example: - add_tags: - - key: team - value: data science - delete_tags: - - string - description: Used to batch add and delete tags from a serving endpoint with - a single API call. - /api/2.0/serving-endpoints/{name}/rate-limits: - put: - operationId: serving-endpoints-put - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/put - x-stackQL-resource: serving_endpoints - x-stackQL-method: put - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Serving endpoint rate limits were updated successfully. - content: - application/json: - schema: - type: object - properties: - rate_limits: - type: array - items: - type: object - properties: - calls: - type: integer - key: - type: string - renewal_period: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - rate_limits: - type: Array of object - example: - rate_limits: - - calls: 15 - key: user - renewal_period: minute - description: 'Used to update the rate limits of a serving endpoint. NOTE: Only - foundation model endpoints are currently supported. For external models, use - AI Gateway to manage rate limits.' - /serving-endpoints/{name}/invocations: - post: - operationId: serving-endpoints-query - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/query - x-stackQL-resource: serving_endpoints - x-stackQL-method: query - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Serving endpoint was queried successfully and returned predictions. - content: - application/json: - schema: - type: object - properties: - id: - type: string - object: - type: string - created: - type: integer - model: - type: string - choices: - type: array - items: - type: object - properties: - text: - type: string - index: - type: integer - logprobs: - type: 'null' - finish_reason: - type: string - usage: - type: object - properties: - prompt_tokens: - type: integer - completion_tokens: - type: integer - total_tokens: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - prompt: - type: object - input: - type: object - temperature: - type: double - stop: - type: Array of string - max_tokens: - type: integer - n: - type: integer - stream: - type: boolean - dataframe_records: - type: Array of object - instances: - type: Array of object - inputs: - type: object - messages: - type: Array of object - extra_params: - type: object - dataframe_split: - type: object - example: - dataframe_records: - - user_latitude: 37 - user_longitude: 122.4 - destination_id: 10 - /api/2.0/serving-endpoints/{name}/config: - put: - operationId: serving-endpoints-updateconfig - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/updateconfig - x-stackQL-resource: serving_endpoints - x-stackQL-method: updateconfig - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Serving endpoint update was issued successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - creator: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - state: - type: object - properties: - ready: - type: string - config: - type: object - properties: - served_entities: - type: array - items: - type: object - properties: - name: - type: string - external_model: - type: object - properties: - provider: - type: string - name: - type: string - task: - type: string - openai_config: - type: object - properties: - openai_api_key: - type: string - state: - type: object - properties: - deployment: - type: string - deployment_state_message: - type: string - creator: - type: string - creation_timestamp: - type: integer - traffic_config: - type: object - properties: - routes: - type: array - items: - type: object - properties: - served_model_name: - type: string - traffic_percentage: - type: integer - config_version: - type: integer - id: - type: string - permission_level: - type: string - route_optimized: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - served_entities: - type: Array of object - served_models: - type: Array of object - traffic_config: - type: object - auto_capture_config: - type: object - example: - served_entities: - - name: openai_chat - external_model: - name: gpt-4 - provider: openai - task: llm/v1/chat - openai_config: - openai_api_key: '{{secrets/my_scope/my_openai_api_key}}' - description: Updates any combination of the serving endpoint's served entities, - the compute configuration of those served entities, and the endpoint's traffic - config. An endpoint that already has an update in progress can not be updated - until the current update completes or fails. - /api/2.0/permissions/serving-endpoints/{serving_endpoint_id}: - get: - operationId: serving-endpoints-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/getpermissions - x-stackQL-resource: serving_endpoints_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: serving_endpoint_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a serving endpoint. Serving endpoints can - inherit permissions from their root object. - put: - operationId: serving-endpoints-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/setpermissions - x-stackQL-resource: serving_endpoints_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: serving_endpoint_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: serving-endpoints-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/updatepermissions - x-stackQL-resource: serving_endpoints_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: serving_endpoint_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a serving endpoint. Serving endpoints - can inherit permissions from their root object. - /api/2.0/permissions/serving-endpoints/{serving_endpoint_id}/permissionLevels: - get: - operationId: serving-endpoints-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/servingendpoints/getpermissionlevels - x-stackQL-resource: serving_endpoints_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: serving_endpoint_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. -components: - x-stackQL-resources: - ai_gateways: - id: databricks_workspace.realtimeserving.ai_gateways - name: ai_gateways - title: Ai_gateways - methods: - putaigateway: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1ai-gateway/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/ai_gateways/methods/putaigateway' - build_logs: - id: databricks_workspace.realtimeserving.build_logs - name: build_logs - title: Build_logs - methods: - buildlogs: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1served-models~1{served_model_name}~1build-logs/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/build_logs/methods/buildlogs' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - logs: - id: databricks_workspace.realtimeserving.logs - name: logs - title: Logs - methods: - logs: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1served-models~1{served_model_name}~1logs/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/logs/methods/logs' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - metrics: - id: databricks_workspace.realtimeserving.metrics - name: metrics - title: Metrics - methods: - exportmetrics: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1metrics/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/metrics/methods/exportmetrics' - schemas: - id: databricks_workspace.realtimeserving.schemas - name: schemas - title: Schemas - methods: - getopenapi: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1openapi/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/schemas/methods/getopenapi' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - serving_endpoints: - id: databricks_workspace.realtimeserving.serving_endpoints - name: serving_endpoints - title: Serving_endpoints - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.endpoints - delete: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - patch: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1tags/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - put: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1rate-limits/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - query: - operation: - $ref: '#/paths/~1serving-endpoints~1{name}~1invocations/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updateconfig: - operation: - $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1config/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/get' - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/updateconfig' - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/patch' - replace: - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/put' - delete: - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/query' - serving_endpoints_permissions: - id: databricks_workspace.realtimeserving.serving_endpoints_permissions - name: serving_endpoints_permissions - title: Serving_endpoints_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/serving_endpoints_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/serving_endpoints_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/serving_endpoints_permissions/methods/setpermissions' - delete: [] - exec: [] - serving_endpoints_permission_levels: - id: databricks_workspace.realtimeserving.serving_endpoints_permission_levels - name: serving_endpoints_permission_levels - title: Serving_endpoints_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/serving_endpoints_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/repos.yaml b/providers/src/databricks_workspace/v00.00.00000/services/repos.yaml deleted file mode 100644 index a0a2830e..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/repos.yaml +++ /dev/null @@ -1,906 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Repos API - description: Repos -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/git-credentials: - post: - operationId: git-credentials-create - externalDocs: - url: https://docs.databricks.com/api/workspace/gitcredentials/create - x-stackQL-resource: git_credentials - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - credential_id: - type: string - git_provider: - type: string - git_username: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - git_provider: - type: required - git_username: - type: string - personal_access_token: - type: string - example: - git_provider: gitHub - git_username: testuser - personal_access_token: something - description: Creates a Git credential entry for the user. Only one Git credential - per user is supported, so any attempts to create credentials if an entry already - exists will fail. Use the PATCH endpoint to update existing credentials, or - the DELETE endpoint to delete existing credentials. - get: - operationId: git-credentials-list - externalDocs: - url: https://docs.databricks.com/api/workspace/gitcredentials/list - x-stackQL-resource: git_credentials - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - credentials: - type: array - items: - type: object - properties: - credential_id: - type: string - git_provider: - type: string - git_username: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.credentials - description: Lists the calling user's Git credentials. One credential per user - is supported. - /api/2.0/git-credentials/{credential_id}: - delete: - operationId: git-credentials-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/gitcredentials/delete - x-stackQL-resource: git_credentials - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: credential_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes the specified Git credential. - get: - operationId: git-credentials-get - externalDocs: - url: https://docs.databricks.com/api/workspace/gitcredentials/get - x-stackQL-resource: git_credentials - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: credential_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - credential_id: - type: string - git_provider: - type: string - git_username: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the Git credential with the specified credential ID. - patch: - operationId: git-credentials-update - externalDocs: - url: https://docs.databricks.com/api/workspace/gitcredentials/update - x-stackQL-resource: git_credentials - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: credential_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - personal_access_token: - type: string - git_provider: - type: required - git_username: - type: string - example: - git_provider: gitHub - git_username: testuser - personal_access_token: something - description: Updates the specified Git credential. - /api/2.0/permissions/repos/{repo_id}: - get: - operationId: permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/getpermissions - x-stackQL-resource: permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: repo_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a repo. Repos can inherit permissions from - their root object. - put: - operationId: permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/setpermissions - x-stackQL-resource: permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: repo_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/updatepermissions - x-stackQL-resource: permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: repo_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a repo. Repos can inherit permissions - from their root object. - /api/2.0/permissions/repos/{repo_id}/permissionLevels: - get: - operationId: permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/getpermissionlevels - x-stackQL-resource: permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: repo_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/repos: - post: - operationId: repos-create - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/create - x-stackQL-resource: repos - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - branch: - type: string - head_commit_id: - type: string - id: - type: string - path: - type: string - provider: - type: string - sparse_checkout: - type: object - properties: - patterns: - type: array - items: - type: string - url: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - url: - type: required - provider: - type: string - path: - type: required - sparse_checkout: - type: object - example: - branch: main - path: /Users/user@company.com/clitest - provider: gitHub - sparse_checkout: - patterns: - - parent-folder/child-folder - - src - - test - url: https://github.com/databricks/cli.git - description: Creates a repo in the workspace and links it to the remote Git - repo specified. Note that repos created programmatically must be linked to - a remote Git repo, unlike repos created in the browser. - get: - operationId: repos-list - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/list - x-stackQL-resource: repos - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: path_prefix - in: query - - name: next_page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - next_page_token: - type: string - repos: - type: array - items: - type: object - properties: - branch: - type: string - path: - type: string - provider: - type: string - sparse_checkout: - type: object - properties: - patterns: - type: array - items: - type: string - url: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.repos - description: Returns repos that the calling user has Manage permissions on. - Use - /api/2.0/repos/{repo_id}: - delete: - operationId: repos-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/delete - x-stackQL-resource: repos - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: repo_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes the specified repo. - get: - operationId: repos-get - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/get - x-stackQL-resource: repos - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: repo_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - branch: - type: string - head_commit_id: - type: string - id: - type: string - path: - type: string - provider: - type: string - url: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Returns the repo with the given repo ID. - patch: - operationId: repos-update - externalDocs: - url: https://docs.databricks.com/api/workspace/repos/update - x-stackQL-resource: repos - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: repo_id - required: true - description: int64 - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - branch: - type: string - tag: - type: string - sparse_checkout: - type: object - example: - branch: main - sparse_checkout: - patterns: - - parent-folder/child-folder - - src - - test - description: Updates the repo to a different branch or tag, or updates the repo - to the latest commit on the same branch. -components: - x-stackQL-resources: - git_credentials: - id: databricks_workspace.repos.git_credentials - name: git_credentials - title: Git_credentials - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1git-credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1git-credentials/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.credentials - delete: - operation: - $ref: '#/paths/~1api~12.0~1git-credentials~1{credential_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1git-credentials~1{credential_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1git-credentials~1{credential_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/git_credentials/methods/get' - - $ref: '#/components/x-stackQL-resources/git_credentials/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/git_credentials/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/git_credentials/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/git_credentials/methods/delete' - exec: [] - permissions: - id: databricks_workspace.repos.permissions - name: permissions - title: Permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1repos~1{repo_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1repos~1{repo_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1repos~1{repo_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/permissions/methods/setpermissions' - delete: [] - exec: [] - permission_levels: - id: databricks_workspace.repos.permission_levels - name: permission_levels - title: Permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1repos~1{repo_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - repos: - id: databricks_workspace.repos.repos - name: repos - title: Repos - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1repos/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1repos/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.repos - delete: - operation: - $ref: '#/paths/~1api~12.0~1repos~1{repo_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1repos~1{repo_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1repos~1{repo_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/repos/methods/get' - - $ref: '#/components/x-stackQL-resources/repos/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/repos/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/repos/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/repos/methods/delete' - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/secrets.yaml b/providers/src/databricks_workspace/v00.00.00000/services/secrets.yaml deleted file mode 100644 index c0f5bfdd..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/secrets.yaml +++ /dev/null @@ -1,618 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Secrets API - description: Secrets -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/secrets/acls/delete: - post: - operationId: acls-deleteacl - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/deleteacl - x-stackQL-resource: acls - x-stackQL-method: deleteacl - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - scope: - type: required - principal: - type: string - example: - scope: my-secret-scope - principal: data-scientists - description: Deletes the given ACL on the given scope. - /api/2.0/secrets/acls/get: - get: - operationId: acls-getacl - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/getacl - x-stackQL-resource: acls - x-stackQL-method: getacl - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: scope - required: true - description: string - in: query - - name: principal - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - principal: - type: string - permission: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the details about the given ACL, such as the group and permission. - Users must have the - /api/2.0/secrets/acls/list: - get: - operationId: acls-listacls - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/listacls - x-stackQL-resource: acls - x-stackQL-method: listacls - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: scope - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - acls: - type: array - items: - type: object - properties: - principal: - type: string - permission: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.acls - description: List the ACLs for a given secret scope. Users must have the - /api/2.0/secrets/acls/put: - post: - operationId: acls-putacl - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/putacl - x-stackQL-resource: acls - x-stackQL-method: putacl - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - scope: - type: required - principal: - type: string - permission: - type: required - example: - scope: my-secret-scope - principal: data-scientists - permission: READ - description: Creates or overwrites the Access Control List (ACL) associated - with the given principal (user or group) on the specified scope point. - /api/2.0/secrets/scopes/create: - post: - operationId: scopes-createscope - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/createscope - x-stackQL-resource: scopes - x-stackQL-method: createscope - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - scope: - type: required - initial_manage_principal: - type: string - scope_backend_type: - type: string - example: - scope: string - initial_manage_principal: string - scope_backend_type: DATABRICKS - description: The scope name must consist of alphanumeric characters, dashes, - underscores, and periods, and may not exceed 128 characters. - /api/2.0/secrets/scopes/delete: - post: - operationId: scopes-deletescope - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/deletescope - x-stackQL-resource: scopes - x-stackQL-method: deletescope - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - scope: - type: required - example: - scope: my-secret-scope - description: Deletes a secret scope. - /api/2.0/secrets/scopes/list: - get: - operationId: scopes-listscopes - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/listscopes - x-stackQL-resource: scopes - x-stackQL-method: listscopes - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - scopes: - type: array - items: - type: object - properties: - name: - type: string - backend_type: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.scopes - description: Lists all secret scopes available in the workspace. - /api/2.0/secrets/delete: - post: - operationId: secrets-deletesecret - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/deletesecret - x-stackQL-resource: secrets - x-stackQL-method: deletesecret - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - scope: - type: required - key: - type: string - example: - scope: my-secret-scope - key: my-secret-key - description: Deletes the secret stored in this secret scope. You must have - /api/2.0/secrets/get: - get: - operationId: secrets-getsecret - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/getsecret - x-stackQL-resource: secrets - x-stackQL-method: getsecret - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: scope - required: true - description: string - in: query - - name: key - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - key: - type: string - value: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the bytes representation of a secret value for the specified - scope and key. - /api/2.0/secrets/list: - get: - operationId: secrets-listsecrets - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/listsecrets - x-stackQL-resource: secrets - x-stackQL-method: listsecrets - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: scope - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - secrets: - type: array - items: - type: object - properties: - key: - type: string - last_updated_timestamp: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.secrets - description: Lists the secret keys that are stored at this scope. This is a - metadata-only operation; secret data cannot be retrieved using this API. Users - need the READ permission to make this call. - /api/2.0/secrets/put: - post: - operationId: secrets-putsecret - externalDocs: - url: https://docs.databricks.com/api/workspace/secrets/putsecret - x-stackQL-resource: secrets - x-stackQL-method: putsecret - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - scope: - type: required - key: - type: string - string_value: - type: required - bytes_value: - type: string - example: - scope: my-databricks-scope - key: my-string-key - string_value: foobar - description: Inserts a secret under the provided scope with the given name. If - a secret already exists with the same name, this command overwrites the existing - secret's value. The server encrypts the secret using the secret scope's encryption - settings before storing it. -components: - x-stackQL-resources: - acls: - id: databricks_workspace.secrets.acls - name: acls - title: Acls - methods: - deleteacl: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1acls~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getacl: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1acls~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listacls: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1acls~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.acls - putacl: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1acls~1put/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/acls/methods/listacls' - - $ref: '#/components/x-stackQL-resources/acls/methods/getacl' - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/acls/methods/putacl' - delete: - - $ref: '#/components/x-stackQL-resources/acls/methods/deleteacl' - exec: [] - scopes: - id: databricks_workspace.secrets.scopes - name: scopes - title: Scopes - methods: - createscope: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1scopes~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deletescope: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1scopes~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listscopes: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1scopes~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.scopes - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/scopes/methods/listscopes' - insert: - - $ref: '#/components/x-stackQL-resources/scopes/methods/createscope' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/scopes/methods/deletescope' - exec: [] - secrets: - id: databricks_workspace.secrets.secrets - name: secrets - title: Secrets - methods: - deletesecret: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getsecret: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listsecrets: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.secrets - putsecret: - operation: - $ref: '#/paths/~1api~12.0~1secrets~1put/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/secrets/methods/listsecrets' - - $ref: '#/components/x-stackQL-resources/secrets/methods/getsecret' - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/secrets/methods/putsecret' - delete: - - $ref: '#/components/x-stackQL-resources/secrets/methods/deletesecret' - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/serving.yaml b/providers/src/databricks_workspace/v00.00.00000/services/serving.yaml new file mode 100644 index 00000000..0fcef3e9 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/serving.yaml @@ -0,0 +1,2850 @@ +openapi: 3.0.0 +info: + title: Databricks Serving API (workspace) + description: OpenAPI specification for the Databricks serving service (workspace-level APIs), generated from the Databricks + Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.serving +servers: +- url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/serving-endpoints/{name}/served-models/{served_model_name}/build-logs: + get: + operationId: serving_endpoints_build_logs + summary: Retrieves the build logs associated with the provided served model. + tags: + - serving + - serving_endpoints + description: "Retrieves the build logs associated with the provided served model.\n\n:param name: str\n The name of\ + \ the serving endpoint that the served model belongs to. This field is required.\n:param served_model_name: str\n\ + \ The name of the served model that build logs will be retrieved for. This field is required.\n\n:returns: :class:`BuildLogsResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint that the served model belongs to. This field is required. + - name: served_model_name + in: path + required: true + schema: + type: string + description: The name of the served model that build logs will be retrieved for. This field is required. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/BuildLogsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints: + post: + operationId: serving_endpoints_create + summary: Create a new serving endpoint. + tags: + - serving + - serving_endpoints + description: "Create a new serving endpoint.\n\n:param name: str\n The name of the serving endpoint. This field is\ + \ required and must be unique across a Databricks\n workspace. An endpoint name can consist of alphanumeric characters,\ + \ dashes, and underscores.\n:param ai_gateway: :class:`AiGatewayConfig` (optional)\n The AI Gateway configuration\ + \ for the serving endpoint. NOTE: External model, provisioned throughput,\n and pay-per-token endpoints are fully\ + \ supported; agent endpoints currently only support inference\n tables.\n:param budget_policy_id: str (optional)\n\ + \ The budget policy to be applied to the serving endpoint.\n:param config: :class:`EndpointCoreConfigInput` (optional)\n\ + \ The core config of the serving endpoint.\n:param description: str (optional)\n:param email_notifications: :class:`EmailNotifications`\ + \ (optional)\n Email notification settings.\n:param rate_limits: List[:class:`RateLimit`] (optional)\n Rate limits\ + \ to be applied to the serving endpoint. NOTE: this field is deprecated, please use AI\n Gateway to manage rate limits.\n\ + :param route_optimized: bool (optional)\n Enable route optimization for the serving endpoint.\n:param tags: List[:class:`EndpointTag`]\ + \ (optional)\n Tags to be attached to the serving endpoint and automatically propagated to billing logs.\n\n:returns:\n\ + \ Long-running operation waiter for :class:`ServingEndpointDetailed`.\n See :method:wait_get_serving_endpoint_not_updating\ + \ for more details." + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: The name of the serving endpoint. This field is required and must be unique across a Databricks + workspace. An endpoint name can consist of alphanumeric characters, dashes, and underscores. + ai_gateway: + type: string + description: 'The AI Gateway configuration for the serving endpoint. NOTE: External model, provisioned throughput, + and pay-per-token endpoints are fully supported; agent endpoints currently only support inference tables.' + budget_policy_id: + type: string + description: The budget policy to be applied to the serving endpoint. + config: + type: string + description: The core config of the serving endpoint. + description: + type: string + description: ':param email_notifications: :class:`EmailNotifications` (optional) Email notification settings.' + email_notifications: + type: string + rate_limits: + type: string + description: 'Rate limits to be applied to the serving endpoint. NOTE: this field is deprecated, please + use AI Gateway to manage rate limits.' + route_optimized: + type: string + description: Enable route optimization for the serving endpoint. + tags: + type: string + description: Tags to be attached to the serving endpoint and automatically propagated to billing logs. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointDetailed' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + get: + operationId: serving_endpoints_list + summary: Get all serving endpoints. + tags: + - serving + - serving_endpoints + description: 'Get all serving endpoints. + + + + :returns: Iterator over :class:`ServingEndpoint`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListEndpointsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/pt: + post: + operationId: serving_endpoints_create_provisioned_throughput_endpoint + summary: Create a new PT serving endpoint. + tags: + - serving + - serving_endpoints + description: "Create a new PT serving endpoint.\n\n:param name: str\n The name of the serving endpoint. This field\ + \ is required and must be unique across a Databricks\n workspace. An endpoint name can consist of alphanumeric characters,\ + \ dashes, and underscores.\n:param config: :class:`PtEndpointCoreConfig`\n The core config of the serving endpoint.\n\ + :param ai_gateway: :class:`AiGatewayConfig` (optional)\n The AI Gateway configuration for the serving endpoint.\n\ + :param budget_policy_id: str (optional)\n The budget policy associated with the endpoint.\n:param email_notifications:\ + \ :class:`EmailNotifications` (optional)\n Email notification settings.\n:param tags: List[:class:`EndpointTag`]\ + \ (optional)\n Tags to be attached to the serving endpoint and automatically propagated to billing logs.\n\n:returns:\n\ + \ Long-running operation waiter for :class:`ServingEndpointDetailed`.\n See :method:wait_get_serving_endpoint_not_updating\ + \ for more details." + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: The name of the serving endpoint. This field is required and must be unique across a Databricks + workspace. An endpoint name can consist of alphanumeric characters, dashes, and underscores. + config: + type: string + description: The core config of the serving endpoint. + ai_gateway: + type: string + description: The AI Gateway configuration for the serving endpoint. + budget_policy_id: + type: string + description: The budget policy associated with the endpoint. + email_notifications: + type: string + description: Email notification settings. + tags: + type: string + description: Tags to be attached to the serving endpoint and automatically propagated to billing logs. + required: + - name + - config + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointDetailed' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}: + delete: + operationId: serving_endpoints_delete + summary: Delete a serving endpoint. + tags: + - serving + - serving_endpoints + description: 'Delete a serving endpoint. + + + :param name: str' + parameters: + - name: name + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + get: + operationId: serving_endpoints_get + summary: Retrieves the details for a single serving endpoint. + tags: + - serving + - serving_endpoints + description: "Retrieves the details for a single serving endpoint.\n\n:param name: str\n The name of the serving endpoint.\ + \ This field is required.\n\n:returns: :class:`ServingEndpointDetailed`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint. This field is required. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointDetailed' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/metrics: + get: + operationId: serving_endpoints_export_metrics + summary: Retrieves the metrics associated with the provided serving endpoint in either Prometheus or + tags: + - serving + - serving_endpoints + description: "Retrieves the metrics associated with the provided serving endpoint in either Prometheus or\nOpenMetrics\ + \ exposition format.\n\n:param name: str\n The name of the serving endpoint to retrieve metrics for. This field is\ + \ required.\n\n:returns: :class:`ExportMetricsResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint to retrieve metrics for. This field is required. + responses: + '200': + description: Success + content: + text/plain: + schema: + type: object + properties: + contents: + type: string + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/openapi: + get: + operationId: serving_endpoints_get_open_api + summary: Get the query schema of the serving endpoint in OpenAPI format. The schema contains information for + tags: + - serving + - serving_endpoints + description: "Get the query schema of the serving endpoint in OpenAPI format. The schema contains information for\n\ + the supported paths, input and output format and datatypes.\n\n:param name: str\n The name of the serving endpoint\ + \ that the served model belongs to. This field is required.\n\n:returns: :class:`GetOpenApiResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint that the served model belongs to. This field is required. + responses: + '200': + description: Success + content: + text/plain: + schema: + type: object + properties: + contents: + type: string + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/permissions/serving-endpoints/{serving_endpoint_id}/permissionLevels: + get: + operationId: serving_endpoints_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - serving + - serving_endpoints + description: "Gets the permission levels that a user can have on an object.\n\n:param serving_endpoint_id: str\n The\ + \ serving endpoint for which to get or manage permissions.\n\n:returns: :class:`GetServingEndpointPermissionLevelsResponse`" + parameters: + - name: serving_endpoint_id + in: path + required: true + schema: + type: string + description: The serving endpoint for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetServingEndpointPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/permissions/serving-endpoints/{serving_endpoint_id}: + get: + operationId: serving_endpoints_get_permissions + summary: Gets the permissions of a serving endpoint. Serving endpoints can inherit permissions from their root + tags: + - serving + - serving_endpoints + description: "Gets the permissions of a serving endpoint. Serving endpoints can inherit permissions from their root\n\ + object.\n\n:param serving_endpoint_id: str\n The serving endpoint for which to get or manage permissions.\n\n:returns:\ + \ :class:`ServingEndpointPermissions`" + parameters: + - name: serving_endpoint_id + in: path + required: true + schema: + type: string + description: The serving endpoint for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + put: + operationId: serving_endpoints_set_permissions + summary: Sets permissions on an object, replacing existing permissions if they exist. Deletes all direct + tags: + - serving + - serving_endpoints + description: "Sets permissions on an object, replacing existing permissions if they exist. Deletes all direct\npermissions\ + \ if none are specified. Objects can inherit permissions from their root object.\n\n:param serving_endpoint_id: str\n\ + \ The serving endpoint for which to get or manage permissions.\n:param access_control_list: List[:class:`ServingEndpointAccessControlRequest`]\ + \ (optional)\n\n:returns: :class:`ServingEndpointPermissions`" + parameters: + - name: serving_endpoint_id + in: path + required: true + schema: + type: string + description: The serving endpoint for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ServingEndpointPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + patch: + operationId: serving_endpoints_update_permissions + summary: Updates the permissions on a serving endpoint. Serving endpoints can inherit permissions from their + tags: + - serving + - serving_endpoints + description: "Updates the permissions on a serving endpoint. Serving endpoints can inherit permissions from their\n\ + root object.\n\n:param serving_endpoint_id: str\n The serving endpoint for which to get or manage permissions.\n\ + :param access_control_list: List[:class:`ServingEndpointAccessControlRequest`] (optional)\n\n:returns: :class:`ServingEndpointPermissions`" + parameters: + - name: serving_endpoint_id + in: path + required: true + schema: + type: string + description: The serving endpoint for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`ServingEndpointPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/external-function: + post: + operationId: serving_endpoints_http_request + summary: Make external services call using the credentials stored in UC Connection. + tags: + - serving + - serving_endpoints + description: "Make external services call using the credentials stored in UC Connection.\n\n:param connection_name:\ + \ str\n The connection name to use. This is required to identify the external connection.\n:param method: :class:`ExternalFunctionRequestHttpMethod`\n\ + \ The HTTP method to use (e.g., 'GET', 'POST').\n:param path: str\n The relative path for the API endpoint. This\ + \ is required.\n:param headers: str (optional)\n Additional headers for the request. If not provided, only auth headers\ + \ from connections would be\n passed.\n:param json: str (optional)\n The JSON payload to send in the request body.\n\ + :param params: str (optional)\n Query parameters for the request.\n\n:returns: :class:`HttpRequestResponse`" + requestBody: + content: + application/json: + schema: + type: object + properties: + connection_name: + type: string + description: The connection name to use. This is required to identify the external connection. + method: + type: string + description: The HTTP method to use (e.g., 'GET', 'POST'). + path: + type: string + description: The relative path for the API endpoint. This is required. + json: + type: string + description: The JSON payload to send in the request body. + params: + type: string + description: Query parameters for the request. + required: + - connection_name + - method + - path + responses: + '200': + description: Success + content: + text/plain: + schema: + type: object + properties: + contents: + type: string + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/served-models/{served_model_name}/logs: + get: + operationId: serving_endpoints_logs + summary: Retrieves the service logs associated with the provided served model. + tags: + - serving + - serving_endpoints + description: "Retrieves the service logs associated with the provided served model.\n\n:param name: str\n The name\ + \ of the serving endpoint that the served model belongs to. This field is required.\n:param served_model_name: str\n\ + \ The name of the served model that logs will be retrieved for. This field is required.\n\n:returns: :class:`ServerLogsResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint that the served model belongs to. This field is required. + - name: served_model_name + in: path + required: true + schema: + type: string + description: The name of the served model that logs will be retrieved for. This field is required. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServerLogsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/tags: + patch: + operationId: serving_endpoints_patch + summary: Used to batch add and delete tags from a serving endpoint with a single API call. + tags: + - serving + - serving_endpoints + description: "Used to batch add and delete tags from a serving endpoint with a single API call.\n\n:param name: str\n\ + \ The name of the serving endpoint who's tags to patch. This field is required.\n:param add_tags: List[:class:`EndpointTag`]\ + \ (optional)\n List of endpoint tags to add\n:param delete_tags: List[str] (optional)\n List of tag keys to delete\n\ + \n:returns: :class:`EndpointTags`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint who's tags to patch. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + add_tags: + type: string + description: List of endpoint tags to add + delete_tags: + type: string + description: List of tag keys to delete + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EndpointTags' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/rate-limits: + put: + operationId: serving_endpoints_put + summary: 'Deprecated: Please use AI Gateway to manage rate limits instead.' + tags: + - serving + - serving_endpoints + description: "Deprecated: Please use AI Gateway to manage rate limits instead.\n\n:param name: str\n The name of the\ + \ serving endpoint whose rate limits are being updated. This field is required.\n:param rate_limits: List[:class:`RateLimit`]\ + \ (optional)\n The list of endpoint rate limits.\n\n:returns: :class:`PutResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint whose rate limits are being updated. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + rate_limits: + type: string + description: The list of endpoint rate limits. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PutResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/ai-gateway: + put: + operationId: serving_endpoints_put_ai_gateway + summary: 'Used to update the AI Gateway of a serving endpoint. NOTE: External model, provisioned throughput, and' + tags: + - serving + - serving_endpoints + description: "Used to update the AI Gateway of a serving endpoint. NOTE: External model, provisioned throughput, and\n\ + pay-per-token endpoints are fully supported; agent endpoints currently only support inference tables.\n\n:param name:\ + \ str\n The name of the serving endpoint whose AI Gateway is being updated. This field is required.\n:param fallback_config:\ + \ :class:`FallbackConfig` (optional)\n Configuration for traffic fallback which auto fallbacks to other served entities\ + \ if the request to a\n served entity fails with certain error codes, to increase availability.\n:param guardrails:\ + \ :class:`AiGatewayGuardrails` (optional)\n Configuration for AI Guardrails to prevent unwanted data and unsafe data\ + \ in requests and responses.\n:param inference_table_config: :class:`AiGatewayInferenceTableConfig` (optional)\n \ + \ Configuration for payload logging using inference tables. Use these tables to monitor and audit data\n being sent\ + \ to and received from model APIs and to improve model quality.\n:param rate_limits: List[:class:`AiGatewayRateLimit`]\ + \ (optional)\n Configuration for rate limits which can be set to limit endpoint traffic.\n:param usage_tracking_config:\ + \ :class:`AiGatewayUsageTrackingConfig` (optional)\n Configuration to enable usage tracking using system tables.\ + \ These tables allow you to monitor\n operational usage on endpoints and their associated costs.\n\n:returns: :class:`PutAiGatewayResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint whose AI Gateway is being updated. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + fallback_config: + type: string + description: Configuration for traffic fallback which auto fallbacks to other served entities if the request + to a served entity fails with certain error codes, to increase availability. + guardrails: + type: string + description: Configuration for AI Guardrails to prevent unwanted data and unsafe data in requests and responses. + inference_table_config: + type: string + description: Configuration for payload logging using inference tables. Use these tables to monitor and audit + data being sent to and received from model APIs and to improve model quality. + rate_limits: + type: string + description: Configuration for rate limits which can be set to limit endpoint traffic. + usage_tracking_config: + type: string + description: Configuration to enable usage tracking using system tables. These tables allow you to monitor + operational usage on endpoints and their associated costs. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PutAiGatewayResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /serving-endpoints/{name}/invocations: + post: + operationId: serving_endpoints_query + summary: Query a serving endpoint + tags: + - serving + - serving_endpoints + description: "Query a serving endpoint\n\n:param name: str\n The name of the serving endpoint. This field is required\ + \ and is provided via the path parameter.\n:param client_request_id: str (optional)\n Optional user-provided request\ + \ identifier that will be recorded in the inference table and the usage\n tracking table.\n:param dataframe_records:\ + \ List[Any] (optional)\n Pandas Dataframe input in the records orientation.\n:param dataframe_split: :class:`DataframeSplitInput`\ + \ (optional)\n Pandas Dataframe input in the split orientation.\n:param extra_params: Dict[str,str] (optional)\n\ + \ The extra parameters field used ONLY for __completions, chat,__ and __embeddings external &\n foundation model__\ + \ serving endpoints. This is a map of strings and should only be used with other\n external/foundation model query\ + \ fields.\n:param input: Any (optional)\n The input string (or array of strings) field used ONLY for __embeddings\ + \ external & foundation\n model__ serving endpoints and is the only field (along with extra_params if needed) used\ + \ by\n embeddings queries.\n:param inputs: Any (optional)\n Tensor-based input in columnar format.\n:param instances:\ + \ List[Any] (optional)\n Tensor-based input in row format.\n:param max_tokens: int (optional)\n The max tokens field\ + \ used ONLY for __completions__ and __chat external & foundation model__ serving\n endpoints. This is an integer\ + \ and should only be used with other chat/completions query fields.\n:param messages: List[:class:`ChatMessage`] (optional)\n\ + \ The messages field used ONLY for __chat external & foundation model__ serving endpoints. This is an\n array of\ + \ ChatMessage objects and should only be used with other chat query fields.\n:param n: int (optional)\n The n (number\ + \ of candidates) field used ONLY for __completions__ and __chat external & foundation\n model__ serving endpoints.\ + \ This is an integer between 1 and 5 with a default of 1 and should only be\n used with other chat/completions query\ + \ fields.\n:param prompt: Any (optional)\n The prompt string (or array of strings) field used ONLY for __completions\ + \ external & foundation\n model__ serving endpoints and should only be used with other completions query fields.\n\ + :param stop: List[str] (optional)\n The stop sequences field used ONLY for __completions__ and __chat external &\ + \ foundation model__\n serving endpoints. This is a list of strings and should only be used with other chat/completions\n\ + \ query fields.\n:param stream: bool (optional)\n The stream field used ONLY for __completions__ and __chat external\ + \ & foundation model__ serving\n endpoints. This is a boolean defaulting to false and should only be used with other\ + \ chat/completions\n query fields.\n:param temperature: float (optional)\n The temperature field used ONLY for __completions__\ + \ and __chat external & foundation model__ serving\n endpoints. This is a float between 0.0 and 2.0 with a default\ + \ of 1.0 and should only be used with\n other chat/completions query fields.\n:param usage_context: Dict[str,str]\ + \ (optional)\n Optional user-provided context that will be recorded in the usage tracking table.\n\n:returns: :class:`QueryEndpointResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint. This field is required and is provided via the path parameter. + requestBody: + content: + application/json: + schema: + type: object + properties: + client_request_id: + type: string + description: Optional user-provided request identifier that will be recorded in the inference table and + the usage tracking table. + dataframe_records: + type: string + description: Pandas Dataframe input in the records orientation. + dataframe_split: + type: string + description: Pandas Dataframe input in the split orientation. + extra_params: + type: string + description: The extra parameters field used ONLY for __completions, chat,__ and __embeddings external & + foundation model__ serving endpoints. This is a map of strings and should only be used with other external/foundation + model query fields. + input: + type: string + description: The input string (or array of strings) field used ONLY for __embeddings external & foundation + model__ serving endpoints and is the only field (along with extra_params if needed) used by embeddings + queries. + inputs: + type: string + description: Tensor-based input in columnar format. + instances: + type: string + description: Tensor-based input in row format. + max_tokens: + type: string + description: The max tokens field used ONLY for __completions__ and __chat external & foundation model__ + serving endpoints. This is an integer and should only be used with other chat/completions query fields. + messages: + type: string + description: The messages field used ONLY for __chat external & foundation model__ serving endpoints. This + is an array of ChatMessage objects and should only be used with other chat query fields. + n: + type: string + description: The n (number of candidates) field used ONLY for __completions__ and __chat external & foundation + model__ serving endpoints. This is an integer between 1 and 5 with a default of 1 and should only be used + with other chat/completions query fields. + prompt: + type: string + description: The prompt string (or array of strings) field used ONLY for __completions external & foundation + model__ serving endpoints and should only be used with other completions query fields. + stop: + type: string + description: The stop sequences field used ONLY for __completions__ and __chat external & foundation model__ + serving endpoints. This is a list of strings and should only be used with other chat/completions query + fields. + stream: + type: string + description: The stream field used ONLY for __completions__ and __chat external & foundation model__ serving + endpoints. This is a boolean defaulting to false and should only be used with other chat/completions query + fields. + temperature: + type: string + description: The temperature field used ONLY for __completions__ and __chat external & foundation model__ + serving endpoints. This is a float between 0.0 and 2.0 with a default of 1.0 and should only be used with + other chat/completions query fields. + usage_context: + type: string + description: Optional user-provided context that will be recorded in the usage tracking table. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/QueryEndpointResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/config: + put: + operationId: serving_endpoints_update_config + summary: Updates any combination of the serving endpoint's served entities, the compute configuration of those + tags: + - serving + - serving_endpoints + description: "Updates any combination of the serving endpoint's served entities, the compute configuration of those\n\ + served entities, and the endpoint's traffic config. An endpoint that already has an update in progress\ncan not be\ + \ updated until the current update completes or fails.\n\n:param name: str\n The name of the serving endpoint to\ + \ update. This field is required.\n:param auto_capture_config: :class:`AutoCaptureConfigInput` (optional)\n Configuration\ + \ for Inference Tables which automatically logs requests and responses to Unity Catalog.\n Note: this field is deprecated\ + \ for creating new provisioned throughput endpoints, or updating\n existing provisioned throughput endpoints that\ + \ never have inference table configured; in these cases\n please use AI Gateway to manage inference tables.\n:param\ + \ served_entities: List[:class:`ServedEntityInput`] (optional)\n The list of served entities under the serving endpoint\ + \ config.\n:param served_models: List[:class:`ServedModelInput`] (optional)\n (Deprecated, use served_entities instead)\ + \ The list of served models under the serving endpoint\n config.\n:param traffic_config: :class:`TrafficConfig` (optional)\n\ + \ The traffic configuration associated with the serving endpoint config.\n\n:returns:\n Long-running operation waiter\ + \ for :class:`ServingEndpointDetailed`.\n See :method:wait_get_serving_endpoint_not_updating for more details." + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint to update. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + auto_capture_config: + type: string + description: 'Configuration for Inference Tables which automatically logs requests and responses to Unity + Catalog. Note: this field is deprecated for creating new provisioned throughput endpoints, or updating + existing provisioned throughput endpoints that never have inference table configured; in these cases please + use AI Gateway to manage inference tables.' + served_entities: + type: string + description: The list of served entities under the serving endpoint config. + served_models: + type: string + description: (Deprecated, use served_entities instead) The list of served models under the serving endpoint + config. + traffic_config: + type: string + description: The traffic configuration associated with the serving endpoint config. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointDetailed' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/{name}/notifications: + patch: + operationId: serving_endpoints_update_notifications + summary: Updates the email and webhook notification settings for an endpoint. + tags: + - serving + - serving_endpoints + description: "Updates the email and webhook notification settings for an endpoint.\n\n:param name: str\n The name of\ + \ the serving endpoint whose notifications are being updated. This field is required.\n:param email_notifications:\ + \ :class:`EmailNotifications` (optional)\n The email notification settings to update. Specify email addresses to\ + \ notify when endpoint state\n changes occur.\n\n:returns: :class:`UpdateInferenceEndpointNotificationsResponse`" + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the serving endpoint whose notifications are being updated. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + email_notifications: + type: string + description: The email notification settings to update. Specify email addresses to notify when endpoint + state changes occur. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateInferenceEndpointNotificationsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI + /api/2.0/serving-endpoints/pt/{name}/config: + put: + operationId: serving_endpoints_update_provisioned_throughput_endpoint_config + summary: Updates any combination of the pt endpoint's served entities, the compute configuration of those + tags: + - serving + - serving_endpoints + description: "Updates any combination of the pt endpoint's served entities, the compute configuration of those\nserved\ + \ entities, and the endpoint's traffic config. Updates are instantaneous and endpoint should be\nupdated instantly\n\ + \n:param name: str\n The name of the pt endpoint to update. This field is required.\n:param config: :class:`PtEndpointCoreConfig`\n\ + \n:returns:\n Long-running operation waiter for :class:`ServingEndpointDetailed`.\n See :method:wait_get_serving_endpoint_not_updating\ + \ for more details." + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the pt endpoint to update. This field is required. + requestBody: + content: + application/json: + schema: + type: object + properties: + config: + type: string + description: ':returns: Long-running operation waiter for :class:`ServingEndpointDetailed`. See :method:wait_get_serving_endpoint_not_updating + for more details.' + required: + - config + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ServingEndpointDetailed' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ServingEndpointsAPI +components: + schemas: + Ai21LabsConfig: + type: object + properties: + ai21labs_api_key: + type: string + ai21labs_api_key_plaintext: + type: string + description: 'An AI21 Labs API key provided as a plaintext string. If you prefer to reference your key using Databricks + Secrets, see `ai21labs_api_key`. You must provide an API key using one of the following fields: `ai21labs_api_key` + or `ai21labs_api_key_plaintext`.' + AiGatewayConfig: + type: object + properties: + fallback_config: + $ref: '#/components/schemas/FallbackConfig' + guardrails: + $ref: '#/components/schemas/AiGatewayGuardrails' + description: Configuration for AI Guardrails to prevent unwanted data and unsafe data in requests and responses. + inference_table_config: + $ref: '#/components/schemas/AiGatewayInferenceTableConfig' + description: Configuration for payload logging using inference tables. Use these tables to monitor and audit data + being sent to and received from model APIs and to improve model quality. + rate_limits: + type: array + items: + $ref: '#/components/schemas/AiGatewayRateLimit' + description: Configuration for rate limits which can be set to limit endpoint traffic. + usage_tracking_config: + $ref: '#/components/schemas/AiGatewayUsageTrackingConfig' + description: Configuration to enable usage tracking using system tables. These tables allow you to monitor operational + usage on endpoints and their associated costs. + AiGatewayGuardrailParameters: + type: object + properties: + invalid_keywords: + type: array + items: + type: string + pii: + $ref: '#/components/schemas/AiGatewayGuardrailPiiBehavior' + description: Configuration for guardrail PII filter. + safety: + type: boolean + description: Indicates whether the safety filter is enabled. + valid_topics: + type: array + items: + type: string + description: The list of allowed topics. Given a chat request, this guardrail flags the request if its topic is + not in the allowed topics. + AiGatewayGuardrailPiiBehavior: + type: object + properties: + behavior: + $ref: '#/components/schemas/AiGatewayGuardrailPiiBehaviorBehavior' + AiGatewayGuardrails: + type: object + properties: + input: + $ref: '#/components/schemas/AiGatewayGuardrailParameters' + output: + $ref: '#/components/schemas/AiGatewayGuardrailParameters' + description: Configuration for output guardrail filters. + AiGatewayInferenceTableConfig: + type: object + properties: + catalog_name: + type: string + enabled: + type: boolean + description: Indicates whether the inference table is enabled. + schema_name: + type: string + description: 'The name of the schema in Unity Catalog. Required when enabling inference tables. NOTE: On update, + you have to disable inference table first in order to change the schema name.' + table_name_prefix: + type: string + description: 'The prefix of the table in Unity Catalog. NOTE: On update, you have to disable inference table first + in order to change the prefix name.' + AiGatewayRateLimit: + type: object + properties: + renewal_period: + $ref: '#/components/schemas/AiGatewayRateLimitRenewalPeriod' + calls: + type: integer + description: Used to specify how many calls are allowed for a key within the renewal_period. + key: + $ref: '#/components/schemas/AiGatewayRateLimitKey' + description: Key field for a rate limit. Currently, 'user', 'user_group, 'service_principal', and 'endpoint' are + supported, with 'endpoint' being the default if not specified. + principal: + type: string + description: Principal field for a user, user group, or service principal to apply rate limiting to. Accepts a user + email, group name, or service principal application ID. + tokens: + type: integer + description: Used to specify how many tokens are allowed for a key within the renewal_period. + required: + - renewal_period + AiGatewayUsageTrackingConfig: + type: object + properties: + enabled: + type: boolean + AmazonBedrockConfig: + type: object + properties: + aws_region: + type: string + bedrock_provider: + $ref: '#/components/schemas/AmazonBedrockConfigBedrockProvider' + description: 'The underlying provider in Amazon Bedrock. Supported values (case insensitive) include: Anthropic, + Cohere, AI21Labs, Amazon.' + aws_access_key_id: + type: string + description: 'The Databricks secret key reference for an AWS access key ID with permissions to interact with Bedrock + services. If you prefer to paste your API key directly, see `aws_access_key_id_plaintext`. You must provide an + API key using one of the following fields: `aws_access_key_id` or `aws_access_key_id_plaintext`.' + aws_access_key_id_plaintext: + type: string + description: 'An AWS access key ID with permissions to interact with Bedrock services provided as a plaintext string. + If you prefer to reference your key using Databricks Secrets, see `aws_access_key_id`. You must provide an API + key using one of the following fields: `aws_access_key_id` or `aws_access_key_id_plaintext`.' + aws_secret_access_key: + type: string + description: 'The Databricks secret key reference for an AWS secret access key paired with the access key ID, with + permissions to interact with Bedrock services. If you prefer to paste your API key directly, see `aws_secret_access_key_plaintext`. + You must provide an API key using one of the following fields: `aws_secret_access_key` or `aws_secret_access_key_plaintext`.' + aws_secret_access_key_plaintext: + type: string + description: 'An AWS secret access key paired with the access key ID, with permissions to interact with Bedrock + services provided as a plaintext string. If you prefer to reference your key using Databricks Secrets, see `aws_secret_access_key`. + You must provide an API key using one of the following fields: `aws_secret_access_key` or `aws_secret_access_key_plaintext`.' + instance_profile_arn: + type: string + description: ARN of the instance profile that the external model will use to access AWS resources. You must authenticate + using an instance profile or access keys. If you prefer to authenticate using access keys, see `aws_access_key_id`, + `aws_access_key_id_plaintext`, `aws_secret_access_key` and `aws_secret_access_key_plaintext`. + required: + - aws_region + - bedrock_provider + AnthropicConfig: + type: object + properties: + anthropic_api_key: + type: string + anthropic_api_key_plaintext: + type: string + description: 'The Anthropic API key provided as a plaintext string. If you prefer to reference your key using Databricks + Secrets, see `anthropic_api_key`. You must provide an API key using one of the following fields: `anthropic_api_key` + or `anthropic_api_key_plaintext`.' + ApiKeyAuth: + type: object + properties: + key: + type: string + value: + type: string + description: The Databricks secret key reference for an API Key. If you prefer to paste your token directly, see + `value_plaintext`. + value_plaintext: + type: string + description: The API Key provided as a plaintext string. If you prefer to reference your token using Databricks + Secrets, see `value`. + required: + - key + AutoCaptureConfigInput: + type: object + properties: + catalog_name: + type: string + enabled: + type: boolean + description: Indicates whether the inference table is enabled. + schema_name: + type: string + description: 'The name of the schema in Unity Catalog. NOTE: On update, you cannot change the schema name if the + inference table is already enabled.' + table_name_prefix: + type: string + description: 'The prefix of the table in Unity Catalog. NOTE: On update, you cannot change the prefix name if the + inference table is already enabled.' + AutoCaptureConfigOutput: + type: object + properties: + catalog_name: + type: string + enabled: + type: boolean + description: Indicates whether the inference table is enabled. + schema_name: + type: string + description: 'The name of the schema in Unity Catalog. NOTE: On update, you cannot change the schema name if the + inference table is already enabled.' + state: + $ref: '#/components/schemas/AutoCaptureState' + table_name_prefix: + type: string + description: 'The prefix of the table in Unity Catalog. NOTE: On update, you cannot change the prefix name if the + inference table is already enabled.' + AutoCaptureState: + type: object + properties: + payload_table: + $ref: '#/components/schemas/PayloadTable' + BearerTokenAuth: + type: object + properties: + token: + type: string + token_plaintext: + type: string + description: The token provided as a plaintext string. If you prefer to reference your token using Databricks Secrets, + see `token`. + BuildLogsResponse: + type: object + properties: + logs: + type: string + required: + - logs + ChatMessage: + type: object + properties: + content: + type: string + role: + $ref: '#/components/schemas/ChatMessageRole' + description: The role of the message. One of [system, user, assistant]. + CohereConfig: + type: object + properties: + cohere_api_base: + type: string + cohere_api_key: + type: string + description: 'The Databricks secret key reference for a Cohere API key. If you prefer to paste your API key directly, + see `cohere_api_key_plaintext`. You must provide an API key using one of the following fields: `cohere_api_key` + or `cohere_api_key_plaintext`.' + cohere_api_key_plaintext: + type: string + description: 'The Cohere API key provided as a plaintext string. If you prefer to reference your key using Databricks + Secrets, see `cohere_api_key`. You must provide an API key using one of the following fields: `cohere_api_key` + or `cohere_api_key_plaintext`.' + CustomProviderConfig: + type: object + properties: + custom_provider_url: + type: string + description: This is a field to provide the URL of the custom provider API. + api_key_auth: + $ref: '#/components/schemas/ApiKeyAuth' + description: This is a field to provide API key authentication for the custom provider API. You can only specify + one authentication method. + bearer_token_auth: + $ref: '#/components/schemas/BearerTokenAuth' + description: This is a field to provide bearer token authentication for the custom provider API. You can only specify + one authentication method. + required: + - custom_provider_url + description: Configs needed to create a custom provider model route. + DataPlaneInfo: + type: object + properties: + authorization_details: + type: string + description: Authorization details as a string. + endpoint_url: + type: string + description: The URL of the endpoint for this operation in the dataplane. + description: Details necessary to query this object's API through the DataPlane APIs. + DatabricksModelServingConfig: + type: object + properties: + databricks_workspace_url: + type: string + databricks_api_token: + type: string + description: 'The Databricks secret key reference for a Databricks API token that corresponds to a user or service + principal with Can Query access to the model serving endpoint pointed to by this external model. If you prefer + to paste your API key directly, see `databricks_api_token_plaintext`. You must provide an API key using one of + the following fields: `databricks_api_token` or `databricks_api_token_plaintext`.' + databricks_api_token_plaintext: + type: string + description: 'The Databricks API token that corresponds to a user or service principal with Can Query access to + the model serving endpoint pointed to by this external model provided as a plaintext string. If you prefer to + reference your key using Databricks Secrets, see `databricks_api_token`. You must provide an API key using one + of the following fields: `databricks_api_token` or `databricks_api_token_plaintext`.' + required: + - databricks_workspace_url + DataframeSplitInput: + type: object + properties: + columns: + type: array + items: + type: object + data: + type: array + items: + type: object + description: Data array for the dataframe + index: + type: array + items: + type: integer + description: Index array for the dataframe + EmailNotifications: + type: object + properties: + on_update_failure: + type: array + items: + type: string + on_update_success: + type: array + items: + type: string + description: A list of email addresses to be notified when an endpoint successfully updates its configuration or + state. + EmbeddingsV1ResponseEmbeddingElement: + type: object + properties: + embedding: + type: array + items: + type: number + index: + type: integer + description: The index of the embedding in the response. + object: + $ref: '#/components/schemas/EmbeddingsV1ResponseEmbeddingElementObject' + description: This will always be 'embedding'. + EndpointCoreConfigInput: + type: object + properties: + name: + type: string + auto_capture_config: + $ref: '#/components/schemas/AutoCaptureConfigInput' + description: 'Configuration for Inference Tables which automatically logs requests and responses to Unity Catalog. + Note: this field is deprecated for creating new provisioned throughput endpoints, or updating existing provisioned + throughput endpoints that never have inference table configured; in these cases please use AI Gateway to manage + inference tables.' + served_entities: + type: array + items: + $ref: '#/components/schemas/ServedEntityInput' + description: The list of served entities under the serving endpoint config. + served_models: + type: array + items: + $ref: '#/components/schemas/ServedModelInput' + description: (Deprecated, use served_entities instead) The list of served models under the serving endpoint config. + traffic_config: + $ref: '#/components/schemas/TrafficConfig' + description: The traffic configuration associated with the serving endpoint config. + required: + - name + EndpointCoreConfigOutput: + type: object + properties: + auto_capture_config: + $ref: '#/components/schemas/AutoCaptureConfigOutput' + config_version: + type: integer + description: The config version that the serving endpoint is currently serving. + served_entities: + type: array + items: + $ref: '#/components/schemas/ServedEntityOutput' + description: The list of served entities under the serving endpoint config. + served_models: + type: array + items: + $ref: '#/components/schemas/ServedModelOutput' + description: (Deprecated, use served_entities instead) The list of served models under the serving endpoint config. + traffic_config: + $ref: '#/components/schemas/TrafficConfig' + description: The traffic configuration associated with the serving endpoint config. + EndpointCoreConfigSummary: + type: object + properties: + served_entities: + type: array + items: + $ref: '#/components/schemas/ServedEntitySpec' + served_models: + type: array + items: + $ref: '#/components/schemas/ServedModelSpec' + description: (Deprecated, use served_entities instead) The list of served models under the serving endpoint config. + EndpointPendingConfig: + type: object + properties: + auto_capture_config: + $ref: '#/components/schemas/AutoCaptureConfigOutput' + config_version: + type: integer + description: The config version that the serving endpoint is currently serving. + served_entities: + type: array + items: + $ref: '#/components/schemas/ServedEntityOutput' + description: The list of served entities belonging to the last issued update to the serving endpoint. + served_models: + type: array + items: + $ref: '#/components/schemas/ServedModelOutput' + description: (Deprecated, use served_entities instead) The list of served models belonging to the last issued update + to the serving endpoint. + start_time: + type: integer + description: The timestamp when the update to the pending config started. + traffic_config: + $ref: '#/components/schemas/TrafficConfig' + description: The traffic config defining how invocations to the serving endpoint should be routed. + EndpointState: + type: object + properties: + config_update: + $ref: '#/components/schemas/EndpointStateConfigUpdate' + ready: + $ref: '#/components/schemas/EndpointStateReady' + description: The state of an endpoint, indicating whether or not the endpoint is queryable. An endpoint is READY + if all of the served entities in its active configuration are ready. If any of the actively served entities are + in a non-ready state, the endpoint state will be NOT_READY. + EndpointTag: + type: object + properties: + key: + type: string + value: + type: string + description: Optional value field for a serving endpoint tag. + required: + - key + EndpointTags: + type: object + properties: + tags: + type: array + items: + $ref: '#/components/schemas/EndpointTag' + ExportMetricsResponse: + type: object + properties: + contents: + type: string + format: binary + ExternalModel: + type: object + properties: + provider: + $ref: '#/components/schemas/ExternalModelProvider' + name: + type: string + description: The name of the external model. + task: + type: string + description: The task type of the external model. + ai21labs_config: + $ref: '#/components/schemas/Ai21LabsConfig' + description: AI21Labs Config. Only required if the provider is 'ai21labs'. + amazon_bedrock_config: + $ref: '#/components/schemas/AmazonBedrockConfig' + description: Amazon Bedrock Config. Only required if the provider is 'amazon-bedrock'. + anthropic_config: + $ref: '#/components/schemas/AnthropicConfig' + description: Anthropic Config. Only required if the provider is 'anthropic'. + cohere_config: + $ref: '#/components/schemas/CohereConfig' + description: Cohere Config. Only required if the provider is 'cohere'. + custom_provider_config: + $ref: '#/components/schemas/CustomProviderConfig' + description: Custom Provider Config. Only required if the provider is 'custom'. + databricks_model_serving_config: + $ref: '#/components/schemas/DatabricksModelServingConfig' + description: Databricks Model Serving Config. Only required if the provider is 'databricks-model-serving'. + google_cloud_vertex_ai_config: + $ref: '#/components/schemas/GoogleCloudVertexAiConfig' + description: Google Cloud Vertex AI Config. Only required if the provider is 'google-cloud-vertex-ai'. + openai_config: + $ref: '#/components/schemas/OpenAiConfig' + description: OpenAI Config. Only required if the provider is 'openai'. + palm_config: + $ref: '#/components/schemas/PaLmConfig' + description: PaLM Config. Only required if the provider is 'palm'. + required: + - provider + - name + - task + ExternalModelUsageElement: + type: object + properties: + completion_tokens: + type: integer + prompt_tokens: + type: integer + description: The number of tokens in the prompt. + total_tokens: + type: integer + description: The total number of tokens in the prompt and response. + FallbackConfig: + type: object + properties: + enabled: + type: boolean + required: + - enabled + FoundationModel: + type: object + properties: + description: + type: string + display_name: + type: string + docs: + type: string + name: + type: string + description: "All fields are not sensitive as they are hard-coded in the system and made available to\n customers." + GetOpenApiResponse: + type: object + properties: + contents: + type: string + format: binary + GetServingEndpointPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/ServingEndpointPermissionsDescription' + GoogleCloudVertexAiConfig: + type: object + properties: + project_id: + type: string + region: + type: string + description: 'This is the region for the Google Cloud Vertex AI Service. See [supported regions] for more details. + Some models are only available in specific regions. [supported regions]: https://cloud.google.com/vertex-ai/docs/general/locations' + private_key: + type: string + description: 'The Databricks secret key reference for a private key for the service account which has access to + the Google Cloud Vertex AI Service. See [Best practices for managing service account keys]. If you prefer to paste + your API key directly, see `private_key_plaintext`. You must provide an API key using one of the following fields: + `private_key` or `private_key_plaintext` [Best practices for managing service account keys]: https://cloud.google.com/iam/docs/best-practices-for-managing-service-account-keys' + private_key_plaintext: + type: string + description: 'The private key for the service account which has access to the Google Cloud Vertex AI Service provided + as a plaintext secret. See [Best practices for managing service account keys]. If you prefer to reference your + key using Databricks Secrets, see `private_key`. You must provide an API key using one of the following fields: + `private_key` or `private_key_plaintext`. [Best practices for managing service account keys]: https://cloud.google.com/iam/docs/best-practices-for-managing-service-account-keys' + required: + - project_id + - region + HttpRequestResponse: + type: object + properties: + contents: + type: string + format: binary + ListEndpointsResponse: + type: object + properties: + endpoints: + type: array + items: + $ref: '#/components/schemas/ServingEndpoint' + ModelDataPlaneInfo: + type: object + properties: + query_info: + $ref: '#/components/schemas/DataPlaneInfo' + description: Information required to query DataPlane API 'query' endpoint. + description: "A representation of all DataPlaneInfo for operations that can be done on a model through Data\n Plane\ + \ APIs." + OpenAiConfig: + type: object + properties: + microsoft_entra_client_id: + type: string + description: This field is only required for Azure AD OpenAI and is the Microsoft Entra Client ID. + microsoft_entra_client_secret: + type: string + description: 'The Databricks secret key reference for a client secret used for Microsoft Entra ID authentication. + If you prefer to paste your client secret directly, see `microsoft_entra_client_secret_plaintext`. You must provide + an API key using one of the following fields: `microsoft_entra_client_secret` or `microsoft_entra_client_secret_plaintext`.' + microsoft_entra_client_secret_plaintext: + type: string + description: 'The client secret used for Microsoft Entra ID authentication provided as a plaintext string. If you + prefer to reference your key using Databricks Secrets, see `microsoft_entra_client_secret`. You must provide an + API key using one of the following fields: `microsoft_entra_client_secret` or `microsoft_entra_client_secret_plaintext`.' + microsoft_entra_tenant_id: + type: string + description: This field is only required for Azure AD OpenAI and is the Microsoft Entra Tenant ID. + openai_api_base: + type: string + description: This is a field to provide a customized base URl for the OpenAI API. For Azure OpenAI, this field is + required, and is the base URL for the Azure OpenAI API service provided by Azure. For other OpenAI API types, + this field is optional, and if left unspecified, the standard OpenAI base URL is used. + openai_api_key: + type: string + description: 'The Databricks secret key reference for an OpenAI API key using the OpenAI or Azure service. If you + prefer to paste your API key directly, see `openai_api_key_plaintext`. You must provide an API key using one of + the following fields: `openai_api_key` or `openai_api_key_plaintext`.' + openai_api_key_plaintext: + type: string + description: 'The OpenAI API key using the OpenAI or Azure service provided as a plaintext string. If you prefer + to reference your key using Databricks Secrets, see `openai_api_key`. You must provide an API key using one of + the following fields: `openai_api_key` or `openai_api_key_plaintext`.' + openai_api_type: + type: string + description: This is an optional field to specify the type of OpenAI API to use. For Azure OpenAI, this field is + required, and adjust this parameter to represent the preferred security access validation protocol. For access + token validation, use azure. For authentication using Azure Active Directory (Azure AD) use, azuread. + openai_api_version: + type: string + description: This is an optional field to specify the OpenAI API version. For Azure OpenAI, this field is required, + and is the version of the Azure OpenAI service to utilize, specified by a date. + openai_deployment_name: + type: string + description: This field is only required for Azure OpenAI and is the name of the deployment resource for the Azure + OpenAI service. + openai_organization: + type: string + description: This is an optional field to specify the organization in OpenAI or Azure OpenAI. + description: Configs needed to create an OpenAI model route. + PaLmConfig: + type: object + properties: + palm_api_key: + type: string + palm_api_key_plaintext: + type: string + description: 'The PaLM API key provided as a plaintext string. If you prefer to reference your key using Databricks + Secrets, see `palm_api_key`. You must provide an API key using one of the following fields: `palm_api_key` or + `palm_api_key_plaintext`.' + PayloadTable: + type: object + properties: + name: + type: string + status: + type: string + status_message: + type: string + PtEndpointCoreConfig: + type: object + properties: + served_entities: + type: array + items: + $ref: '#/components/schemas/PtServedModel' + traffic_config: + $ref: '#/components/schemas/TrafficConfig' + PtServedModel: + type: object + properties: + entity_name: + type: string + provisioned_model_units: + type: integer + description: The number of model units to be provisioned. + burst_scaling_enabled: + type: boolean + description: Whether burst scaling is enabled. When enabled (default), the endpoint can automatically scale up beyond + provisioned capacity to handle traffic spikes. When disabled, the endpoint maintains fixed capacity at provisioned_model_units. + entity_version: + type: string + name: + type: string + description: The name of a served entity. It must be unique across an endpoint. A served entity name can consist + of alphanumeric characters, dashes, and underscores. If not specified for an external model, this field defaults + to external_model.name, with '.' and ':' replaced with '-', and if not specified for other entities, it defaults + to entity_name-entity_version. + required: + - entity_name + - provisioned_model_units + PutAiGatewayResponse: + type: object + properties: + fallback_config: + $ref: '#/components/schemas/FallbackConfig' + guardrails: + $ref: '#/components/schemas/AiGatewayGuardrails' + description: Configuration for AI Guardrails to prevent unwanted data and unsafe data in requests and responses. + inference_table_config: + $ref: '#/components/schemas/AiGatewayInferenceTableConfig' + description: Configuration for payload logging using inference tables. Use these tables to monitor and audit data + being sent to and received from model APIs and to improve model quality. + rate_limits: + type: array + items: + $ref: '#/components/schemas/AiGatewayRateLimit' + description: Configuration for rate limits which can be set to limit endpoint traffic. + usage_tracking_config: + $ref: '#/components/schemas/AiGatewayUsageTrackingConfig' + description: Configuration to enable usage tracking using system tables. These tables allow you to monitor operational + usage on endpoints and their associated costs. + PutResponse: + type: object + properties: + rate_limits: + type: array + items: + $ref: '#/components/schemas/RateLimit' + QueryEndpointResponse: + type: object + properties: + choices: + type: array + items: + $ref: '#/components/schemas/V1ResponseChoiceElement' + created: + type: integer + description: The timestamp in seconds when the query was created in Unix time returned by a __completions or chat + external/foundation model__ serving endpoint. + data: + type: array + items: + $ref: '#/components/schemas/EmbeddingsV1ResponseEmbeddingElement' + description: The list of the embeddings returned by the __embeddings external/foundation model__ serving endpoint. + id: + type: string + description: The ID of the query that may be returned by a __completions or chat external/foundation model__ serving + endpoint. + model: + type: string + description: The name of the __external/foundation model__ used for querying. This is the name of the model that + was specified in the endpoint config. + object: + $ref: '#/components/schemas/QueryEndpointResponseObject' + description: The type of object returned by the __external/foundation model__ serving endpoint, one of [text_completion, + chat.completion, list (of embeddings)]. + outputs: + type: array + items: + type: object + description: The outputs of the feature serving endpoint. + predictions: + type: array + items: + type: object + description: The predictions returned by the serving endpoint. + served-model-name: + type: string + description: The name of the served model that served the request. This is useful when there are multiple models + behind the same endpoint with traffic split. + usage: + $ref: '#/components/schemas/ExternalModelUsageElement' + description: The usage object that may be returned by the __external/foundation model__ serving endpoint. This contains + information about the number of tokens used in the prompt and response. + RateLimit: + type: object + properties: + calls: + type: integer + renewal_period: + $ref: '#/components/schemas/RateLimitRenewalPeriod' + description: Renewal period field for a serving endpoint rate limit. Currently, only 'minute' is supported. + key: + $ref: '#/components/schemas/RateLimitKey' + description: Key field for a serving endpoint rate limit. Currently, only 'user' and 'endpoint' are supported, with + 'endpoint' being the default if not specified. + required: + - calls + - renewal_period + Route: + type: object + properties: + traffic_percentage: + type: integer + served_entity_name: + type: string + served_model_name: + type: string + description: The name of the served model this route configures traffic for. + required: + - traffic_percentage + ServedEntityInput: + type: object + properties: + burst_scaling_enabled: + type: boolean + entity_name: + type: string + description: The name of the entity to be served. The entity may be a model in the Databricks Model Registry, a + model in the Unity Catalog (UC), or a function of type FEATURE_SPEC in the UC. If it is a UC object, the full + name of the object should be given in the form of **catalog_name.schema_name.model_name**. + entity_version: + type: string + environment_vars: + type: object + description: 'An object containing a set of optional, user-specified environment variable key-value pairs used for + serving this entity. Note: this is an experimental feature and subject to change. Example entity environment variables + that refer to Databricks secrets: `{"OPENAI_API_KEY": "{{secrets/my_scope/my_key}}", "DATABRICKS_TOKEN": "{{secrets/my_scope2/my_key2}}"}`' + external_model: + $ref: '#/components/schemas/ExternalModel' + description: 'The external model to be served. NOTE: Only one of external_model and (entity_name, entity_version, + workload_size, workload_type, and scale_to_zero_enabled) can be specified with the latter set being used for custom + model serving for a Databricks registered model. For an existing endpoint with external_model, it cannot be updated + to an endpoint without external_model. If the endpoint is created without external_model, users cannot update + it to add external_model later. The task type of all external models within an endpoint must be the same.' + instance_profile_arn: + type: string + description: ARN of the instance profile that the served entity uses to access AWS resources. + max_provisioned_concurrency: + type: integer + description: The maximum provisioned concurrency that the endpoint can scale up to. Do not use if workload_size + is specified. + max_provisioned_throughput: + type: integer + description: The maximum tokens per second that the endpoint can scale up to. + min_provisioned_concurrency: + type: integer + description: The minimum provisioned concurrency that the endpoint can scale down to. Do not use if workload_size + is specified. + min_provisioned_throughput: + type: integer + description: The minimum tokens per second that the endpoint can scale down to. + name: + type: string + description: The name of a served entity. It must be unique across an endpoint. A served entity name can consist + of alphanumeric characters, dashes, and underscores. If not specified for an external model, this field defaults + to external_model.name, with '.' and ':' replaced with '-', and if not specified for other entities, it defaults + to entity_name-entity_version. + provisioned_model_units: + type: integer + description: The number of model units provisioned. + scale_to_zero_enabled: + type: boolean + description: Whether the compute resources for the served entity should scale down to zero. + workload_size: + type: string + description: The workload size of the served entity. The workload size corresponds to a range of provisioned concurrency + that the compute autoscales between. A single unit of provisioned concurrency can process one request at a time. + Valid workload sizes are "Small" (4 - 4 provisioned concurrency), "Medium" (8 - 16 provisioned concurrency), and + "Large" (16 - 64 provisioned concurrency). Additional custom workload sizes can also be used when available in + the workspace. If scale-to-zero is enabled, the lower bound of the provisioned concurrency for each workload size + is 0. Do not use if min_provisioned_concurrency and max_provisioned_concurrency are specified. + workload_type: + $ref: '#/components/schemas/ServingModelWorkloadType' + description: 'The workload type of the served entity. The workload type selects which type of compute to use in + the endpoint. The default value for this parameter is "CPU". For deep learning workloads, GPU acceleration is + available by selecting workload types like GPU_SMALL and others. See the available [GPU types]. [GPU types]: https://docs.databricks.com/en/machine-learning/model-serving/create-manage-serving-endpoints.html#gpu-workload-types' + ServedEntityOutput: + type: object + properties: + burst_scaling_enabled: + type: boolean + creation_timestamp: + type: integer + creator: + type: string + entity_name: + type: string + description: The name of the entity to be served. The entity may be a model in the Databricks Model Registry, a + model in the Unity Catalog (UC), or a function of type FEATURE_SPEC in the UC. If it is a UC object, the full + name of the object should be given in the form of **catalog_name.schema_name.model_name**. + entity_version: + type: string + environment_vars: + type: object + description: 'An object containing a set of optional, user-specified environment variable key-value pairs used for + serving this entity. Note: this is an experimental feature and subject to change. Example entity environment variables + that refer to Databricks secrets: `{"OPENAI_API_KEY": "{{secrets/my_scope/my_key}}", "DATABRICKS_TOKEN": "{{secrets/my_scope2/my_key2}}"}`' + external_model: + $ref: '#/components/schemas/ExternalModel' + description: 'The external model to be served. NOTE: Only one of external_model and (entity_name, entity_version, + workload_size, workload_type, and scale_to_zero_enabled) can be specified with the latter set being used for custom + model serving for a Databricks registered model. For an existing endpoint with external_model, it cannot be updated + to an endpoint without external_model. If the endpoint is created without external_model, users cannot update + it to add external_model later. The task type of all external models within an endpoint must be the same.' + foundation_model: + $ref: '#/components/schemas/FoundationModel' + instance_profile_arn: + type: string + description: ARN of the instance profile that the served entity uses to access AWS resources. + max_provisioned_concurrency: + type: integer + description: The maximum provisioned concurrency that the endpoint can scale up to. Do not use if workload_size + is specified. + max_provisioned_throughput: + type: integer + description: The maximum tokens per second that the endpoint can scale up to. + min_provisioned_concurrency: + type: integer + description: The minimum provisioned concurrency that the endpoint can scale down to. Do not use if workload_size + is specified. + min_provisioned_throughput: + type: integer + description: The minimum tokens per second that the endpoint can scale down to. + name: + type: string + description: The name of a served entity. It must be unique across an endpoint. A served entity name can consist + of alphanumeric characters, dashes, and underscores. If not specified for an external model, this field defaults + to external_model.name, with '.' and ':' replaced with '-', and if not specified for other entities, it defaults + to entity_name-entity_version. + provisioned_model_units: + type: integer + description: The number of model units provisioned. + scale_to_zero_enabled: + type: boolean + description: Whether the compute resources for the served entity should scale down to zero. + state: + $ref: '#/components/schemas/ServedModelState' + workload_size: + type: string + description: The workload size of the served entity. The workload size corresponds to a range of provisioned concurrency + that the compute autoscales between. A single unit of provisioned concurrency can process one request at a time. + Valid workload sizes are "Small" (4 - 4 provisioned concurrency), "Medium" (8 - 16 provisioned concurrency), and + "Large" (16 - 64 provisioned concurrency). Additional custom workload sizes can also be used when available in + the workspace. If scale-to-zero is enabled, the lower bound of the provisioned concurrency for each workload size + is 0. Do not use if min_provisioned_concurrency and max_provisioned_concurrency are specified. + workload_type: + $ref: '#/components/schemas/ServingModelWorkloadType' + description: 'The workload type of the served entity. The workload type selects which type of compute to use in + the endpoint. The default value for this parameter is "CPU". For deep learning workloads, GPU acceleration is + available by selecting workload types like GPU_SMALL and others. See the available [GPU types]. [GPU types]: https://docs.databricks.com/en/machine-learning/model-serving/create-manage-serving-endpoints.html#gpu-workload-types' + ServedEntitySpec: + type: object + properties: + entity_name: + type: string + entity_version: + type: string + external_model: + $ref: '#/components/schemas/ExternalModel' + foundation_model: + $ref: '#/components/schemas/FoundationModel' + name: + type: string + ServedModelInput: + type: object + properties: + scale_to_zero_enabled: + type: boolean + model_name: + type: string + model_version: + type: string + burst_scaling_enabled: + type: boolean + description: Whether burst scaling is enabled. When enabled (default), the endpoint can automatically scale up beyond + provisioned capacity to handle traffic spikes. When disabled, the endpoint maintains fixed capacity at provisioned_model_units. + environment_vars: + type: object + description: 'An object containing a set of optional, user-specified environment variable key-value pairs used for + serving this entity. Note: this is an experimental feature and subject to change. Example entity environment variables + that refer to Databricks secrets: `{"OPENAI_API_KEY": "{{secrets/my_scope/my_key}}", "DATABRICKS_TOKEN": "{{secrets/my_scope2/my_key2}}"}`' + instance_profile_arn: + type: string + description: ARN of the instance profile that the served entity uses to access AWS resources. + max_provisioned_concurrency: + type: integer + description: The maximum provisioned concurrency that the endpoint can scale up to. Do not use if workload_size + is specified. + max_provisioned_throughput: + type: integer + description: The maximum tokens per second that the endpoint can scale up to. + min_provisioned_concurrency: + type: integer + description: The minimum provisioned concurrency that the endpoint can scale down to. Do not use if workload_size + is specified. + min_provisioned_throughput: + type: integer + description: The minimum tokens per second that the endpoint can scale down to. + name: + type: string + description: The name of a served entity. It must be unique across an endpoint. A served entity name can consist + of alphanumeric characters, dashes, and underscores. If not specified for an external model, this field defaults + to external_model.name, with '.' and ':' replaced with '-', and if not specified for other entities, it defaults + to entity_name-entity_version. + provisioned_model_units: + type: integer + description: The number of model units provisioned. + workload_size: + type: string + description: The workload size of the served entity. The workload size corresponds to a range of provisioned concurrency + that the compute autoscales between. A single unit of provisioned concurrency can process one request at a time. + Valid workload sizes are "Small" (4 - 4 provisioned concurrency), "Medium" (8 - 16 provisioned concurrency), and + "Large" (16 - 64 provisioned concurrency). Additional custom workload sizes can also be used when available in + the workspace. If scale-to-zero is enabled, the lower bound of the provisioned concurrency for each workload size + is 0. Do not use if min_provisioned_concurrency and max_provisioned_concurrency are specified. + workload_type: + $ref: '#/components/schemas/ServedModelInputWorkloadType' + description: 'The workload type of the served entity. The workload type selects which type of compute to use in + the endpoint. The default value for this parameter is "CPU". For deep learning workloads, GPU acceleration is + available by selecting workload types like GPU_SMALL and others. See the available [GPU types]. [GPU types]: https://docs.databricks.com/en/machine-learning/model-serving/create-manage-serving-endpoints.html#gpu-workload-types' + required: + - scale_to_zero_enabled + - model_name + - model_version + ServedModelOutput: + type: object + properties: + burst_scaling_enabled: + type: boolean + creation_timestamp: + type: integer + creator: + type: string + environment_vars: + type: object + description: 'An object containing a set of optional, user-specified environment variable key-value pairs used for + serving this entity. Note: this is an experimental feature and subject to change. Example entity environment variables + that refer to Databricks secrets: `{"OPENAI_API_KEY": "{{secrets/my_scope/my_key}}", "DATABRICKS_TOKEN": "{{secrets/my_scope2/my_key2}}"}`' + instance_profile_arn: + type: string + description: ARN of the instance profile that the served entity uses to access AWS resources. + max_provisioned_concurrency: + type: integer + description: The maximum provisioned concurrency that the endpoint can scale up to. Do not use if workload_size + is specified. + min_provisioned_concurrency: + type: integer + description: The minimum provisioned concurrency that the endpoint can scale down to. Do not use if workload_size + is specified. + model_name: + type: string + model_version: + type: string + name: + type: string + description: The name of a served entity. It must be unique across an endpoint. A served entity name can consist + of alphanumeric characters, dashes, and underscores. If not specified for an external model, this field defaults + to external_model.name, with '.' and ':' replaced with '-', and if not specified for other entities, it defaults + to entity_name-entity_version. + provisioned_model_units: + type: integer + description: The number of model units provisioned. + scale_to_zero_enabled: + type: boolean + description: Whether the compute resources for the served entity should scale down to zero. + state: + $ref: '#/components/schemas/ServedModelState' + workload_size: + type: string + description: The workload size of the served entity. The workload size corresponds to a range of provisioned concurrency + that the compute autoscales between. A single unit of provisioned concurrency can process one request at a time. + Valid workload sizes are "Small" (4 - 4 provisioned concurrency), "Medium" (8 - 16 provisioned concurrency), and + "Large" (16 - 64 provisioned concurrency). Additional custom workload sizes can also be used when available in + the workspace. If scale-to-zero is enabled, the lower bound of the provisioned concurrency for each workload size + is 0. Do not use if min_provisioned_concurrency and max_provisioned_concurrency are specified. + workload_type: + $ref: '#/components/schemas/ServingModelWorkloadType' + description: 'The workload type of the served entity. The workload type selects which type of compute to use in + the endpoint. The default value for this parameter is "CPU". For deep learning workloads, GPU acceleration is + available by selecting workload types like GPU_SMALL and others. See the available [GPU types]. [GPU types]: https://docs.databricks.com/en/machine-learning/model-serving/create-manage-serving-endpoints.html#gpu-workload-types' + ServedModelSpec: + type: object + properties: + model_name: + type: string + model_version: + type: string + description: Only one of model_version and entity_version should be populated + name: + type: string + ServedModelState: + type: object + properties: + deployment: + $ref: '#/components/schemas/ServedModelStateDeployment' + deployment_state_message: + type: string + ServerLogsResponse: + type: object + properties: + logs: + type: string + required: + - logs + ServingEndpoint: + type: object + properties: + ai_gateway: + $ref: '#/components/schemas/AiGatewayConfig' + budget_policy_id: + type: string + description: The budget policy associated with the endpoint. + config: + $ref: '#/components/schemas/EndpointCoreConfigSummary' + description: The config that is currently being served by the endpoint. + creation_timestamp: + type: integer + description: The timestamp when the endpoint was created in Unix time. + creator: + type: string + description: The email of the user who created the serving endpoint. + description: + type: string + description: Description of the endpoint + id: + type: string + description: System-generated ID of the endpoint, included to be used by the Permissions API. + last_updated_timestamp: + type: integer + description: The timestamp when the endpoint was last updated by a user in Unix time. + name: + type: string + description: The name of the serving endpoint. + state: + $ref: '#/components/schemas/EndpointState' + description: Information corresponding to the state of the serving endpoint. + tags: + type: array + items: + $ref: '#/components/schemas/EndpointTag' + description: Tags attached to the serving endpoint. + task: + type: string + description: The task type of the serving endpoint. + usage_policy_id: + type: string + description: The usage policy associated with serving endpoint. + ServingEndpointAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/ServingEndpointPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + ServingEndpointAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/ServingEndpointPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + ServingEndpointDetailed: + type: object + properties: + ai_gateway: + $ref: '#/components/schemas/AiGatewayConfig' + budget_policy_id: + type: string + description: The budget policy associated with the endpoint. + config: + $ref: '#/components/schemas/EndpointCoreConfigOutput' + description: The config that is currently being served by the endpoint. + creation_timestamp: + type: integer + description: The timestamp when the endpoint was created in Unix time. + creator: + type: string + description: The email of the user who created the serving endpoint. + data_plane_info: + $ref: '#/components/schemas/ModelDataPlaneInfo' + description: Information required to query DataPlane APIs. + description: + type: string + description: Description of the serving model + email_notifications: + $ref: '#/components/schemas/EmailNotifications' + description: Email notification settings. + endpoint_url: + type: string + description: Endpoint invocation url if route optimization is enabled for endpoint + id: + type: string + description: System-generated ID of the endpoint. This is used to refer to the endpoint in the Permissions API + last_updated_timestamp: + type: integer + description: The timestamp when the endpoint was last updated by a user in Unix time. + name: + type: string + description: The name of the serving endpoint. + pending_config: + $ref: '#/components/schemas/EndpointPendingConfig' + description: The config that the endpoint is attempting to update to. + permission_level: + $ref: '#/components/schemas/ServingEndpointDetailedPermissionLevel' + description: The permission level of the principal making the request. + route_optimized: + type: boolean + description: Boolean representing if route optimization has been enabled for the endpoint + state: + $ref: '#/components/schemas/EndpointState' + description: Information corresponding to the state of the serving endpoint. + tags: + type: array + items: + $ref: '#/components/schemas/EndpointTag' + description: Tags attached to the serving endpoint. + task: + type: string + description: The task type of the serving endpoint. + ServingEndpointPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/ServingEndpointPermissionLevel' + ServingEndpointPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/ServingEndpointAccessControlResponse' + object_id: + type: string + object_type: + type: string + ServingEndpointPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/ServingEndpointPermissionLevel' + TrafficConfig: + type: object + properties: + routes: + type: array + items: + $ref: '#/components/schemas/Route' + UpdateInferenceEndpointNotificationsResponse: + type: object + properties: + email_notifications: + $ref: '#/components/schemas/EmailNotifications' + name: + type: string + V1ResponseChoiceElement: + type: object + properties: + finishReason: + type: string + index: + type: integer + description: The index of the choice in the __chat or completions__ response. + logprobs: + type: integer + description: The logprobs returned only by the __completions__ endpoint. + message: + $ref: '#/components/schemas/ChatMessage' + description: The message response from the __chat__ endpoint. + text: + type: string + description: The text response from the __completions__ endpoint. + AiGatewayGuardrailPiiBehaviorBehavior: + type: string + enum: + - BLOCK + - MASK + - NONE + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + AiGatewayRateLimitKey: + type: string + enum: + - endpoint + - service_principal + - user + - user_group + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + AiGatewayRateLimitRenewalPeriod: + type: string + enum: + - minute + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + AmazonBedrockConfigBedrockProvider: + type: string + enum: + - ai21labs + - amazon + - anthropic + - cohere + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + ChatMessageRole: + type: string + enum: + - assistant + - system + - user + description: The role of the message. One of [system, user, assistant]. + EmbeddingsV1ResponseEmbeddingElementObject: + type: string + enum: + - embedding + description: This will always be 'embedding'. + EndpointStateConfigUpdate: + type: string + enum: + - IN_PROGRESS + - NOT_UPDATING + - UPDATE_CANCELED + - UPDATE_FAILED + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + EndpointStateReady: + type: string + enum: + - NOT_READY + - READY + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + ExternalFunctionRequestHttpMethod: + type: string + enum: + - DELETE + - GET + - PATCH + - POST + - PUT + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + ExternalModelProvider: + type: string + enum: + - ai21labs + - amazon-bedrock + - anthropic + - cohere + - custom + - databricks-model-serving + - google-cloud-vertex-ai + - openai + - palm + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + QueryEndpointResponseObject: + type: string + enum: + - chat.completion + - list + - text_completion + description: 'The type of object returned by the __external/foundation model__ serving endpoint, one of + + [text_completion, chat.completion, list (of embeddings)].' + RateLimitKey: + type: string + enum: + - endpoint + - user + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + RateLimitRenewalPeriod: + type: string + enum: + - minute + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + ServedModelInputWorkloadType: + type: string + enum: + - CPU + - GPU_LARGE + - GPU_MEDIUM + - GPU_SMALL + - MULTIGPU_MEDIUM + description: Please keep this in sync with with workload types in InferenceEndpointEntities.scala + ServedModelStateDeployment: + type: string + enum: + - DEPLOYMENT_ABORTED + - DEPLOYMENT_CREATING + - DEPLOYMENT_FAILED + - DEPLOYMENT_READY + - DEPLOYMENT_RECOVERING + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + ServingEndpointDetailedPermissionLevel: + type: string + enum: + - CAN_MANAGE + - CAN_QUERY + - CAN_VIEW + description: "Create a collection of name/value pairs.\n\nExample enumeration:\n\n>>> class Color(Enum):\n... RED\ + \ = 1\n... BLUE = 2\n... GREEN = 3\n\nAccess them by:\n\n- attribute access:\n\n >>> Color.RED\n \n\n- value lookup:\n\n >>> Color(1)\n \n\n- name lookup:\n\n >>> Color['RED']\n \n\nEnumerations can be iterated over, and know how many members they have:\n\n>>> len(Color)\n3\n\n>>> list(Color)\n\ + [, , ]\n\nMethods can be added to enumerations, and members can have\ + \ their own\nattributes -- see the documentation for details." + ServingEndpointPermissionLevel: + type: string + enum: + - CAN_MANAGE + - CAN_QUERY + - CAN_VIEW + description: Permission level + ServingModelWorkloadType: + type: string + enum: + - CPU + - GPU_LARGE + - GPU_MEDIUM + - GPU_SMALL + - MULTIGPU_MEDIUM + description: Please keep this in sync with with workload types in InferenceEndpointEntities.scala + x-stackQL-resources: + serving_endpoint_build_logs: + id: databricks_workspace.serving.serving_endpoint_build_logs + name: serving_endpoint_build_logs + title: Serving Endpoint Build Logs + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1served-models~1{served_model_name}~1build-logs/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_build_logs/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + serving_endpoints: + id: databricks_workspace.serving.serving_endpoints + name: serving_endpoints + title: Serving Endpoints + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.endpoints + delete: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1tags/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + query: + operation: + $ref: '#/paths/~1serving-endpoints~1{name}~1invocations/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update_config: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1config/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/get' + - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/serving_endpoints/methods/update_config' + serving_endpoints_pt: + id: databricks_workspace.serving.serving_endpoints_pt + name: serving_endpoints_pt + title: Serving Endpoints Pt + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1pt/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update_config: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1pt~1{name}~1config/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: '#/components/x-stackQL-resources/serving_endpoints_pt/methods/create' + update: [] + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/serving_endpoints_pt/methods/update_config' + serving_endpoint_metrics: + id: databricks_workspace.serving.serving_endpoint_metrics + name: serving_endpoint_metrics + title: Serving Endpoint Metrics + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1metrics/get' + response: + mediaType: text/plain + openAPIDocKey: '200' + overrideMediaType: application/json + transform: + body: '[{"contents": {{ toJson . }}}]' + type: golang_template_text_v0.3.0 + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_metrics/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + serving_endpoint_openapi: + id: databricks_workspace.serving.serving_endpoint_openapi + name: serving_endpoint_openapi + title: Serving Endpoint Openapi + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1openapi/get' + response: + mediaType: text/plain + openAPIDocKey: '200' + overrideMediaType: application/json + transform: + body: '[{"contents": {{ toJson . }}}]' + type: golang_template_text_v0.3.0 + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_openapi/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + serving_endpoint_permission_levels: + id: databricks_workspace.serving.serving_endpoint_permission_levels + name: serving_endpoint_permission_levels + title: Serving Endpoint Permission Levels + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}~1permissionLevels/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_permission_levels/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + serving_endpoint_permissions: + id: databricks_workspace.serving.serving_endpoint_permissions + name: serving_endpoint_permissions + title: Serving Endpoint Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1serving-endpoints~1{serving_endpoint_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_permissions/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_permissions/methods/update' + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_permissions/methods/set' + serving_endpoints_http_request: + id: databricks_workspace.serving.serving_endpoints_http_request + name: serving_endpoints_http_request + title: Serving Endpoints Http Request + methods: + serving_endpoints_http_request: + operation: + $ref: '#/paths/~1api~12.0~1external-function/post' + response: + mediaType: text/plain + openAPIDocKey: '200' + overrideMediaType: application/json + transform: + body: '[{"contents": {{ toJson . }}}]' + type: golang_template_text_v0.3.0 + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoints_http_request/methods/serving_endpoints_http_request' + insert: [] + update: [] + delete: [] + replace: [] + serving_endpoint_logs: + id: databricks_workspace.serving.serving_endpoint_logs + name: serving_endpoint_logs + title: Serving Endpoint Logs + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1served-models~1{served_model_name}~1logs/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_logs/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + serving_endpoint_rate_limits: + id: databricks_workspace.serving.serving_endpoint_rate_limits + name: serving_endpoint_rate_limits + title: Serving Endpoint Rate Limits + methods: + update: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1rate-limits/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: [] + update: [] + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_rate_limits/methods/update' + serving_endpoint_ai_gateway: + id: databricks_workspace.serving.serving_endpoint_ai_gateway + name: serving_endpoint_ai_gateway + title: Serving Endpoint Ai Gateway + methods: + update: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1ai-gateway/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: [] + update: [] + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_ai_gateway/methods/update' + serving_endpoint_notifications: + id: databricks_workspace.serving.serving_endpoint_notifications + name: serving_endpoint_notifications + title: Serving Endpoint Notifications + methods: + update: + operation: + $ref: '#/paths/~1api~12.0~1serving-endpoints~1{name}~1notifications/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/serving_endpoint_notifications/methods/update' + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/settings.yaml b/providers/src/databricks_workspace/v00.00.00000/services/settings.yaml new file mode 100644 index 00000000..c22959d6 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/settings.yaml @@ -0,0 +1,8851 @@ +openapi: 3.0.0 +info: + title: Databricks Settings API (workspace) + description: >- + OpenAPI specification for the Databricks settings service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.settings +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/settings/types/aibi_dash_embed_ws_acc_policy/names/default: + delete: + operationId: aibi_dashboard_embedding_access_policy_delete + summary: >- + Delete the AI/BI dashboard embedding access policy, reverting back to + the default. + tags: + - settings + - aibi_dashboard_embedding_access_policy + description: >- + Delete the AI/BI dashboard embedding access policy, reverting back to + the default. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: + :class:`DeleteAibiDashboardEmbeddingAccessPolicySettingResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/DeleteAibiDashboardEmbeddingAccessPolicySettingResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AibiDashboardEmbeddingAccessPolicyAPI + get: + operationId: aibi_dashboard_embedding_access_policy_get + summary: >- + Retrieves the AI/BI dashboard embedding access policy. The default + setting is ALLOW_APPROVED_DOMAINS, + tags: + - settings + - aibi_dashboard_embedding_access_policy + description: >- + Retrieves the AI/BI dashboard embedding access policy. The default + setting is ALLOW_APPROVED_DOMAINS, + + permitting AI/BI dashboards to be embedded on approved domains. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`AibiDashboardEmbeddingAccessPolicySetting` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicySetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AibiDashboardEmbeddingAccessPolicyAPI + patch: + operationId: aibi_dashboard_embedding_access_policy_update + summary: >- + Updates the AI/BI dashboard embedding access policy at the workspace + level. + tags: + - settings + - aibi_dashboard_embedding_access_policy + description: >- + Updates the AI/BI dashboard embedding access policy at the workspace + level. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`AibiDashboardEmbeddingAccessPolicySetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`AibiDashboardEmbeddingAccessPolicySetting` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicySetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AibiDashboardEmbeddingAccessPolicyAPI + /api/2.0/settings/types/aibi_dash_embed_ws_apprvd_domains/names/default: + delete: + operationId: aibi_dashboard_embedding_approved_domains_delete + summary: >- + Delete the list of domains approved to host embedded AI/BI dashboards, + reverting back to the default + tags: + - settings + - aibi_dashboard_embedding_approved_domains + description: >- + Delete the list of domains approved to host embedded AI/BI dashboards, + reverting back to the default + + empty list. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: + :class:`DeleteAibiDashboardEmbeddingApprovedDomainsSettingResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/DeleteAibiDashboardEmbeddingApprovedDomainsSettingResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AibiDashboardEmbeddingApprovedDomainsAPI + get: + operationId: aibi_dashboard_embedding_approved_domains_get + summary: >- + Retrieves the list of domains approved to host embedded AI/BI + dashboards. + tags: + - settings + - aibi_dashboard_embedding_approved_domains + description: >- + Retrieves the list of domains approved to host embedded AI/BI + dashboards. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`AibiDashboardEmbeddingApprovedDomainsSetting` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/AibiDashboardEmbeddingApprovedDomainsSetting + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AibiDashboardEmbeddingApprovedDomainsAPI + patch: + operationId: aibi_dashboard_embedding_approved_domains_update + summary: >- + Updates the list of domains approved to host embedded AI/BI dashboards. + This update will fail if the + tags: + - settings + - aibi_dashboard_embedding_approved_domains + description: >- + Updates the list of domains approved to host embedded AI/BI dashboards. + This update will fail if the + + current workspace access policy is not ALLOW_APPROVED_DOMAINS. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`AibiDashboardEmbeddingApprovedDomainsSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`AibiDashboardEmbeddingApprovedDomainsSetting` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/AibiDashboardEmbeddingApprovedDomainsSetting + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AibiDashboardEmbeddingApprovedDomainsAPI + /api/2.0/settings/types/automatic_cluster_update/names/default: + get: + operationId: automatic_cluster_update_get + summary: Gets the automatic cluster update setting. + tags: + - settings + - automatic_cluster_update + description: |- + Gets the automatic cluster update setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`AutomaticClusterUpdateSetting` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AutomaticClusterUpdateSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AutomaticClusterUpdateAPI + patch: + operationId: automatic_cluster_update_update + summary: >- + Updates the automatic cluster update setting for the workspace. A fresh + etag needs to be provided in + tags: + - settings + - automatic_cluster_update + description: >- + Updates the automatic cluster update setting for the workspace. A fresh + etag needs to be provided in + + `PATCH` requests (as part of the setting field). The etag can be + retrieved by making a `GET` request + + before the `PATCH` request. If the setting is updated concurrently, + `PATCH` fails with 409 and the + + request must be retried by using the fresh etag in the 409 response. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`AutomaticClusterUpdateSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`AutomaticClusterUpdateSetting` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AutomaticClusterUpdateSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AutomaticClusterUpdateAPI + /api/2.0/settings/types/shield_csp_enablement_ws_db/names/default: + get: + operationId: compliance_security_profile_get + summary: Gets the compliance security profile setting. + tags: + - settings + - compliance_security_profile + description: |- + Gets the compliance security profile setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`ComplianceSecurityProfileSetting` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ComplianceSecurityProfileSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ComplianceSecurityProfileAPI + patch: + operationId: compliance_security_profile_update + summary: >- + Updates the compliance security profile setting for the workspace. A + fresh etag needs to be provided + tags: + - settings + - compliance_security_profile + description: >- + Updates the compliance security profile setting for the workspace. A + fresh etag needs to be provided + + in `PATCH` requests (as part of the setting field). The etag can be + retrieved by making a `GET` + + request before the `PATCH` request. If the setting is updated + concurrently, `PATCH` fails with 409 and + + the request must be retried by using the fresh etag in the 409 response. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`ComplianceSecurityProfileSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`ComplianceSecurityProfileSetting` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ComplianceSecurityProfileSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ComplianceSecurityProfileAPI + /api/2.0/credentials-manager/exchange-tokens/token: + post: + operationId: credentials_manager_exchange_token + summary: >- + Exchange tokens with an Identity Provider to get a new access token. It + allows specifying scopes to + tags: + - settings + - credentials_manager + description: >- + Exchange tokens with an Identity Provider to get a new access token. It + allows specifying scopes to + + determine token permissions. + + + :param partition_id: :class:`PartitionId` + The partition of Credentials store + :param token_type: List[:class:`TokenType`] + A list of token types being requested + :param scopes: List[str] + Array of scopes for the token request. + + :returns: :class:`ExchangeTokenResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + partition_id: + type: string + description: The partition of Credentials store + token_type: + type: string + description: A list of token types being requested + scopes: + type: string + description: Array of scopes for the token request. + required: + - partition_id + - token_type + - scopes + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExchangeTokenResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CredentialsManagerAPI + /api/2.0/accounts/{account_id}/settings/types/shield_csp_enablement_ac/names/default: + get: + operationId: csp_enablement_account_get + summary: Gets the compliance security profile setting for new workspaces. + tags: + - settings + - csp_enablement_account + description: |- + Gets the compliance security profile setting for new workspaces. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`CspEnablementAccountSetting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CspEnablementAccountSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CspEnablementAccountAPI + patch: + operationId: csp_enablement_account_update + summary: >- + Updates the value of the compliance security profile setting for new + workspaces. + tags: + - settings + - csp_enablement_account + description: >- + Updates the value of the compliance security profile setting for new + workspaces. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`CspEnablementAccountSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`CspEnablementAccountSetting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CspEnablementAccountSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: CspEnablementAccountAPI + /api/2.0/settings/types/dashboard_email_subscriptions/names/default: + delete: + operationId: dashboard_email_subscriptions_delete + summary: Reverts the Dashboard Email Subscriptions setting to its default value. + tags: + - settings + - dashboard_email_subscriptions + description: |- + Reverts the Dashboard Email Subscriptions setting to its default value. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteDashboardEmailSubscriptionsResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteDashboardEmailSubscriptionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardEmailSubscriptionsAPI + get: + operationId: dashboard_email_subscriptions_get + summary: Gets the Dashboard Email Subscriptions setting. + tags: + - settings + - dashboard_email_subscriptions + description: |- + Gets the Dashboard Email Subscriptions setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DashboardEmailSubscriptions` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DashboardEmailSubscriptions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardEmailSubscriptionsAPI + patch: + operationId: dashboard_email_subscriptions_update + summary: Updates the Dashboard Email Subscriptions setting. + tags: + - settings + - dashboard_email_subscriptions + description: |- + Updates the Dashboard Email Subscriptions setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`DashboardEmailSubscriptions` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`DashboardEmailSubscriptions` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DashboardEmailSubscriptions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardEmailSubscriptionsAPI + /api/2.0/settings/types/default_namespace_ws/names/default: + delete: + operationId: default_namespace_delete + summary: >- + Deletes the default namespace setting for the workspace. A fresh etag + needs to be provided in `DELETE` + tags: + - settings + - default_namespace + description: >- + Deletes the default namespace setting for the workspace. A fresh etag + needs to be provided in `DELETE` + + requests (as a query parameter). The etag can be retrieved by making a + `GET` request before the + + `DELETE` request. If the setting is updated/deleted concurrently, + `DELETE` fails with 409 and the + + request must be retried by using the fresh etag in the 409 response. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteDefaultNamespaceSettingResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteDefaultNamespaceSettingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DefaultNamespaceAPI + get: + operationId: default_namespace_get + summary: Gets the default namespace setting. + tags: + - settings + - default_namespace + description: |- + Gets the default namespace setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DefaultNamespaceSetting` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DefaultNamespaceSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DefaultNamespaceAPI + patch: + operationId: default_namespace_update + summary: >- + Updates the default namespace setting for the workspace. A fresh etag + needs to be provided in `PATCH` + tags: + - settings + - default_namespace + description: >- + Updates the default namespace setting for the workspace. A fresh etag + needs to be provided in `PATCH` + + requests (as part of the setting field). The etag can be retrieved by + making a `GET` request before + + the `PATCH` request. Note that if the setting does not exist, `GET` + returns a NOT_FOUND error and the + + etag is present in the error response, which should be set in the + `PATCH` request. If the setting is + + updated concurrently, `PATCH` fails with 409 and the request must be + retried by using the fresh etag + + in the 409 response. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`DefaultNamespaceSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`DefaultNamespaceSetting` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DefaultNamespaceSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DefaultNamespaceAPI + /api/2.0/settings/types/default_warehouse_id/names/default: + delete: + operationId: default_warehouse_id_delete + summary: Reverts the Default Warehouse Id setting to its default value. + tags: + - settings + - default_warehouse_id + description: |- + Reverts the Default Warehouse Id setting to its default value. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteDefaultWarehouseIdResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteDefaultWarehouseIdResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DefaultWarehouseIdAPI + get: + operationId: default_warehouse_id_get + summary: Gets the Default Warehouse Id setting. + tags: + - settings + - default_warehouse_id + description: |- + Gets the Default Warehouse Id setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DefaultWarehouseId` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DefaultWarehouseId' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DefaultWarehouseIdAPI + patch: + operationId: default_warehouse_id_update + summary: Updates the Default Warehouse Id setting. + tags: + - settings + - default_warehouse_id + description: |- + Updates the Default Warehouse Id setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`DefaultWarehouseId` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`DefaultWarehouseId` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DefaultWarehouseId' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DefaultWarehouseIdAPI + /api/2.0/settings/types/disable_legacy_access/names/default: + delete: + operationId: disable_legacy_access_delete + summary: Deletes legacy access disablement status. + tags: + - settings + - disable_legacy_access + description: |- + Deletes legacy access disablement status. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteDisableLegacyAccessResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteDisableLegacyAccessResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyAccessAPI + get: + operationId: disable_legacy_access_get + summary: Retrieves legacy access disablement Status. + tags: + - settings + - disable_legacy_access + description: |- + Retrieves legacy access disablement Status. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DisableLegacyAccess` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DisableLegacyAccess' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyAccessAPI + patch: + operationId: disable_legacy_access_update + summary: Updates legacy access disablement status. + tags: + - settings + - disable_legacy_access + description: |- + Updates legacy access disablement status. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`DisableLegacyAccess` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`DisableLegacyAccess` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DisableLegacyAccess' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyAccessAPI + /api/2.0/settings/types/disable_legacy_dbfs/names/default: + delete: + operationId: disable_legacy_dbfs_delete + summary: >- + Deletes the disable legacy DBFS setting for a workspace, reverting back + to the default. + tags: + - settings + - disable_legacy_dbfs + description: >- + Deletes the disable legacy DBFS setting for a workspace, reverting back + to the default. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteDisableLegacyDbfsResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteDisableLegacyDbfsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyDbfsAPI + get: + operationId: disable_legacy_dbfs_get + summary: Gets the disable legacy DBFS setting. + tags: + - settings + - disable_legacy_dbfs + description: |- + Gets the disable legacy DBFS setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DisableLegacyDbfs` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DisableLegacyDbfs' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyDbfsAPI + patch: + operationId: disable_legacy_dbfs_update + summary: Updates the disable legacy DBFS setting for the workspace. + tags: + - settings + - disable_legacy_dbfs + description: |- + Updates the disable legacy DBFS setting for the workspace. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`DisableLegacyDbfs` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`DisableLegacyDbfs` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DisableLegacyDbfs' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyDbfsAPI + /api/2.0/accounts/{account_id}/settings/types/disable_legacy_features/names/default: + delete: + operationId: disable_legacy_features_delete + summary: Deletes the disable legacy features setting. + tags: + - settings + - disable_legacy_features + description: |- + Deletes the disable legacy features setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteDisableLegacyFeaturesResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteDisableLegacyFeaturesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyFeaturesAPI + get: + operationId: disable_legacy_features_get + summary: Gets the value of the disable legacy features setting. + tags: + - settings + - disable_legacy_features + description: |- + Gets the value of the disable legacy features setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DisableLegacyFeatures` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DisableLegacyFeatures' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyFeaturesAPI + patch: + operationId: disable_legacy_features_update + summary: Updates the value of the disable legacy features setting. + tags: + - settings + - disable_legacy_features + description: |- + Updates the value of the disable legacy features setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`DisableLegacyFeatures` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`DisableLegacyFeatures` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DisableLegacyFeatures' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DisableLegacyFeaturesAPI + /api/2.0/settings/types/enable-export-notebook/names/default: + get: + operationId: enable_export_notebook_get_enable_export_notebook + summary: Gets the Notebook and File exporting setting. + tags: + - settings + - enable_export_notebook + description: |- + Gets the Notebook and File exporting setting. + + + :returns: :class:`EnableExportNotebook` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnableExportNotebook' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableExportNotebookAPI + patch: + operationId: enable_export_notebook_patch_enable_export_notebook + summary: >- + Updates the Notebook and File exporting setting. The model follows + eventual consistency, which means + tags: + - settings + - enable_export_notebook + description: >- + Updates the Notebook and File exporting setting. The model follows + eventual consistency, which means + + the get after the update operation might receive stale values for some + time. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`EnableExportNotebook` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`EnableExportNotebook` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnableExportNotebook' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableExportNotebookAPI + /api/2.0/accounts/{account_id}/settings/types/acct_ip_acl_enable/names/default: + delete: + operationId: enable_ip_access_lists_delete + summary: >- + Reverts the value of the account IP access toggle setting to default + (ON) + tags: + - settings + - enable_ip_access_lists + description: >- + Reverts the value of the account IP access toggle setting to default + (ON) + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteAccountIpAccessEnableResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteAccountIpAccessEnableResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableIpAccessListsAPI + get: + operationId: enable_ip_access_lists_get + summary: Gets the value of the account IP access toggle setting. + tags: + - settings + - enable_ip_access_lists + description: |- + Gets the value of the account IP access toggle setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`AccountIpAccessEnable` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountIpAccessEnable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableIpAccessListsAPI + patch: + operationId: enable_ip_access_lists_update + summary: Updates the value of the account IP access toggle setting. + tags: + - settings + - enable_ip_access_lists + description: |- + Updates the value of the account IP access toggle setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`AccountIpAccessEnable` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`AccountIpAccessEnable` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AccountIpAccessEnable' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableIpAccessListsAPI + /api/2.0/settings/types/enable-notebook-table-clipboard/names/default: + get: + operationId: enable_notebook_table_clipboard_get_enable_notebook_table_clipboard + summary: Gets the Results Table Clipboard features setting. + tags: + - settings + - enable_notebook_table_clipboard + description: |- + Gets the Results Table Clipboard features setting. + + + :returns: :class:`EnableNotebookTableClipboard` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnableNotebookTableClipboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableNotebookTableClipboardAPI + patch: + operationId: enable_notebook_table_clipboard_patch_enable_notebook_table_clipboard + summary: >- + Updates the Results Table Clipboard features setting. The model follows + eventual consistency, which + tags: + - settings + - enable_notebook_table_clipboard + description: >- + Updates the Results Table Clipboard features setting. The model follows + eventual consistency, which + + means the get after the update operation might receive stale values for + some time. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`EnableNotebookTableClipboard` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`EnableNotebookTableClipboard` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnableNotebookTableClipboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableNotebookTableClipboardAPI + /api/2.0/settings/types/enable-results-downloading/names/default: + get: + operationId: enable_results_downloading_get_enable_results_downloading + summary: Gets the Notebook results download setting. + tags: + - settings + - enable_results_downloading + description: |- + Gets the Notebook results download setting. + + + :returns: :class:`EnableResultsDownloading` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnableResultsDownloading' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableResultsDownloadingAPI + patch: + operationId: enable_results_downloading_patch_enable_results_downloading + summary: >- + Updates the Notebook results download setting. The model follows + eventual consistency, which means the + tags: + - settings + - enable_results_downloading + description: >- + Updates the Notebook results download setting. The model follows + eventual consistency, which means the + + get after the update operation might receive stale values for some time. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`EnableResultsDownloading` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`EnableResultsDownloading` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnableResultsDownloading' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnableResultsDownloadingAPI + /api/2.0/settings/types/shield_esm_enablement_ws_db/names/default: + get: + operationId: enhanced_security_monitoring_get + summary: Gets the enhanced security monitoring setting. + tags: + - settings + - enhanced_security_monitoring + description: |- + Gets the enhanced security monitoring setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`EnhancedSecurityMonitoringSetting` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnhancedSecurityMonitoringSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnhancedSecurityMonitoringAPI + patch: + operationId: enhanced_security_monitoring_update + summary: >- + Updates the enhanced security monitoring setting for the workspace. A + fresh etag needs to be provided + tags: + - settings + - enhanced_security_monitoring + description: >- + Updates the enhanced security monitoring setting for the workspace. A + fresh etag needs to be provided + + in `PATCH` requests (as part of the setting field). The etag can be + retrieved by making a `GET` + + request before the `PATCH` request. If the setting is updated + concurrently, `PATCH` fails with 409 and + + the request must be retried by using the fresh etag in the 409 response. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`EnhancedSecurityMonitoringSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`EnhancedSecurityMonitoringSetting` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EnhancedSecurityMonitoringSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EnhancedSecurityMonitoringAPI + /api/2.0/accounts/{account_id}/settings/types/shield_esm_enablement_ac/names/default: + get: + operationId: esm_enablement_account_get + summary: Gets the enhanced security monitoring setting for new workspaces. + tags: + - settings + - esm_enablement_account + description: |- + Gets the enhanced security monitoring setting for new workspaces. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`EsmEnablementAccountSetting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EsmEnablementAccountSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EsmEnablementAccountAPI + patch: + operationId: esm_enablement_account_update + summary: >- + Updates the value of the enhanced security monitoring setting for new + workspaces. + tags: + - settings + - esm_enablement_account + description: >- + Updates the value of the enhanced security monitoring setting for new + workspaces. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`EsmEnablementAccountSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`EsmEnablementAccountSetting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EsmEnablementAccountSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: EsmEnablementAccountAPI + /api/2.0/ip-access-lists: + post: + operationId: ip_access_lists_create + summary: Creates an IP access list for this workspace. + tags: + - settings + - ip_access_lists + description: >- + Creates an IP access list for this workspace. + + + A list can be an allow list or a block list. See the top of this file + for a description of how the + + server treats allow lists and block lists at runtime. + + + When creating or updating an IP access list: + + + * For all allow lists and block lists combined, the API supports a + maximum of 1000 IP/CIDR values, + + where one CIDR counts as a single value. Attempts to exceed that number + return error 400 with + + `error_code` value `QUOTA_EXCEEDED`. * If the new list would block the + calling user's current IP, + + error 400 is returned with `error_code` value `INVALID_STATE`. + + + It can take a few minutes for the changes to take effect. **Note**: Your + new IP access list has no + + effect until you enable the feature. See :method:workspaceconf/setStatus + + + :param label: str + Label for the IP access list. This **cannot** be empty. + :param list_type: :class:`ListType` + + :param ip_addresses: List[str] (optional) + + + :returns: :class:`CreateIpAccessListResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + label: + type: string + description: Label for the IP access list. This **cannot** be empty. + list_type: + type: string + description: ':param ip_addresses: List[str] (optional)' + ip_addresses: + type: string + required: + - label + - list_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateIpAccessListResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: IpAccessListsAPI + get: + operationId: ip_access_lists_list + summary: Gets all IP access lists for the specified workspace. + tags: + - settings + - ip_access_lists + description: |- + Gets all IP access lists for the specified workspace. + + + :returns: Iterator over :class:`IpAccessListInfo` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetIpAccessListsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: IpAccessListsAPI + /api/2.0/ip-access-lists/{ip_access_list_id}: + delete: + operationId: ip_access_lists_delete + summary: Deletes an IP access list, specified by its list ID. + tags: + - settings + - ip_access_lists + description: |- + Deletes an IP access list, specified by its list ID. + + :param ip_access_list_id: str + The ID for the corresponding IP access list + parameters: + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: IpAccessListsAPI + get: + operationId: ip_access_lists_get + summary: Gets an IP access list, specified by its list ID. + tags: + - settings + - ip_access_lists + description: |- + Gets an IP access list, specified by its list ID. + + :param ip_access_list_id: str + The ID for the corresponding IP access list + + :returns: :class:`FetchIpAccessListResponse` + parameters: + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FetchIpAccessListResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: IpAccessListsAPI + put: + operationId: ip_access_lists_replace + summary: Replaces an IP access list, specified by its ID. + tags: + - settings + - ip_access_lists + description: >- + Replaces an IP access list, specified by its ID. + + + A list can include allow lists and block lists. See the top of this file + for a description of how the + + server treats allow lists and block lists at run time. When replacing an + IP access list: * For all + + allow lists and block lists combined, the API supports a maximum of 1000 + IP/CIDR values, where one + + CIDR counts as a single value. Attempts to exceed that number return + error 400 with `error_code` value + + `QUOTA_EXCEEDED`. * If the resulting list would block the calling user's + current IP, error 400 is + + returned with `error_code` value `INVALID_STATE`. It can take a few + minutes for the changes to take + + effect. Note that your resulting IP access list has no effect until you + enable the feature. See + + :method:workspaceconf/setStatus. + + + :param ip_access_list_id: str + The ID for the corresponding IP access list + :param label: str + Label for the IP access list. This **cannot** be empty. + :param list_type: :class:`ListType` + + :param enabled: bool + Specifies whether this IP access list is enabled. + :param ip_addresses: List[str] (optional) + parameters: + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + requestBody: + content: + application/json: + schema: + type: object + properties: + label: + type: string + description: Label for the IP access list. This **cannot** be empty. + list_type: + type: string + description: >- + :param enabled: bool Specifies whether this IP access list + is enabled. + enabled: + type: boolean + ip_addresses: + type: string + description: List[str] (optional) + required: + - label + - list_type + - enabled + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: IpAccessListsAPI + patch: + operationId: ip_access_lists_update + summary: Updates an existing IP access list, specified by its ID. + tags: + - settings + - ip_access_lists + description: >- + Updates an existing IP access list, specified by its ID. + + + A list can include allow lists and block lists. See the top of this file + for a description of how the + + server treats allow lists and block lists at run time. + + + When updating an IP access list: + + + * For all allow lists and block lists combined, the API supports a + maximum of 1000 IP/CIDR values, + + where one CIDR counts as a single value. Attempts to exceed that number + return error 400 with + + `error_code` value `QUOTA_EXCEEDED`. * If the updated list would block + the calling user's current IP, + + error 400 is returned with `error_code` value `INVALID_STATE`. + + + It can take a few minutes for the changes to take effect. Note that your + resulting IP access list has + + no effect until you enable the feature. See + :method:workspaceconf/setStatus. + + + :param ip_access_list_id: str + The ID for the corresponding IP access list + :param enabled: bool (optional) + Specifies whether this IP access list is enabled. + :param ip_addresses: List[str] (optional) + + :param label: str (optional) + Label for the IP access list. This **cannot** be empty. + :param list_type: :class:`ListType` (optional) + parameters: + - name: ip_access_list_id + in: path + required: true + schema: + type: string + description: The ID for the corresponding IP access list + requestBody: + content: + application/json: + schema: + type: object + properties: + enabled: + type: string + description: Specifies whether this IP access list is enabled. + ip_addresses: + type: string + description: >- + :param label: str (optional) Label for the IP access list. + This **cannot** be empty. + label: + type: string + list_type: + type: string + description: ':class:`ListType` (optional)' + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: IpAccessListsAPI + /api/2.0/accounts/{account_id}/settings/types/llm_proxy_partner_powered/names/default: + get: + operationId: llm_proxy_partner_powered_account_get + summary: Gets the enable partner powered AI features account setting. + tags: + - settings + - llm_proxy_partner_powered_account + description: |- + Gets the enable partner powered AI features account setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`LlmProxyPartnerPoweredAccount` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LlmProxyPartnerPoweredAccount' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LlmProxyPartnerPoweredAccountAPI + patch: + operationId: llm_proxy_partner_powered_account_update + summary: Updates the enable partner powered AI features account setting. + tags: + - settings + - llm_proxy_partner_powered_account + description: |- + Updates the enable partner powered AI features account setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`LlmProxyPartnerPoweredAccount` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`LlmProxyPartnerPoweredAccount` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LlmProxyPartnerPoweredAccount' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LlmProxyPartnerPoweredAccountAPI + /api/2.0/accounts/{account_id}/settings/types/llm_proxy_partner_powered_enforce/names/default: + get: + operationId: llm_proxy_partner_powered_enforce_get + summary: >- + Gets the enforcement status of partner powered AI features account + setting. + tags: + - settings + - llm_proxy_partner_powered_enforce + description: >- + Gets the enforcement status of partner powered AI features account + setting. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`LlmProxyPartnerPoweredEnforce` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LlmProxyPartnerPoweredEnforce' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LlmProxyPartnerPoweredEnforceAPI + patch: + operationId: llm_proxy_partner_powered_enforce_update + summary: >- + Updates the enable enforcement status of partner powered AI features + account setting. + tags: + - settings + - llm_proxy_partner_powered_enforce + description: >- + Updates the enable enforcement status of partner powered AI features + account setting. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`LlmProxyPartnerPoweredEnforce` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`LlmProxyPartnerPoweredEnforce` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LlmProxyPartnerPoweredEnforce' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LlmProxyPartnerPoweredEnforceAPI + /api/2.0/settings/types/llm_proxy_partner_powered/names/default: + delete: + operationId: llm_proxy_partner_powered_workspace_delete + summary: >- + Reverts the enable partner powered AI features workspace setting to its + default value. + tags: + - settings + - llm_proxy_partner_powered_workspace + description: >- + Reverts the enable partner powered AI features workspace setting to its + default value. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteLlmProxyPartnerPoweredWorkspaceResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/DeleteLlmProxyPartnerPoweredWorkspaceResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LlmProxyPartnerPoweredWorkspaceAPI + get: + operationId: llm_proxy_partner_powered_workspace_get + summary: Gets the enable partner powered AI features workspace setting. + tags: + - settings + - llm_proxy_partner_powered_workspace + description: |- + Gets the enable partner powered AI features workspace setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`LlmProxyPartnerPoweredWorkspace` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LlmProxyPartnerPoweredWorkspace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LlmProxyPartnerPoweredWorkspaceAPI + patch: + operationId: llm_proxy_partner_powered_workspace_update + summary: Updates the enable partner powered AI features workspace setting. + tags: + - settings + - llm_proxy_partner_powered_workspace + description: |- + Updates the enable partner powered AI features workspace setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`LlmProxyPartnerPoweredWorkspace` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`LlmProxyPartnerPoweredWorkspace` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LlmProxyPartnerPoweredWorkspace' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: LlmProxyPartnerPoweredWorkspaceAPI + /api/2.0/notification-destinations: + post: + operationId: notification_destinations_create + summary: >- + Creates a notification destination. Requires workspace admin + permissions. + tags: + - settings + - notification_destinations + description: >- + Creates a notification destination. Requires workspace admin + permissions. + + + :param config: :class:`Config` (optional) + The configuration for the notification destination. Must wrap EXACTLY one of the nested configs. + :param display_name: str (optional) + The display name for the notification destination. + + :returns: :class:`NotificationDestination` + requestBody: + content: + application/json: + schema: + type: object + properties: + config: + type: string + description: >- + The configuration for the notification destination. Must + wrap EXACTLY one of the nested configs. + display_name: + type: string + description: The display name for the notification destination. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NotificationDestination' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NotificationDestinationsAPI + get: + operationId: notification_destinations_list + summary: Lists notification destinations. + tags: + - settings + - notification_destinations + description: |- + Lists notification destinations. + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`ListNotificationDestinationsResult` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListNotificationDestinationsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NotificationDestinationsAPI + /api/2.0/notification-destinations/{id}: + delete: + operationId: notification_destinations_delete + summary: >- + Deletes a notification destination. Requires workspace admin + permissions. + tags: + - settings + - notification_destinations + description: >- + Deletes a notification destination. Requires workspace admin + permissions. + + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NotificationDestinationsAPI + get: + operationId: notification_destinations_get + summary: Gets a notification destination. + tags: + - settings + - notification_destinations + description: |- + Gets a notification destination. + + :param id: str + + :returns: :class:`NotificationDestination` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`NotificationDestination`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NotificationDestination' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NotificationDestinationsAPI + patch: + operationId: notification_destinations_update + summary: >- + Updates a notification destination. Requires workspace admin + permissions. At least one field is + tags: + - settings + - notification_destinations + description: >- + Updates a notification destination. Requires workspace admin + permissions. At least one field is + + required in the request body. + + + :param id: str + UUID identifying notification destination. + :param config: :class:`Config` (optional) + The configuration for the notification destination. Must wrap EXACTLY one of the nested configs. + :param display_name: str (optional) + The display name for the notification destination. + + :returns: :class:`NotificationDestination` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: UUID identifying notification destination. + requestBody: + content: + application/json: + schema: + type: object + properties: + config: + type: string + description: >- + The configuration for the notification destination. Must + wrap EXACTLY one of the nested configs. + display_name: + type: string + description: The display name for the notification destination. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/NotificationDestination' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: NotificationDestinationsAPI + /api/2.0/accounts/{account_id}/settings/types/dcp_acct_enable/names/default: + delete: + operationId: personal_compute_delete + summary: Reverts back the Personal Compute setting value to default (ON) + tags: + - settings + - personal_compute + description: |- + Reverts back the Personal Compute setting value to default (ON) + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeletePersonalComputeSettingResponse` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeletePersonalComputeSettingResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PersonalComputeAPI + get: + operationId: personal_compute_get + summary: Gets the value of the Personal Compute setting. + tags: + - settings + - personal_compute + description: |- + Gets the value of the Personal Compute setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`PersonalComputeSetting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PersonalComputeSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PersonalComputeAPI + patch: + operationId: personal_compute_update + summary: Updates the value of the Personal Compute setting. + tags: + - settings + - personal_compute + description: |- + Updates the value of the Personal Compute setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`PersonalComputeSetting` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`PersonalComputeSetting` + parameters: + - name: account_id + in: path + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PersonalComputeSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: PersonalComputeAPI + /api/2.0/settings/types/restrict_workspace_admins/names/default: + delete: + operationId: restrict_workspace_admins_delete + summary: >- + Reverts the restrict workspace admins setting status for the workspace. + A fresh etag needs to be + tags: + - settings + - restrict_workspace_admins + description: >- + Reverts the restrict workspace admins setting status for the workspace. + A fresh etag needs to be + + provided in `DELETE` requests (as a query parameter). The etag can be + retrieved by making a `GET` + + request before the DELETE request. If the setting is updated/deleted + concurrently, `DELETE` fails with + + 409 and the request must be retried by using the fresh etag in the 409 + response. + + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteRestrictWorkspaceAdminsSettingResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/DeleteRestrictWorkspaceAdminsSettingResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RestrictWorkspaceAdminsAPI + get: + operationId: restrict_workspace_admins_get + summary: Gets the restrict workspace admins setting. + tags: + - settings + - restrict_workspace_admins + description: |- + Gets the restrict workspace admins setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`RestrictWorkspaceAdminsSetting` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RestrictWorkspaceAdminsSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RestrictWorkspaceAdminsAPI + patch: + operationId: restrict_workspace_admins_update + summary: >- + Updates the restrict workspace admins setting for the workspace. A fresh + etag needs to be provided in + tags: + - settings + - restrict_workspace_admins + description: >- + Updates the restrict workspace admins setting for the workspace. A fresh + etag needs to be provided in + + `PATCH` requests (as part of the setting field). The etag can be + retrieved by making a GET request + + before the `PATCH` request. If the setting is updated concurrently, + `PATCH` fails with 409 and the + + request must be retried by using the fresh etag in the 409 response. + + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`RestrictWorkspaceAdminsSetting` + + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`RestrictWorkspaceAdminsSetting` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RestrictWorkspaceAdminsSetting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RestrictWorkspaceAdminsAPI + /api/2.0/settings/types/sql_results_download/names/default: + delete: + operationId: sql_results_download_delete + summary: Reverts the SQL Results Download setting to its default value. + tags: + - settings + - sql_results_download + description: |- + Reverts the SQL Results Download setting to its default value. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`DeleteSqlResultsDownloadResponse` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteSqlResultsDownloadResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SqlResultsDownloadAPI + get: + operationId: sql_results_download_get + summary: Gets the SQL Results Download setting. + tags: + - settings + - sql_results_download + description: |- + Gets the SQL Results Download setting. + + :param etag: str (optional) + etag used for versioning. The response is at least as fresh as the eTag provided. This is used for + optimistic concurrency control as a way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the etag in the read -> delete pattern + to perform setting deletions in order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule set version you are deleting. + + :returns: :class:`SqlResultsDownload` + parameters: + - name: etag + in: query + required: false + schema: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SqlResultsDownload' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SqlResultsDownloadAPI + patch: + operationId: sql_results_download_update + summary: Updates the SQL Results Download setting. + tags: + - settings + - sql_results_download + description: |- + Updates the SQL Results Download setting. + + :param allow_missing: bool + This should always be set to true for Settings API. Added for AIP compliance. + :param setting: :class:`SqlResultsDownload` + :param field_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`SqlResultsDownload` + requestBody: + content: + application/json: + schema: + type: object + properties: + allow_missing: + type: boolean + description: >- + This should always be set to true for Settings API. Added + for AIP compliance. + setting: + type: string + description: >- + :param field_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + field_mask: + type: string + required: + - allow_missing + - setting + - field_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SqlResultsDownload' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SqlResultsDownloadAPI + /api/2.0/token-management/on-behalf-of/tokens: + post: + operationId: token_management_create_obo_token + summary: Creates a token on behalf of a service principal. + tags: + - settings + - token_management + description: |- + Creates a token on behalf of a service principal. + + :param application_id: str + Application ID of the service principal. + :param comment: str (optional) + Comment that describes the purpose of the token. + :param lifetime_seconds: int (optional) + The number of seconds before the token expires. + + :returns: :class:`CreateOboTokenResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + application_id: + type: string + description: Application ID of the service principal. + comment: + type: string + description: Comment that describes the purpose of the token. + lifetime_seconds: + type: string + description: The number of seconds before the token expires. + required: + - application_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateOboTokenResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + /api/2.0/token-management/tokens/{token_id}: + delete: + operationId: token_management_delete + summary: Deletes a token, specified by its ID. + tags: + - settings + - token_management + description: |- + Deletes a token, specified by its ID. + + :param token_id: str + The ID of the token to revoke. + parameters: + - name: token_id + in: path + required: true + schema: + type: string + description: The ID of the token to revoke. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + get: + operationId: token_management_get + summary: Gets information about a token, specified by its ID. + tags: + - settings + - token_management + description: |- + Gets information about a token, specified by its ID. + + :param token_id: str + The ID of the token to get. + + :returns: :class:`GetTokenResponse` + parameters: + - name: token_id + in: path + required: true + schema: + type: string + description: The ID of the token to get. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetTokenResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + /api/2.0/permissions/authorization/tokens/permissionLevels: + get: + operationId: token_management_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - settings + - token_management + description: |- + Gets the permission levels that a user can have on an object. + + + :returns: :class:`GetTokenPermissionLevelsResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetTokenPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + /api/2.0/permissions/authorization/tokens: + get: + operationId: token_management_get_permissions + summary: >- + Gets the permissions of all tokens. Tokens can inherit permissions from + their root object. + tags: + - settings + - token_management + description: >- + Gets the permissions of all tokens. Tokens can inherit permissions from + their root object. + + + + :returns: :class:`TokenPermissions` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TokenPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + put: + operationId: token_management_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - settings + - token_management + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param access_control_list: List[:class:`TokenAccessControlRequest`] + (optional) + + + :returns: :class:`TokenPermissions` + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`TokenPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TokenPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + patch: + operationId: token_management_update_permissions + summary: >- + Updates the permissions on all tokens. Tokens can inherit permissions + from their root object. + tags: + - settings + - token_management + description: >- + Updates the permissions on all tokens. Tokens can inherit permissions + from their root object. + + + :param access_control_list: List[:class:`TokenAccessControlRequest`] + (optional) + + + :returns: :class:`TokenPermissions` + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`TokenPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TokenPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + /api/2.0/token-management/tokens: + get: + operationId: token_management_list + summary: Lists all tokens associated with the specified workspace or user. + tags: + - settings + - token_management + description: |- + Lists all tokens associated with the specified workspace or user. + + :param created_by_id: int (optional) + User ID of the user that created the token. + :param created_by_username: str (optional) + Username of the user that created the token. + + :returns: Iterator over :class:`TokenInfo` + parameters: + - name: created_by_id + in: query + required: false + schema: + type: string + description: User ID of the user that created the token. + - name: created_by_username + in: query + required: false + schema: + type: string + description: Username of the user that created the token. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListTokensResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokenManagementAPI + /api/2.0/token/create: + post: + operationId: tokens_create + summary: >- + Creates and returns a token for a user. If this call is made through + token authentication, it creates + tags: + - settings + - tokens + description: >- + Creates and returns a token for a user. If this call is made through + token authentication, it creates + + a token with the same client ID as the authenticated token. If the + user's token quota is exceeded, + + this call returns an error **QUOTA_EXCEEDED**. + + + :param comment: str (optional) + Optional description to attach to the token. + :param lifetime_seconds: int (optional) + The lifetime of the token, in seconds. + + If the lifetime is not specified, this token remains valid for 2 years. + + :returns: :class:`CreateTokenResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: Optional description to attach to the token. + lifetime_seconds: + type: string + description: >- + The lifetime of the token, in seconds. If the lifetime is + not specified, this token remains valid for 2 years. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateTokenResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokensAPI + /api/2.0/token/delete: + post: + operationId: tokens_delete + summary: Revokes an access token. + tags: + - settings + - tokens + description: >- + Revokes an access token. + + + If a token with the specified ID is not valid, this call returns an + error **RESOURCE_DOES_NOT_EXIST**. + + + :param token_id: str + The ID of the token to be revoked. + requestBody: + content: + application/json: + schema: + type: object + properties: + token_id: + type: string + description: The ID of the token to be revoked. + required: + - token_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokensAPI + /api/2.0/token/list: + get: + operationId: tokens_list + summary: Lists all the valid tokens for a user-workspace pair. + tags: + - settings + - tokens + description: |- + Lists all the valid tokens for a user-workspace pair. + + + :returns: Iterator over :class:`PublicTokenInfo` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListPublicTokensResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TokensAPI + /api/2.0/workspace-conf: + get: + operationId: workspace_conf_get_status + summary: Gets the configuration status for a workspace. + tags: + - settings + - workspace_conf + description: |- + Gets the configuration status for a workspace. + + :param keys: str + + :returns: Dict[str,str] + parameters: + - name: keys + in: query + required: true + schema: + type: string + description: ':returns: Dict[str,str]' + responses: + '200': + description: Success + content: + application/json: + schema: + type: object + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceConfAPI + patch: + operationId: workspace_conf_set_status + summary: >- + Sets the configuration status for a workspace, including enabling or + disabling it. + tags: + - settings + - workspace_conf + description: >- + Sets the configuration status for a workspace, including enabling or + disabling it. + requestBody: + content: + application/json: + schema: + type: object + properties: + contents: + type: string + required: + - contents + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceConfAPI +components: + schemas: + AccountIpAccessEnable: + type: object + properties: + acct_ip_acl_enable: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - acct_ip_acl_enable + AccountNetworkPolicy: + type: object + properties: + account_id: + type: string + egress: + $ref: '#/components/schemas/NetworkPolicyEgress' + description: The network policies applying for egress traffic. + network_policy_id: + type: string + description: The unique identifier for the network policy. + AibiDashboardEmbeddingAccessPolicy: + type: object + properties: + access_policy_type: + $ref: >- + #/components/schemas/AibiDashboardEmbeddingAccessPolicyAccessPolicyType + required: + - access_policy_type + AibiDashboardEmbeddingAccessPolicySetting: + type: object + properties: + aibi_dashboard_embedding_access_policy: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicy' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - aibi_dashboard_embedding_access_policy + AibiDashboardEmbeddingApprovedDomains: + type: object + properties: + approved_domains: + type: array + items: + type: string + AibiDashboardEmbeddingApprovedDomainsSetting: + type: object + properties: + aibi_dashboard_embedding_approved_domains: + $ref: '#/components/schemas/AibiDashboardEmbeddingApprovedDomains' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - aibi_dashboard_embedding_approved_domains + AutomaticClusterUpdateSetting: + type: object + properties: + automatic_cluster_update_workspace: + $ref: '#/components/schemas/ClusterAutoRestartMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - automatic_cluster_update_workspace + BooleanMessage: + type: object + properties: + value: + type: boolean + ClusterAutoRestartMessage: + type: object + properties: + can_toggle: + type: boolean + enabled: + type: boolean + enablement_details: + $ref: '#/components/schemas/ClusterAutoRestartMessageEnablementDetails' + maintenance_window: + $ref: '#/components/schemas/ClusterAutoRestartMessageMaintenanceWindow' + restart_even_if_no_updates_available: + type: boolean + ClusterAutoRestartMessageEnablementDetails: + type: object + properties: + forced_for_compliance_mode: + type: boolean + description: The feature is force enabled if compliance mode is active + unavailable_for_disabled_entitlement: + type: boolean + description: >- + The feature is unavailable if the corresponding entitlement disabled + (see getShieldEntitlementEnable) + unavailable_for_non_enterprise_tier: + type: boolean + description: >- + The feature is unavailable if the customer doesn't have enterprise + tier + description: >- + Contains an information about the enablement status judging (e.g. + whether the enterprise tier is + enabled) This is only additional information that MUST NOT be used to decide whether the setting + is enabled or not. This is intended to use only for purposes like showing an error message to + the customer with the additional details. For example, using these details we can check why + exactly the feature is disabled for this customer. + ClusterAutoRestartMessageMaintenanceWindow: + type: object + properties: + week_day_based_schedule: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule + ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule: + type: object + properties: + day_of_week: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowDayOfWeek + frequency: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency + window_start_time: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWindowStartTime + ClusterAutoRestartMessageMaintenanceWindowWindowStartTime: + type: object + properties: + hours: + type: integer + minutes: + type: integer + ComplianceSecurityProfile: + type: object + properties: + compliance_standards: + type: array + items: + $ref: '#/components/schemas/ComplianceStandard' + description: Set by customers when they request Compliance Security Profile (CSP) + is_enabled: + type: boolean + description: 'SHIELD feature: CSP' + ComplianceSecurityProfileSetting: + type: object + properties: + compliance_security_profile_workspace: + $ref: '#/components/schemas/ComplianceSecurityProfile' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - compliance_security_profile_workspace + Config: + type: object + properties: + email: + $ref: '#/components/schemas/EmailConfig' + generic_webhook: + $ref: '#/components/schemas/GenericWebhookConfig' + microsoft_teams: + $ref: '#/components/schemas/MicrosoftTeamsConfig' + pagerduty: + $ref: '#/components/schemas/PagerdutyConfig' + slack: + $ref: '#/components/schemas/SlackConfig' + CreateIpAccessListResponse: + type: object + properties: + ip_access_list: + $ref: '#/components/schemas/IpAccessListInfo' + description: An IP access list was successfully created. + CreateNetworkConnectivityConfiguration: + type: object + properties: + name: + type: string + description: >- + The name of the network connectivity configuration. The name can + contain alphanumeric characters, hyphens, and underscores. The + length must be between 3 and 30 characters. The name must match the + regular expression ^[0-9a-zA-Z-_]{3,30}$ + region: + type: string + description: >- + The region for the network connectivity configuration. Only + workspaces in the same region can be attached to the network + connectivity configuration. + required: + - name + - region + description: Properties of the new network connectivity configuration. + CreateOboTokenResponse: + type: object + properties: + token_info: + $ref: '#/components/schemas/TokenInfo' + token_value: + type: string + description: Value of the token. + description: An on-behalf token was successfully created for the service principal. + CreatePrivateEndpointRule: + type: object + properties: + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints to customer-managed private endpoint + services. Domain names of target private link service. When updating + this field, the full list of target domain_names must be specified. + endpoint_service: + type: string + description: >- + The full target AWS endpoint service name that connects to the + destination resources of the private endpoint. + error_message: + type: string + group_id: + type: string + description: >- + Not used by customer-managed private endpoint services. The + sub-resource type (group ID) of the target resource. Note that to + connect to workspace root storage (root DBFS), you need two + endpoints, one for blob and one for dfs. + resource_id: + type: string + description: The Azure resource ID of the target resource. + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + CreateTokenResponse: + type: object + properties: + token_info: + $ref: '#/components/schemas/PublicTokenInfo' + token_value: + type: string + description: The value of the new token. + CspEnablementAccount: + type: object + properties: + compliance_standards: + type: array + items: + $ref: '#/components/schemas/ComplianceStandard' + description: >- + Set by customers when they request Compliance Security Profile (CSP) + Invariants are enforced in Settings policy. + is_enforced: + type: boolean + description: Enforced = it cannot be overriden at workspace level. + description: Account level policy for CSP + CspEnablementAccountSetting: + type: object + properties: + csp_enablement_account: + $ref: '#/components/schemas/CspEnablementAccount' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - csp_enablement_account + CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRule: + type: object + properties: + account_id: + type: string + description: >- + Databricks account ID. You can find your account ID from the + Accounts Console. + connection_state: + $ref: >- + #/components/schemas/CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRulePrivateLinkConnectionState + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + deactivated: + type: boolean + description: Whether this private endpoint is deactivated. + deactivated_at: + type: integer + description: Time in epoch milliseconds when this object was deactivated. + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards a VPC endpoint service for + customer-managed VPC endpoint service. The target AWS resource FQDNs + accessible via the VPC endpoint service. When updating this field, + we perform full update on this field. Please ensure a full list of + desired domain_names is provided. + enabled: + type: boolean + description: >- + Only used by private endpoints towards an AWS S3 service. Update + this field to activate/deactivate this private endpoint to allow + egress access from serverless compute resources. + endpoint_service: + type: string + description: >- + The full target AWS endpoint service name that connects to the + destination resources of the private endpoint. + error_message: + type: string + network_connectivity_config_id: + type: string + description: >- + The ID of a network connectivity configuration, which is the parent + resource of this private endpoint rule object. + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + rule_id: + type: string + description: The ID of a private endpoint rule. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + vpc_endpoint_id: + type: string + description: >- + The AWS VPC endpoint ID. You can use this ID to identify VPC + endpoint created by Databricks. + description: >- + Properties of the new private endpoint rule. Note that for private + endpoints towards a VPC + endpoint service behind a customer-managed NLB, you must approve the endpoint in AWS console + after initialization. + DashboardEmailSubscriptions: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + DefaultNamespaceSetting: + type: object + properties: + namespace: + $ref: '#/components/schemas/StringMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - namespace + description: >- + This represents the setting configuration for the default namespace in + the Databricks workspace. + Setting the default catalog for the workspace determines the catalog that is used when queries + do not reference a fully qualified 3 level name. For example, if the default catalog is set to + 'retail_prod' then a query 'SELECT * FROM myTable' would reference the object + 'retail_prod.default.myTable' (the schema 'default' is always assumed). This setting requires a + restart of clusters and SQL warehouses to take effect. Additionally, the default namespace only + applies when using Unity Catalog-enabled compute. + DefaultWarehouseId: + type: object + properties: + string_val: + $ref: '#/components/schemas/StringMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - string_val + DeleteAccountIpAccessEnableResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteAibiDashboardEmbeddingAccessPolicySettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteAibiDashboardEmbeddingApprovedDomainsSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDashboardEmailSubscriptionsResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDefaultNamespaceSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDefaultWarehouseIdResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDisableLegacyAccessResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDisableLegacyDbfsResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteDisableLegacyFeaturesResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteLlmProxyPartnerPoweredWorkspaceResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeletePersonalComputeSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteRestrictWorkspaceAdminsSettingResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DeleteSqlResultsDownloadResponse: + type: object + properties: + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> delete pattern to perform setting deletions in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the DELETE request to identify the rule + set version you are deleting. + required: + - etag + description: The etag is returned. + DisableLegacyAccess: + type: object + properties: + disable_legacy_access: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - disable_legacy_access + DisableLegacyDbfs: + type: object + properties: + disable_legacy_dbfs: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - disable_legacy_dbfs + DisableLegacyFeatures: + type: object + properties: + disable_legacy_features: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - disable_legacy_features + EgressNetworkPolicy: + type: object + properties: + internet_access: + $ref: '#/components/schemas/EgressNetworkPolicyInternetAccessPolicy' + description: The access policy enforced for egress traffic to the internet. + description: >- + The network policies applying for egress traffic. This message is used + by the UI/REST API. We + translate this message to the format expected by the dataplane in Lakehouse Network Manager (for + the format expected by the dataplane, see networkconfig.textproto). + EgressNetworkPolicyInternetAccessPolicy: + type: object + properties: + allowed_internet_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyInternetDestination + allowed_storage_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyStorageDestination + log_only_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyLogOnlyMode + description: >- + Optional. If not specified, assume the policy is enforced for all + workloads. + restriction_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyRestrictionMode + EgressNetworkPolicyInternetAccessPolicyInternetDestination: + type: object + properties: + destination: + type: string + protocol: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationFilteringProtocol + type: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationType + description: >- + Users can specify accessible internet destinations when outbound access + is restricted. We only + support domain name (FQDN) destinations for the time being, though going forwards we want to + support host names and IP addresses. + EgressNetworkPolicyInternetAccessPolicyLogOnlyMode: + type: object + properties: + log_only_mode_type: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyLogOnlyModeLogOnlyModeType + workloads: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyLogOnlyModeWorkloadType + EgressNetworkPolicyInternetAccessPolicyStorageDestination: + type: object + properties: + allowed_paths: + type: array + items: + type: string + azure_container: + type: string + azure_dns_zone: + type: string + azure_storage_account: + type: string + azure_storage_service: + type: string + bucket_name: + type: string + region: + type: string + type: + $ref: >- + #/components/schemas/EgressNetworkPolicyInternetAccessPolicyStorageDestinationStorageDestinationType + description: Users can specify accessible storage destinations. + EgressNetworkPolicyNetworkAccessPolicy: + type: object + properties: + restriction_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyRestrictionMode + allowed_internet_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyInternetDestination + description: >- + List of internet destinations that serverless workloads are allowed + to access when in RESTRICTED_ACCESS mode. + allowed_storage_destinations: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyStorageDestination + description: >- + List of storage destinations that serverless workloads are allowed + to access when in RESTRICTED_ACCESS mode. + policy_enforcement: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcement + description: >- + Optional. When policy_enforcement is not provided, we default to + ENFORCE_MODE_ALL_SERVICES + required: + - restriction_mode + EgressNetworkPolicyNetworkAccessPolicyInternetDestination: + type: object + properties: + destination: + type: string + description: >- + The internet destination to which access will be allowed. Format + dependent on the destination type. + internet_destination_type: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyInternetDestinationInternetDestinationType + description: >- + Users can specify accessible internet destinations when outbound access + is restricted. We only + support DNS_NAME (FQDN format) destinations for the time being. Going forward we may extend + support to host names and IP addresses. + EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcement: + type: object + properties: + dry_run_mode_product_filter: + type: array + items: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementDryRunModeProductFilter + enforcement_mode: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementEnforcementMode + description: >- + The mode of policy enforcement. ENFORCED blocks traffic that + violates policy, while DRY_RUN only logs violations without + blocking. When not specified, defaults to ENFORCED. + EgressNetworkPolicyNetworkAccessPolicyStorageDestination: + type: object + properties: + azure_storage_account: + type: string + description: The Azure storage account name. + azure_storage_service: + type: string + description: The Azure storage service type (blob, dfs, etc.). + bucket_name: + type: string + region: + type: string + storage_destination_type: + $ref: >- + #/components/schemas/EgressNetworkPolicyNetworkAccessPolicyStorageDestinationStorageDestinationType + description: Users can specify accessible storage destinations. + EmailConfig: + type: object + properties: + addresses: + type: array + items: + type: string + Empty: + type: object + properties: {} + EnableExportNotebook: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + EnableNotebookTableClipboard: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + EnableResultsDownloading: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + EnhancedSecurityMonitoring: + type: object + properties: + is_enabled: + type: boolean + description: 'SHIELD feature: ESM' + EnhancedSecurityMonitoringSetting: + type: object + properties: + enhanced_security_monitoring_workspace: + $ref: '#/components/schemas/EnhancedSecurityMonitoring' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - enhanced_security_monitoring_workspace + EsmEnablementAccount: + type: object + properties: + is_enforced: + type: boolean + description: Account level policy for ESM + EsmEnablementAccountSetting: + type: object + properties: + esm_enablement_account: + $ref: '#/components/schemas/EsmEnablementAccount' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - esm_enablement_account + ExchangeToken: + type: object + properties: + credential: + type: string + description: The requested token. + credentialEolTime: + type: integer + description: >- + The end-of-life timestamp of the token. The value is in milliseconds + since the Unix epoch. + ownerId: + type: integer + description: User ID of the user that owns this token. + scopes: + type: array + items: + type: string + description: The scopes of access granted in the token. + tokenType: + $ref: '#/components/schemas/TokenType' + description: The type of this exchange token + description: The exchange token is the result of the token exchange with the IdP + ExchangeTokenResponse: + type: object + properties: + values: + type: array + items: + $ref: '#/components/schemas/ExchangeToken' + description: Exhanged tokens were successfully returned. + FetchIpAccessListResponse: + type: object + properties: + ip_access_list: + $ref: '#/components/schemas/IpAccessListInfo' + description: An IP access list was successfully returned. + GenericWebhookConfig: + type: object + properties: + password: + type: string + password_set: + type: boolean + description: '[Output-Only] Whether password is set.' + url: + type: string + description: '[Input-Only] URL for webhook.' + url_set: + type: boolean + description: '[Output-Only] Whether URL is set.' + username: + type: string + description: '[Input-Only][Optional] Username for webhook.' + username_set: + type: boolean + description: '[Output-Only] Whether username is set.' + GetIpAccessListResponse: + type: object + properties: + ip_access_list: + $ref: '#/components/schemas/IpAccessListInfo' + GetIpAccessListsResponse: + type: object + properties: + ip_access_lists: + type: array + items: + $ref: '#/components/schemas/IpAccessListInfo' + description: IP access lists were successfully returned. + GetTokenPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/TokenPermissionsDescription' + GetTokenResponse: + type: object + properties: + token_info: + $ref: '#/components/schemas/TokenInfo' + description: Token with specified Token ID was successfully returned. + IpAccessListInfo: + type: object + properties: + address_count: + type: integer + description: Total number of IP or CIDR values. + created_at: + type: integer + description: Creation timestamp in milliseconds. + created_by: + type: integer + description: User ID of the user who created this list. + enabled: + type: boolean + description: Specifies whether this IP access list is enabled. + ip_addresses: + type: array + items: + type: string + label: + type: string + description: Label for the IP access list. This **cannot** be empty. + list_id: + type: string + description: Universally unique identifier (UUID) of the IP access list. + list_type: + $ref: '#/components/schemas/ListType' + updated_at: + type: integer + description: Update timestamp in milliseconds. + updated_by: + type: integer + description: User ID of the user who updated this list. + description: Definition of an IP Access list + ListIpAccessListResponse: + type: object + properties: + ip_access_lists: + type: array + items: + $ref: '#/components/schemas/IpAccessListInfo' + description: IP access lists were successfully returned. + ListNetworkConnectivityConfigurationsResponse: + type: object + properties: + items: + type: array + items: + $ref: '#/components/schemas/NetworkConnectivityConfiguration' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If null, + there are no more results to show. + description: The network connectivity configuration list was successfully retrieved. + ListNetworkPoliciesResponse: + type: object + properties: + items: + type: array + items: + $ref: '#/components/schemas/AccountNetworkPolicy' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If null, + there are no more results to show. + ListNotificationDestinationsResponse: + type: object + properties: + next_page_token: + type: string + results: + type: array + items: + $ref: '#/components/schemas/ListNotificationDestinationsResult' + ListNotificationDestinationsResult: + type: object + properties: + destination_type: + $ref: '#/components/schemas/DestinationType' + display_name: + type: string + description: The display name for the notification destination. + id: + type: string + description: UUID identifying notification destination. + ListPrivateEndpointRulesResponse: + type: object + properties: + items: + type: array + items: + $ref: '#/components/schemas/NccPrivateEndpointRule' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If null, + there are no more results to show. + description: The private endpoint rule list was successfully retrieved. + ListPublicTokensResponse: + type: object + properties: + token_infos: + type: array + items: + $ref: '#/components/schemas/PublicTokenInfo' + ListTokensResponse: + type: object + properties: + token_infos: + type: array + items: + $ref: '#/components/schemas/TokenInfo' + description: Token metadata of each user-created token in the workspace + description: Tokens were successfully returned. + LlmProxyPartnerPoweredAccount: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + LlmProxyPartnerPoweredEnforce: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + LlmProxyPartnerPoweredWorkspace: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + MicrosoftTeamsConfig: + type: object + properties: + app_id: + type: string + app_id_set: + type: boolean + description: '[Output-Only] Whether App ID is set.' + auth_secret: + type: string + description: '[Input-Only] Secret for Microsoft Teams App authentication.' + auth_secret_set: + type: boolean + description: '[Output-Only] Whether secret is set.' + channel_url: + type: string + description: '[Input-Only] Channel URL for Microsoft Teams App.' + channel_url_set: + type: boolean + description: '[Output-Only] Whether Channel URL is set.' + tenant_id: + type: string + description: '[Input-Only] Tenant ID for Microsoft Teams App.' + tenant_id_set: + type: boolean + description: '[Output-Only] Whether Tenant ID is set.' + url: + type: string + description: '[Input-Only] URL for Microsoft Teams webhook.' + url_set: + type: boolean + description: '[Output-Only] Whether URL is set.' + NccAwsStableIpRule: + type: object + properties: + cidr_blocks: + type: array + items: + type: string + description: >- + The list of stable IP CIDR blocks from which Databricks network + traffic originates when accessing your resources. + description: >- + The stable AWS IP CIDR blocks. You can use these to configure the + firewall of your resources to + allow traffic from your Databricks workspace. + NccAzurePrivateEndpointRule: + type: object + properties: + connection_state: + $ref: '#/components/schemas/NccAzurePrivateEndpointRuleConnectionState' + description: >- + The current status of this private endpoint. The private endpoint + rules are effective only if the connection state is ESTABLISHED. + Remember that you must approve new endpoints on your resources in + the Azure portal before they take effect. The possible values are: - + INIT: (deprecated) The endpoint has been created and pending + approval. - PENDING: The endpoint has been created and pending + approval. - ESTABLISHED: The endpoint has been approved and is ready + to use in your serverless compute resources. - REJECTED: Connection + was rejected by the private link resource owner. - DISCONNECTED: + Connection was removed by the private link resource owner, the + private endpoint becomes informative and should be deleted for + clean-up. - EXPIRED: If the endpoint was created but not approved in + 14 days, it will be EXPIRED. + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + deactivated: + type: boolean + description: Whether this private endpoint is deactivated. + deactivated_at: + type: integer + description: Time in epoch milliseconds when this object was deactivated. + domain_names: + type: array + items: + type: string + description: >- + Not used by customer-managed private endpoint services. Domain names + of target private link service. When updating this field, the full + list of target domain_names must be specified. + endpoint_name: + type: string + description: The name of the Azure private endpoint resource. + error_message: + type: string + group_id: + type: string + description: >- + Only used by private endpoints to Azure first-party services. The + sub-resource type (group ID) of the target resource. Note that to + connect to workspace root storage (root DBFS), you need two + endpoints, one for blob and one for dfs. + network_connectivity_config_id: + type: string + description: >- + The ID of a network connectivity configuration, which is the parent + resource of this private endpoint rule object. + resource_id: + type: string + description: The Azure resource ID of the target resource. + rule_id: + type: string + description: The ID of a private endpoint rule. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + NccAzureServiceEndpointRule: + type: object + properties: + subnets: + type: array + items: + type: string + description: >- + The list of subnets from which Databricks network traffic originates + when accessing your Azure resources. + target_region: + type: string + description: The Azure region in which this service endpoint rule applies.. + target_services: + type: array + items: + $ref: '#/components/schemas/EgressResourceType' + description: The Azure services to which this service endpoint rule applies to. + description: >- + The stable Azure service endpoints. You can configure the firewall of + your Azure resources to + allow traffic from your Databricks serverless compute resources. + NccEgressConfig: + type: object + properties: + default_rules: + $ref: '#/components/schemas/NccEgressDefaultRules' + target_rules: + $ref: '#/components/schemas/NccEgressTargetRules' + description: >- + The network connectivity rules that configured for each + destinations. These rules override default rules. + NccEgressDefaultRules: + type: object + properties: + aws_stable_ip_rule: + $ref: '#/components/schemas/NccAwsStableIpRule' + azure_service_endpoint_rule: + $ref: '#/components/schemas/NccAzureServiceEndpointRule' + description: Default rules don't have specific targets. + NccEgressTargetRules: + type: object + properties: + aws_private_endpoint_rules: + type: array + items: + $ref: >- + #/components/schemas/CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRule + description: >- + AWS private endpoint rule controls the AWS private endpoint based + egress rules. + azure_private_endpoint_rules: + type: array + items: + $ref: '#/components/schemas/NccAzurePrivateEndpointRule' + description: >- + Target rule controls the egress rules that are dedicated to specific + resources. + NccPrivateEndpointRule: + type: object + properties: + account_id: + type: string + description: >- + Databricks account ID. You can find your account ID from the + Accounts Console. + connection_state: + $ref: >- + #/components/schemas/NccPrivateEndpointRulePrivateLinkConnectionState + description: >- + The current status of this private endpoint. The private endpoint + rules are effective only if the connection state is ESTABLISHED. + Remember that you must approve new endpoints on your resources in + the Cloud console before they take effect. The possible values are: + - PENDING: The endpoint has been created and pending approval. - + ESTABLISHED: The endpoint has been approved and is ready to use in + your serverless compute resources. - REJECTED: Connection was + rejected by the private link resource owner. - DISCONNECTED: + Connection was removed by the private link resource owner, the + private endpoint becomes informative and should be deleted for + clean-up. - EXPIRED: If the endpoint was created but not approved in + 14 days, it will be EXPIRED. - CREATING: The endpoint creation is in + progress. Once successfully created, the state will transition to + PENDING. - CREATE_FAILED: The endpoint creation failed. You can + check the error_message field for more details. + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + deactivated: + type: boolean + description: Whether this private endpoint is deactivated. + deactivated_at: + type: integer + description: Time in epoch milliseconds when this object was deactivated. + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints to customer-managed private endpoint + services. Domain names of target private link service. When updating + this field, the full list of target domain_names must be specified. + enabled: + type: boolean + description: >- + Only used by private endpoints towards an AWS S3 service. Update + this field to activate/deactivate this private endpoint to allow + egress access from serverless compute resources. + endpoint_name: + type: string + description: The name of the Azure private endpoint resource. + endpoint_service: + type: string + description: >- + The full target AWS endpoint service name that connects to the + destination resources of the private endpoint. + error_message: + type: string + group_id: + type: string + description: >- + Not used by customer-managed private endpoint services. The + sub-resource type (group ID) of the target resource. Note that to + connect to workspace root storage (root DBFS), you need two + endpoints, one for blob and one for dfs. + network_connectivity_config_id: + type: string + description: >- + The ID of a network connectivity configuration, which is the parent + resource of this private endpoint rule object. + resource_id: + type: string + description: The Azure resource ID of the target resource. + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + rule_id: + type: string + description: The ID of a private endpoint rule. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + vpc_endpoint_id: + type: string + description: >- + The AWS VPC endpoint ID. You can use this ID to identify the VPC + endpoint created by Databricks. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + NetworkConnectivityConfiguration: + type: object + properties: + account_id: + type: string + description: >- + Your Databricks account ID. You can find your account ID in your + Databricks accounts console. + creation_time: + type: integer + description: Time in epoch milliseconds when this object was created. + egress_config: + $ref: '#/components/schemas/NccEgressConfig' + description: >- + The network connectivity rules that apply to network traffic from + your serverless compute resources. + name: + type: string + description: >- + The name of the network connectivity configuration. The name can + contain alphanumeric characters, hyphens, and underscores. The + length must be between 3 and 30 characters. The name must match the + regular expression ^[0-9a-zA-Z-_]{3,30}$ + network_connectivity_config_id: + type: string + description: Databricks network connectivity configuration ID. + region: + type: string + description: >- + The region for the network connectivity configuration. Only + workspaces in the same region can be attached to the network + connectivity configuration. + updated_time: + type: integer + description: Time in epoch milliseconds when this object was updated. + description: Properties of the new network connectivity configuration. + NetworkPolicyEgress: + type: object + properties: + network_access: + $ref: '#/components/schemas/EgressNetworkPolicyNetworkAccessPolicy' + description: The access policy enforced for egress traffic to the internet. + description: >- + The network policies applying for egress traffic. This message is used + by the UI/REST API. We + translate this message to the format expected by the dataplane in Lakehouse Network Manager (for + the format expected by the dataplane, see networkconfig.textproto). This policy should be + consistent with [[com.databricks.api.proto.settingspolicy.EgressNetworkPolicy]]. Details see + API-design: https://docs.google.com/document/d/1DKWO_FpZMCY4cF2O62LpwII1lx8gsnDGG-qgE3t3TOA/ + NotificationDestination: + type: object + properties: + config: + $ref: '#/components/schemas/Config' + destination_type: + $ref: '#/components/schemas/DestinationType' + description: >- + [Output-only] The type of the notification destination. The type can + not be changed once set. + display_name: + type: string + description: The display name for the notification destination. + id: + type: string + description: UUID identifying notification destination. + PagerdutyConfig: + type: object + properties: + integration_key: + type: string + integration_key_set: + type: boolean + description: '[Output-Only] Whether integration key is set.' + PartitionId: + type: object + properties: + workspaceId: + type: integer + description: The ID of the workspace. + description: Partition by workspace or account + PersonalComputeMessage: + type: object + properties: + value: + $ref: '#/components/schemas/PersonalComputeMessageEnum' + required: + - value + PersonalComputeSetting: + type: object + properties: + personal_compute: + $ref: '#/components/schemas/PersonalComputeMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - personal_compute + PublicTokenInfo: + type: object + properties: + comment: + type: string + creation_time: + type: integer + description: Server time (in epoch milliseconds) when the token was created. + expiry_time: + type: integer + description: >- + Server time (in epoch milliseconds) when the token will expire, or + -1 if not applicable. + token_id: + type: string + description: The ID of this token. + RestrictWorkspaceAdminsMessage: + type: object + properties: + status: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessageStatus' + required: + - status + RestrictWorkspaceAdminsSetting: + type: object + properties: + restrict_workspace_admins: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - restrict_workspace_admins + RevokeTokenResponse: + type: object + properties: {} + SlackConfig: + type: object + properties: + channel_id: + type: string + channel_id_set: + type: boolean + description: '[Output-Only] Whether channel ID is set.' + oauth_token: + type: string + description: '[Input-Only] OAuth token for Slack authentication.' + oauth_token_set: + type: boolean + description: '[Output-Only] Whether OAuth token is set.' + url: + type: string + description: '[Input-Only] URL for Slack destination.' + url_set: + type: boolean + description: '[Output-Only] Whether URL is set.' + SqlResultsDownload: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + etag: + type: string + description: >- + etag used for versioning. The response is at least as fresh as the + eTag provided. This is used for optimistic concurrency control as a + way to help prevent simultaneous writes of a setting overwriting + each other. It is strongly suggested that systems make use of the + etag in the read -> update pattern to perform setting updates in + order to avoid race conditions. That is, get an etag from a GET + request, and pass it with the PATCH request to identify the setting + version you are updating. + setting_name: + type: string + description: >- + Name of the corresponding setting. This field is populated in the + response, but it will not be respected even if it's set in the + request body. The setting name in the path parameter will be + respected instead. Setting name is required to be 'default' if the + setting only has one instance per workspace. + required: + - boolean_val + StringMessage: + type: object + properties: + value: + type: string + TokenAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/TokenPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + TokenAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/TokenPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + TokenInfo: + type: object + properties: + comment: + type: string + created_by_id: + type: integer + description: User ID of the user that created the token. + created_by_username: + type: string + description: Username of the user that created the token. + creation_time: + type: integer + description: Timestamp when the token was created. + expiry_time: + type: integer + description: Timestamp when the token expires. + last_used_day: + type: integer + description: >- + Approximate timestamp for the day the token was last used. Accurate + up to 1 day. + owner_id: + type: integer + description: User ID of the user that owns the token. + token_id: + type: string + description: ID of the token. + workspace_id: + type: integer + description: >- + If applicable, the ID of the workspace that the token was created + in. + TokenPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/TokenPermissionLevel' + TokenPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/TokenAccessControlResponse' + object_id: + type: string + object_type: + type: string + TokenPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/TokenPermissionLevel' + UpdatePrivateEndpointRule: + type: object + properties: + domain_names: + type: array + items: + type: string + description: >- + Only used by private endpoints to customer-managed private endpoint + services. Domain names of target private link service. When updating + this field, the full list of target domain_names must be specified. + enabled: + type: boolean + description: >- + Only used by private endpoints towards an AWS S3 service. Update + this field to activate/deactivate this private endpoint to allow + egress access from serverless compute resources. + error_message: + type: string + resource_names: + type: array + items: + type: string + description: >- + Only used by private endpoints towards AWS S3 service. The globally + unique S3 bucket names that will be accessed via the VPC endpoint. + The bucket names must be in the same region as the NCC/endpoint + service. When updating this field, we perform full update on this + field. Please ensure a full list of desired resource_names is + provided. + description: >- + Properties of the new private endpoint rule. Note that you must approve + the endpoint in Azure + portal after initialization. + WorkspaceNetworkOption: + type: object + properties: + network_policy_id: + type: string + workspace_id: + type: integer + description: The workspace ID. + AibiDashboardEmbeddingAccessPolicyAccessPolicyType: + type: string + enum: + - ALLOW_ALL_DOMAINS + - ALLOW_APPROVED_DOMAINS + - DENY_ALL_DOMAINS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowDayOfWeek: + type: string + enum: + - FRIDAY + - MONDAY + - SATURDAY + - SUNDAY + - THURSDAY + - TUESDAY + - WEDNESDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency: + type: string + enum: + - EVERY_WEEK + - FIRST_AND_THIRD_OF_MONTH + - FIRST_OF_MONTH + - FOURTH_OF_MONTH + - SECOND_AND_FOURTH_OF_MONTH + - SECOND_OF_MONTH + - THIRD_OF_MONTH + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ComplianceStandard: + type: string + enum: + - CANADA_PROTECTED_B + - CYBER_ESSENTIAL_PLUS + - FEDRAMP_HIGH + - FEDRAMP_IL5 + - FEDRAMP_MODERATE + - GERMANY_C5 + - GERMANY_TISAX + - HIPAA + - HITRUST + - IRAP_PROTECTED + - ISMAP + - ITAR_EAR + - K_FSI + - NONE + - PCI_DSS + description: >- + Compliance standard for SHIELD customers. See README.md for how + instructions of how to add new + + standards. + CustomerFacingNetworkConnectivityConfigAwsPrivateEndpointRulePrivateLinkConnectionState: + type: string + enum: + - CREATE_FAILED + - CREATING + - DISCONNECTED + - ESTABLISHED + - EXPIRED + - PENDING + - REJECTED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DestinationType: + type: string + enum: + - EMAIL + - MICROSOFT_TEAMS + - PAGERDUTY + - SLACK + - WEBHOOK + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationFilteringProtocol: + type: string + enum: + - TCP + description: >- + The filtering protocol used by the DP. For private and public preview, + SEG will only support TCP + + filtering (i.e. DNS based filtering, filtering by destination IP + address), so protocol will be + + set to TCP by default and hidden from the user. In the future, users may + be able to select HTTP + + filtering (i.e. SNI based filtering, filtering by FQDN). + EgressNetworkPolicyInternetAccessPolicyInternetDestinationInternetDestinationType: + type: string + enum: + - FQDN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyInternetAccessPolicyLogOnlyModeLogOnlyModeType: + type: string + enum: + - ALL_SERVICES + - SELECTED_SERVICES + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyInternetAccessPolicyLogOnlyModeWorkloadType: + type: string + enum: + - DBSQL + - ML_SERVING + description: >- + The values should match the list of workloads used in + networkconfig.proto + EgressNetworkPolicyInternetAccessPolicyRestrictionMode: + type: string + enum: + - FULL_ACCESS + - PRIVATE_ACCESS_ONLY + - RESTRICTED_ACCESS + description: >- + At which level can Databricks and Databricks managed compute access + Internet. FULL_ACCESS: + + Databricks can access Internet. No blocking rules will apply. + RESTRICTED_ACCESS: Databricks can + + only access explicitly allowed internet and storage destinations, as + well as UC connections and + + external locations. PRIVATE_ACCESS_ONLY (not used): Databricks can only + access destinations via + + private link. + EgressNetworkPolicyInternetAccessPolicyStorageDestinationStorageDestinationType: + type: string + enum: + - AWS_S3 + - AZURE_STORAGE + - CLOUDFLARE_R2 + - GOOGLE_CLOUD_STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyNetworkAccessPolicyInternetDestinationInternetDestinationType: + type: string + enum: + - DNS_NAME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementDryRunModeProductFilter: + type: string + enum: + - DBSQL + - ML_SERVING + description: >- + The values should match the list of workloads used in + networkconfig.proto + EgressNetworkPolicyNetworkAccessPolicyPolicyEnforcementEnforcementMode: + type: string + enum: + - DRY_RUN + - ENFORCED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressNetworkPolicyNetworkAccessPolicyRestrictionMode: + type: string + enum: + - FULL_ACCESS + - RESTRICTED_ACCESS + description: >- + At which level can Databricks and Databricks managed compute access + Internet. FULL_ACCESS: + + Databricks can access Internet. No blocking rules will apply. + RESTRICTED_ACCESS: Databricks can + + only access explicitly allowed internet and storage destinations, as + well as UC connections and + + external locations. + EgressNetworkPolicyNetworkAccessPolicyStorageDestinationStorageDestinationType: + type: string + enum: + - AWS_S3 + - AZURE_STORAGE + - GOOGLE_CLOUD_STORAGE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EgressResourceType: + type: string + enum: + - AZURE_BLOB_STORAGE + description: >- + The target resources that are supported by Network Connectivity Config. + Note: some egress types + + can support general types that are not defined in EgressResourceType. + E.g.: Azure private + + endpoint supports private link enabled Azure services. + ListType: + type: string + enum: + - ALLOW + - BLOCK + description: >- + Type of IP access list. Valid values are as follows and are + case-sensitive: + + + * `ALLOW`: An allow list. Include this IP or range. * `BLOCK`: A block + list. Exclude this IP or + + range. IP addresses in the block list are excluded even if they are + included in an allow list. + NccAzurePrivateEndpointRuleConnectionState: + type: string + enum: + - CREATE_FAILED + - CREATING + - DISCONNECTED + - ESTABLISHED + - EXPIRED + - INIT + - PENDING + - REJECTED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + NccPrivateEndpointRulePrivateLinkConnectionState: + type: string + enum: + - CREATE_FAILED + - CREATING + - DISCONNECTED + - ESTABLISHED + - EXPIRED + - PENDING + - REJECTED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PersonalComputeMessageEnum: + type: string + enum: + - DELEGATE + - 'ON' + description: >- + ON: Grants all users in all workspaces access to the Personal Compute + default policy, allowing + + all users to create single-machine compute resources. DELEGATE: Moves + access control for the + + Personal Compute default policy to individual workspaces and requires a + workspace’s users or + + groups to be added to the ACLs of that workspace’s Personal Compute + default policy before they + + will be able to create compute resources through that policy. + RestrictWorkspaceAdminsMessageStatus: + type: string + enum: + - ALLOW_ALL + - RESTRICT_TOKENS_AND_JOB_RUN_AS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TokenPermissionLevel: + type: string + enum: + - CAN_USE + description: Permission level + TokenType: + type: string + enum: + - ARCLIGHT_AZURE_EXCHANGE_TOKEN + - ARCLIGHT_AZURE_EXCHANGE_TOKEN_WITH_USER_DELEGATION_KEY + - ARCLIGHT_MULTI_TENANT_AZURE_EXCHANGE_TOKEN + - ARCLIGHT_MULTI_TENANT_AZURE_EXCHANGE_TOKEN_WITH_USER_DELEGATION_KEY + - AZURE_ACTIVE_DIRECTORY_TOKEN + description: >- + The type of token request. As of now, only + `AZURE_ACTIVE_DIRECTORY_TOKEN` is supported. + x-stackQL-resources: + aibi_dashboard_embedding_access_policy: + id: databricks_workspace.settings.aibi_dashboard_embedding_access_policy + name: aibi_dashboard_embedding_access_policy + title: Aibi Dashboard Embedding Access Policy + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_acc_policy~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_acc_policy~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_acc_policy~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/aibi_dashboard_embedding_access_policy/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/aibi_dashboard_embedding_access_policy/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/aibi_dashboard_embedding_access_policy/methods/delete + replace: [] + aibi_dashboard_embedding_approved_domains: + id: databricks_workspace.settings.aibi_dashboard_embedding_approved_domains + name: aibi_dashboard_embedding_approved_domains + title: Aibi Dashboard Embedding Approved Domains + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_apprvd_domains~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_apprvd_domains~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_apprvd_domains~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/aibi_dashboard_embedding_approved_domains/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/aibi_dashboard_embedding_approved_domains/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/aibi_dashboard_embedding_approved_domains/methods/delete + replace: [] + automatic_cluster_update: + id: databricks_workspace.settings.automatic_cluster_update + name: automatic_cluster_update + title: Automatic Cluster Update + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1automatic_cluster_update~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1automatic_cluster_update~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/automatic_cluster_update/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/automatic_cluster_update/methods/update + delete: [] + replace: [] + compliance_security_profile: + id: databricks_workspace.settings.compliance_security_profile + name: compliance_security_profile + title: Compliance Security Profile + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1shield_csp_enablement_ws_db~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1shield_csp_enablement_ws_db~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/compliance_security_profile/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/compliance_security_profile/methods/update + delete: [] + replace: [] + credentials_manager: + id: databricks_workspace.settings.credentials_manager + name: credentials_manager + title: Credentials Manager + methods: + exchange_token: + operation: + $ref: >- + #/paths/~1api~12.0~1credentials-manager~1exchange-tokens~1token/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: >- + #/components/x-stackQL-resources/credentials_manager/methods/exchange_token + update: [] + delete: [] + replace: [] + csp_enablement_account: + id: databricks_workspace.settings.csp_enablement_account + name: csp_enablement_account + title: Csp Enablement Account + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_csp_enablement_ac~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_csp_enablement_ac~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/csp_enablement_account/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/csp_enablement_account/methods/update + delete: [] + replace: [] + dashboard_email_subscriptions: + id: databricks_workspace.settings.dashboard_email_subscriptions + name: dashboard_email_subscriptions + title: Dashboard Email Subscriptions + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1dashboard_email_subscriptions~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1dashboard_email_subscriptions~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1dashboard_email_subscriptions~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/dashboard_email_subscriptions/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/dashboard_email_subscriptions/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/dashboard_email_subscriptions/methods/delete + replace: [] + default_namespace: + id: databricks_workspace.settings.default_namespace + name: default_namespace + title: Default Namespace + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1default_namespace_ws~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1default_namespace_ws~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1default_namespace_ws~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/default_namespace/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/default_namespace/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/default_namespace/methods/delete' + replace: [] + default_warehouse_id: + id: databricks_workspace.settings.default_warehouse_id + name: default_warehouse_id + title: Default Warehouse Id + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1default_warehouse_id~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1default_warehouse_id~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1default_warehouse_id~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/default_warehouse_id/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/default_warehouse_id/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/default_warehouse_id/methods/delete + replace: [] + disable_legacy_access: + id: databricks_workspace.settings.disable_legacy_access + name: disable_legacy_access + title: Disable Legacy Access + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1disable_legacy_access~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1disable_legacy_access~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1disable_legacy_access~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/disable_legacy_access/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/disable_legacy_access/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/disable_legacy_access/methods/delete + replace: [] + disable_legacy_dbfs: + id: databricks_workspace.settings.disable_legacy_dbfs + name: disable_legacy_dbfs + title: Disable Legacy Dbfs + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1disable_legacy_dbfs~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1disable_legacy_dbfs~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1disable_legacy_dbfs~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/disable_legacy_dbfs/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/disable_legacy_dbfs/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/disable_legacy_dbfs/methods/delete + replace: [] + disable_legacy_features: + id: databricks_workspace.settings.disable_legacy_features + name: disable_legacy_features + title: Disable Legacy Features + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1disable_legacy_features~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1disable_legacy_features~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1disable_legacy_features~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/disable_legacy_features/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/disable_legacy_features/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/disable_legacy_features/methods/delete + replace: [] + enable_export_notebook: + id: databricks_workspace.settings.enable_export_notebook + name: enable_export_notebook + title: Enable Export Notebook + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1enable-export-notebook~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1enable-export-notebook~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/enable_export_notebook/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/enable_export_notebook/methods/update + delete: [] + replace: [] + enable_ip_access_lists: + id: databricks_workspace.settings.enable_ip_access_lists + name: enable_ip_access_lists + title: Enable Ip Access Lists + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1acct_ip_acl_enable~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1acct_ip_acl_enable~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1acct_ip_acl_enable~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/enable_ip_access_lists/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/enable_ip_access_lists/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/enable_ip_access_lists/methods/delete + replace: [] + enable_notebook_table_clipboard: + id: databricks_workspace.settings.enable_notebook_table_clipboard + name: enable_notebook_table_clipboard + title: Enable Notebook Table Clipboard + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1enable-notebook-table-clipboard~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1enable-notebook-table-clipboard~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/enable_notebook_table_clipboard/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/enable_notebook_table_clipboard/methods/update + delete: [] + replace: [] + enable_results_downloading: + id: databricks_workspace.settings.enable_results_downloading + name: enable_results_downloading + title: Enable Results Downloading + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1enable-results-downloading~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1enable-results-downloading~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/enable_results_downloading/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/enable_results_downloading/methods/update + delete: [] + replace: [] + enhanced_security_monitoring: + id: databricks_workspace.settings.enhanced_security_monitoring + name: enhanced_security_monitoring + title: Enhanced Security Monitoring + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1shield_esm_enablement_ws_db~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1shield_esm_enablement_ws_db~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/enhanced_security_monitoring/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/enhanced_security_monitoring/methods/update + delete: [] + replace: [] + esm_enablement_account: + id: databricks_workspace.settings.esm_enablement_account + name: esm_enablement_account + title: Esm Enablement Account + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_esm_enablement_ac~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1shield_esm_enablement_ac~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/esm_enablement_account/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/esm_enablement_account/methods/update + delete: [] + replace: [] + ip_access_lists: + id: databricks_workspace.settings.ip_access_lists + name: ip_access_lists + title: Ip Access Lists + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1ip-access-lists/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1ip-access-lists/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.ip_access_lists + delete: + operation: + $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/get' + - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/replace' + llm_proxy_partner_powered_account: + id: databricks_workspace.settings.llm_proxy_partner_powered_account + name: llm_proxy_partner_powered_account + title: Llm Proxy Partner Powered Account + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1llm_proxy_partner_powered~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1llm_proxy_partner_powered~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/llm_proxy_partner_powered_account/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/llm_proxy_partner_powered_account/methods/update + delete: [] + replace: [] + llm_proxy_partner_powered_enforce: + id: databricks_workspace.settings.llm_proxy_partner_powered_enforce + name: llm_proxy_partner_powered_enforce + title: Llm Proxy Partner Powered Enforce + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1llm_proxy_partner_powered_enforce~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1llm_proxy_partner_powered_enforce~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/llm_proxy_partner_powered_enforce/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/llm_proxy_partner_powered_enforce/methods/update + delete: [] + replace: [] + llm_proxy_partner_powered_workspace: + id: databricks_workspace.settings.llm_proxy_partner_powered_workspace + name: llm_proxy_partner_powered_workspace + title: Llm Proxy Partner Powered Workspace + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1llm_proxy_partner_powered~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1llm_proxy_partner_powered~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1llm_proxy_partner_powered~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/llm_proxy_partner_powered_workspace/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/llm_proxy_partner_powered_workspace/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/llm_proxy_partner_powered_workspace/methods/delete + replace: [] + notification_destinations: + id: databricks_workspace.settings.notification_destinations + name: notification_destinations + title: Notification Destinations + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1notification-destinations/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1notification-destinations/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.results + delete: + operation: + $ref: '#/paths/~1api~12.0~1notification-destinations~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1notification-destinations~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1notification-destinations~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/notification_destinations/methods/get + - $ref: >- + #/components/x-stackQL-resources/notification_destinations/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/notification_destinations/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/notification_destinations/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/notification_destinations/methods/delete + replace: [] + personal_compute: + id: databricks_workspace.settings.personal_compute + name: personal_compute + title: Personal Compute + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1dcp_acct_enable~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1dcp_acct_enable~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1accounts~1{account_id}~1settings~1types~1dcp_acct_enable~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/personal_compute/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/personal_compute/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/personal_compute/methods/delete' + replace: [] + restrict_workspace_admins: + id: databricks_workspace.settings.restrict_workspace_admins + name: restrict_workspace_admins + title: Restrict Workspace Admins + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1restrict_workspace_admins~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1restrict_workspace_admins~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1restrict_workspace_admins~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/restrict_workspace_admins/methods/get + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/restrict_workspace_admins/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/restrict_workspace_admins/methods/delete + replace: [] + sql_results_download: + id: databricks_workspace.settings.sql_results_download + name: sql_results_download + title: Sql Results Download + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1sql_results_download~1names~1default/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1sql_results_download~1names~1default/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1settings~1types~1sql_results_download~1names~1default/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/sql_results_download/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/sql_results_download/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/sql_results_download/methods/delete + replace: [] + token_management: + id: databricks_workspace.settings.token_management + name: token_management + title: Token Management + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1token-management~1on-behalf-of~1tokens/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1token-management~1tokens~1{token_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1token-management~1tokens~1{token_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1token-management~1tokens/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.token_infos + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/token_management/methods/get' + - $ref: '#/components/x-stackQL-resources/token_management/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/token_management/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/token_management/methods/delete' + replace: [] + token_permission_levels: + id: databricks_workspace.settings.token_permission_levels + name: token_permission_levels + title: Token Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1authorization~1tokens~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/token_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + token_permissions: + id: databricks_workspace.settings.token_permissions + name: token_permissions + title: Token Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1authorization~1tokens/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1authorization~1tokens/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1authorization~1tokens/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/token_permissions/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/token_permissions/methods/update' + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/token_permissions/methods/set' + tokens: + id: databricks_workspace.settings.tokens + name: tokens + title: Tokens + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1token~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1token~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1token~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.token_infos + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/tokens/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/tokens/methods/create' + update: [] + delete: [] + replace: [] + workspace_config: + id: databricks_workspace.settings.workspace_config + name: workspace_config + title: Workspace Config + methods: + get_workspace_config: + operation: + $ref: '#/paths/~1api~12.0~1workspace-conf/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set_workspace_config: + operation: + $ref: '#/paths/~1api~12.0~1workspace-conf/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: [] + update: [] + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/settingsv2.yaml b/providers/src/databricks_workspace/v00.00.00000/services/settingsv2.yaml new file mode 100644 index 00000000..bab5ddbf --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/settingsv2.yaml @@ -0,0 +1,707 @@ +openapi: 3.0.0 +info: + title: Databricks Settingsv2 API (workspace) + description: >- + OpenAPI specification for the Databricks settingsv2 service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.settingsv2 +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.1/settings/{name}: + get: + operationId: workspace_settings_v2_get_public_workspace_setting + summary: >- + Get a setting value at workspace level. See + :method:settingsv2/listworkspacesettingsmetadata for list + tags: + - settingsv2 + - workspace_settings_v2 + description: >- + Get a setting value at workspace level. See + :method:settingsv2/listworkspacesettingsmetadata for list + + of setting available via public APIs. + + + :param name: str + Name of the setting + + :returns: :class:`Setting` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the setting + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Setting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceSettingsV2API + patch: + operationId: workspace_settings_v2_patch_public_workspace_setting + summary: >- + Patch a setting value at workspace level. See + :method:settingsv2/listworkspacesettingsmetadata for + tags: + - settingsv2 + - workspace_settings_v2 + description: >- + Patch a setting value at workspace level. See + :method:settingsv2/listworkspacesettingsmetadata for + + list of setting available via public APIs at workspace level. To + determine the correct field to + + include in a patch request, refer to the type field of the setting + returned in the + + :method:settingsv2/listworkspacesettingsmetadata response. + + + Note: Page refresh is required for changes to take effect in UI. + + + :param name: str + Name of the setting + :param setting: :class:`Setting` + + + :returns: :class:`Setting` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the setting + requestBody: + content: + application/json: + schema: + type: object + properties: + setting: + type: string + description: ':returns: :class:`Setting`' + required: + - setting + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Setting' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceSettingsV2API + /api/2.1/settings-metadata: + get: + operationId: workspace_settings_v2_list_workspace_settings_metadata + summary: >- + List valid setting keys and metadata. These settings are available to be + referenced via GET + tags: + - settingsv2 + - workspace_settings_v2 + description: >- + List valid setting keys and metadata. These settings are available to be + referenced via GET + + :method:settingsv2/getpublicworkspacesetting and PATCH + :method:settingsv2/patchpublicworkspacesetting + + APIs + + + :param page_size: int (optional) + The maximum number of settings to return. The service may return fewer than this value. If + unspecified, at most 200 settings will be returned. The maximum value is 1000; values above 1000 + will be coerced to 1000. + :param page_token: str (optional) + A page token, received from a previous `ListWorkspaceSettingsMetadataRequest` call. Provide this to + retrieve the subsequent page. + + When paginating, all other parameters provided to `ListWorkspaceSettingsMetadataRequest` must match + the call that provided the page token. + + :returns: Iterator over :class:`SettingsMetadata` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of settings to return. The service may return + fewer than this value. If unspecified, at most 200 settings will be + returned. The maximum value is 1000; values above 1000 will be + coerced to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous + `ListWorkspaceSettingsMetadataRequest` call. Provide this to + retrieve the subsequent page. When paginating, all other parameters + provided to `ListWorkspaceSettingsMetadataRequest` must match the + call that provided the page token. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAccountSettingsMetadataResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceSettingsV2API +components: + schemas: + AibiDashboardEmbeddingAccessPolicy: + type: object + properties: + access_policy_type: + $ref: >- + #/components/schemas/AibiDashboardEmbeddingAccessPolicyAccessPolicyType + required: + - access_policy_type + AibiDashboardEmbeddingApprovedDomains: + type: object + properties: + approved_domains: + type: array + items: + type: string + BooleanMessage: + type: object + properties: + value: + type: boolean + ClusterAutoRestartMessage: + type: object + properties: + can_toggle: + type: boolean + enabled: + type: boolean + enablement_details: + $ref: '#/components/schemas/ClusterAutoRestartMessageEnablementDetails' + maintenance_window: + $ref: '#/components/schemas/ClusterAutoRestartMessageMaintenanceWindow' + restart_even_if_no_updates_available: + type: boolean + ClusterAutoRestartMessageEnablementDetails: + type: object + properties: + forced_for_compliance_mode: + type: boolean + description: The feature is force enabled if compliance mode is active + unavailable_for_disabled_entitlement: + type: boolean + description: >- + The feature is unavailable if the corresponding entitlement disabled + (see getShieldEntitlementEnable) + unavailable_for_non_enterprise_tier: + type: boolean + description: >- + The feature is unavailable if the customer doesn't have enterprise + tier + description: >- + Contains an information about the enablement status judging (e.g. + whether the enterprise tier is + enabled) This is only additional information that MUST NOT be used to decide whether the setting + is enabled or not. This is intended to use only for purposes like showing an error message to + the customer with the additional details. For example, using these details we can check why + exactly the feature is disabled for this customer. + ClusterAutoRestartMessageMaintenanceWindow: + type: object + properties: + week_day_based_schedule: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule + ClusterAutoRestartMessageMaintenanceWindowWeekDayBasedSchedule: + type: object + properties: + day_of_week: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowDayOfWeek + frequency: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency + window_start_time: + $ref: >- + #/components/schemas/ClusterAutoRestartMessageMaintenanceWindowWindowStartTime + ClusterAutoRestartMessageMaintenanceWindowWindowStartTime: + type: object + properties: + hours: + type: integer + minutes: + type: integer + IntegerMessage: + type: object + properties: + value: + type: integer + ListAccountSettingsMetadataResponse: + type: object + properties: + next_page_token: + type: string + settings_metadata: + type: array + items: + $ref: '#/components/schemas/SettingsMetadata' + description: List of all settings available via public APIs and their metadata + ListAccountUserPreferencesMetadataResponse: + type: object + properties: + next_page_token: + type: string + settings_metadata: + type: array + items: + $ref: '#/components/schemas/SettingsMetadata' + description: List of all settings available via public APIs and their metadata + ListWorkspaceSettingsMetadataResponse: + type: object + properties: + next_page_token: + type: string + settings_metadata: + type: array + items: + $ref: '#/components/schemas/SettingsMetadata' + description: List of all settings available via public APIs and their metadata + PersonalComputeMessage: + type: object + properties: + value: + $ref: >- + #/components/schemas/PersonalComputeMessagePersonalComputeMessageEnum + RestrictWorkspaceAdminsMessage: + type: object + properties: + status: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessageStatus' + required: + - status + Setting: + type: object + properties: + aibi_dashboard_embedding_access_policy: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicy' + aibi_dashboard_embedding_approved_domains: + $ref: '#/components/schemas/AibiDashboardEmbeddingApprovedDomains' + description: >- + Setting value for aibi_dashboard_embedding_approved_domains setting. + This is the setting value set by consumers, check + effective_aibi_dashboard_embedding_approved_domains for final + setting value. + automatic_cluster_update_workspace: + $ref: '#/components/schemas/ClusterAutoRestartMessage' + description: >- + Setting value for automatic_cluster_update_workspace setting. This + is the setting value set by consumers, check + effective_automatic_cluster_update_workspace for final setting + value. + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + description: >- + Setting value for boolean type setting. This is the setting value + set by consumers, check effective_boolean_val for final setting + value. + effective_aibi_dashboard_embedding_access_policy: + $ref: '#/components/schemas/AibiDashboardEmbeddingAccessPolicy' + description: >- + Effective setting value for aibi_dashboard_embedding_access_policy + setting. This is the final effective value of setting. To set a + value use aibi_dashboard_embedding_access_policy. + effective_aibi_dashboard_embedding_approved_domains: + $ref: '#/components/schemas/AibiDashboardEmbeddingApprovedDomains' + description: >- + Effective setting value for + aibi_dashboard_embedding_approved_domains setting. This is the final + effective value of setting. To set a value use + aibi_dashboard_embedding_approved_domains. + effective_automatic_cluster_update_workspace: + $ref: '#/components/schemas/ClusterAutoRestartMessage' + description: >- + Effective setting value for automatic_cluster_update_workspace + setting. This is the final effective value of setting. To set a + value use automatic_cluster_update_workspace. + effective_boolean_val: + $ref: '#/components/schemas/BooleanMessage' + description: >- + Effective setting value for boolean type setting. This is the final + effective value of setting. To set a value use boolean_val. + effective_integer_val: + $ref: '#/components/schemas/IntegerMessage' + description: >- + Effective setting value for integer type setting. This is the final + effective value of setting. To set a value use integer_val. + effective_personal_compute: + $ref: '#/components/schemas/PersonalComputeMessage' + description: >- + Effective setting value for personal_compute setting. This is the + final effective value of setting. To set a value use + personal_compute. + effective_restrict_workspace_admins: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessage' + description: >- + Effective setting value for restrict_workspace_admins setting. This + is the final effective value of setting. To set a value use + restrict_workspace_admins. + effective_string_val: + $ref: '#/components/schemas/StringMessage' + description: >- + Effective setting value for string type setting. This is the final + effective value of setting. To set a value use string_val. + integer_val: + $ref: '#/components/schemas/IntegerMessage' + description: >- + Setting value for integer type setting. This is the setting value + set by consumers, check effective_integer_val for final setting + value. + name: + type: string + description: Name of the setting. + personal_compute: + $ref: '#/components/schemas/PersonalComputeMessage' + description: >- + Setting value for personal_compute setting. This is the setting + value set by consumers, check effective_personal_compute for final + setting value. + restrict_workspace_admins: + $ref: '#/components/schemas/RestrictWorkspaceAdminsMessage' + description: >- + Setting value for restrict_workspace_admins setting. This is the + setting value set by consumers, check + effective_restrict_workspace_admins for final setting value. + string_val: + $ref: '#/components/schemas/StringMessage' + description: >- + Setting value for string type setting. This is the setting value set + by consumers, check effective_string_val for final setting value. + SettingsMetadata: + type: object + properties: + description: + type: string + docs_link: + type: string + description: Link to databricks documentation for the setting + name: + type: string + description: Name of the setting. + type: + type: string + description: >- + Sample message depicting the type of the setting. To set this + setting, the value sent must match this type. + StringMessage: + type: object + properties: + value: + type: string + UserPreference: + type: object + properties: + boolean_val: + $ref: '#/components/schemas/BooleanMessage' + effective_boolean_val: + $ref: '#/components/schemas/BooleanMessage' + effective_string_val: + $ref: '#/components/schemas/StringMessage' + name: + type: string + description: Name of the setting. + string_val: + $ref: '#/components/schemas/StringMessage' + user_id: + type: string + description: User ID of the user. + description: >- + User Preference represents a user-specific setting scoped to an + individual user within an + account. Unlike workspace or account settings that apply to all users, user preferences allow + personal customization (e.g., UI theme, editor preferences) without affecting other users. + AibiDashboardEmbeddingAccessPolicyAccessPolicyType: + type: string + enum: + - ALLOW_ALL_DOMAINS + - ALLOW_APPROVED_DOMAINS + - DENY_ALL_DOMAINS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowDayOfWeek: + type: string + enum: + - FRIDAY + - MONDAY + - SATURDAY + - SUNDAY + - THURSDAY + - TUESDAY + - WEDNESDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ClusterAutoRestartMessageMaintenanceWindowWeekDayFrequency: + type: string + enum: + - EVERY_WEEK + - FIRST_AND_THIRD_OF_MONTH + - FIRST_OF_MONTH + - FOURTH_OF_MONTH + - SECOND_AND_FOURTH_OF_MONTH + - SECOND_OF_MONTH + - THIRD_OF_MONTH + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PersonalComputeMessagePersonalComputeMessageEnum: + type: string + enum: + - DELEGATE + - 'ON' + description: >- + ON: Grants all users in all workspaces access to the Personal Compute + default policy, allowing + + all users to create single-machine compute resources. DELEGATE: Moves + access control for the + + Personal Compute default policy to individual workspaces and requires a + workspace’s users or + + groups to be added to the ACLs of that workspace’s Personal Compute + default policy before they + + will be able to create compute resources through that policy. + RestrictWorkspaceAdminsMessageStatus: + type: string + enum: + - ALLOW_ALL + - RESTRICT_TOKENS_AND_JOB_RUN_AS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + workspace_settings_v2: + id: databricks_workspace.settingsv2.workspace_settings_v2 + name: workspace_settings_v2 + title: Workspace Settings V2 + methods: + get: + operation: + $ref: '#/paths/~1api~12.1~1settings~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + patch: + operation: + $ref: '#/paths/~1api~12.1~1settings~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1settings-metadata/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.settings_metadata + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/workspace_settings_v2/methods/get' + - $ref: >- + #/components/x-stackQL-resources/workspace_settings_v2/methods/list + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/workspace_settings_v2/methods/patch + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/sharing.yaml b/providers/src/databricks_workspace/v00.00.00000/services/sharing.yaml new file mode 100644 index 00000000..c9f7caf9 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/sharing.yaml @@ -0,0 +1,3550 @@ +openapi: 3.0.0 +info: + title: Databricks Sharing API (workspace) + description: >- + OpenAPI specification for the Databricks sharing service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.sharing +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.1/unity-catalog/providers: + post: + operationId: providers_create + summary: >- + Creates a new authentication provider minimally based on a name and + authentication type. The caller + tags: + - sharing + - providers + description: >- + Creates a new authentication provider minimally based on a name and + authentication type. The caller + + must be an admin on the metastore. + + + :param name: str + The name of the Provider. + :param authentication_type: :class:`AuthenticationType` + + :param comment: str (optional) + Description about the provider. + :param recipient_profile_str: str (optional) + This field is required when the __authentication_type__ is **TOKEN**, **OAUTH_CLIENT_CREDENTIALS** + or not provided. + + :returns: :class:`ProviderInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: The name of the Provider. + authentication_type: + type: string + description: >- + :param comment: str (optional) Description about the + provider. + comment: + type: string + recipient_profile_str: + type: string + description: >- + This field is required when the __authentication_type__ is + **TOKEN**, **OAUTH_CLIENT_CREDENTIALS** or not provided. + required: + - name + - authentication_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ProviderInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProvidersAPI + get: + operationId: providers_list + summary: >- + Gets an array of available authentication providers. The caller must + either be a metastore admin, have + tags: + - sharing + - providers + description: >- + Gets an array of available authentication providers. The caller must + either be a metastore admin, have + + the **USE_PROVIDER** privilege on the providers, or be the owner of the + providers. Providers not owned + + by the caller and for which the caller does not have the + **USE_PROVIDER** privilege are not included + + in the response. There is no guarantee of a specific ordering of the + elements in the array. + + + :param data_provider_global_metastore_id: str (optional) + If not provided, all providers will be returned. If no providers exist with this ID, no results will + be returned. + :param max_results: int (optional) + Maximum number of providers to return. - when set to 0, the page length is set to a server + configured value (recommended); - when set to a value greater than 0, the page length is the minimum + of this value and a server configured value; - when set to a value less than 0, an invalid parameter + error is returned; - If not set, all valid providers are returned (not recommended). - Note: The + number of returned providers might be less than the specified max_results size, even zero. The only + definitive indication that no further providers can be fetched is when the next_page_token is unset + from the response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ProviderInfo` + parameters: + - name: data_provider_global_metastore_id + in: query + required: false + schema: + type: string + description: >- + If not provided, all providers will be returned. If no providers + exist with this ID, no results will be returned. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of providers to return. - when set to 0, the page + length is set to a server configured value (recommended); - when set + to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than + 0, an invalid parameter error is returned; - If not set, all valid + providers are returned (not recommended). - Note: The number of + returned providers might be less than the specified max_results + size, even zero. The only definitive indication that no further + providers can be fetched is when the next_page_token is unset from + the response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListProvidersResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProvidersAPI + /api/2.1/unity-catalog/providers/{name}: + delete: + operationId: providers_delete + summary: >- + Deletes an authentication provider, if the caller is a metastore admin + or is the owner of the + tags: + - sharing + - providers + description: >- + Deletes an authentication provider, if the caller is a metastore admin + or is the owner of the + + provider. + + + :param name: str + Name of the provider. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the provider. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProvidersAPI + get: + operationId: providers_get + summary: >- + Gets a specific authentication provider. The caller must supply the name + of the provider, and must + tags: + - sharing + - providers + description: >- + Gets a specific authentication provider. The caller must supply the name + of the provider, and must + + either be a metastore admin or the owner of the provider. + + + :param name: str + Name of the provider. + + :returns: :class:`ProviderInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the provider. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ProviderInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProvidersAPI + patch: + operationId: providers_update + summary: >- + Updates the information for an authentication provider, if the caller is + a metastore admin or is the + tags: + - sharing + - providers + description: >- + Updates the information for an authentication provider, if the caller is + a metastore admin or is the + + owner of the provider. If the update changes the provider name, the + caller must be both a metastore + + admin and the owner of the provider. + + + :param name: str + Name of the provider. + :param comment: str (optional) + Description about the provider. + :param new_name: str (optional) + New name for the provider. + :param owner: str (optional) + Username of Provider owner. + :param recipient_profile_str: str (optional) + This field is required when the __authentication_type__ is **TOKEN**, **OAUTH_CLIENT_CREDENTIALS** + or not provided. + + :returns: :class:`ProviderInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the provider. + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: Description about the provider. + new_name: + type: string + description: New name for the provider. + owner: + type: string + description: Username of Provider owner. + recipient_profile_str: + type: string + description: >- + This field is required when the __authentication_type__ is + **TOKEN**, **OAUTH_CLIENT_CREDENTIALS** or not provided. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ProviderInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProvidersAPI + /api/2.1/data-sharing/providers/{provider_name}/shares/{share_name}: + get: + operationId: providers_list_provider_share_assets + summary: >- + Get arrays of assets associated with a specified provider's share. The + caller is the recipient of the + tags: + - sharing + - providers + description: >- + Get arrays of assets associated with a specified provider's share. The + caller is the recipient of the + + share. + + + :param provider_name: str + The name of the provider who owns the share. + :param share_name: str + The name of the share. + :param function_max_results: int (optional) + Maximum number of functions to return. + :param notebook_max_results: int (optional) + Maximum number of notebooks to return. + :param table_max_results: int (optional) + Maximum number of tables to return. + :param volume_max_results: int (optional) + Maximum number of volumes to return. + + :returns: :class:`ListProviderShareAssetsResponse` + parameters: + - name: provider_name + in: path + required: true + schema: + type: string + description: The name of the provider who owns the share. + - name: share_name + in: path + required: true + schema: + type: string + description: The name of the share. + - name: function_max_results + in: query + required: false + schema: + type: string + description: Maximum number of functions to return. + - name: notebook_max_results + in: query + required: false + schema: + type: string + description: Maximum number of notebooks to return. + - name: table_max_results + in: query + required: false + schema: + type: string + description: Maximum number of tables to return. + - name: volume_max_results + in: query + required: false + schema: + type: string + description: Maximum number of volumes to return. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListProviderShareAssetsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProvidersAPI + /api/2.1/unity-catalog/providers/{name}/shares: + get: + operationId: providers_list_shares + summary: >- + Gets an array of a specified provider's shares within the metastore + where: + tags: + - sharing + - providers + description: >- + Gets an array of a specified provider's shares within the metastore + where: + + + * the caller is a metastore admin, or * the caller is the owner. + + + :param name: str + Name of the provider in which to list shares. + :param max_results: int (optional) + Maximum number of shares to return. - when set to 0, the page length is set to a server configured + value (recommended); - when set to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than 0, an invalid parameter error + is returned; - If not set, all valid shares are returned (not recommended). - Note: The number of + returned shares might be less than the specified max_results size, even zero. The only definitive + indication that no further shares can be fetched is when the next_page_token is unset from the + response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ProviderShare` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the provider in which to list shares. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of shares to return. - when set to 0, the page length + is set to a server configured value (recommended); - when set to a + value greater than 0, the page length is the minimum of this value + and a server configured value; - when set to a value less than 0, an + invalid parameter error is returned; - If not set, all valid shares + are returned (not recommended). - Note: The number of returned + shares might be less than the specified max_results size, even zero. + The only definitive indication that no further shares can be fetched + is when the next_page_token is unset from the response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListProviderSharesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ProvidersAPI + /api/2.1/unity-catalog/public/data_sharing_activation_info/{activation_url}: + get: + operationId: recipient_activation_get_activation_url_info + summary: Gets an activation URL for a share. + tags: + - sharing + - recipient_activation + description: |- + Gets an activation URL for a share. + + :param activation_url: str + The one time activation url. It also accepts activation token. + parameters: + - name: activation_url + in: path + required: true + schema: + type: string + description: The one time activation url. It also accepts activation token. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientActivationAPI + /api/2.1/unity-catalog/public/data_sharing_activation/{activation_url}: + get: + operationId: recipient_activation_retrieve_token + summary: >- + Retrieve access token with an activation url. This is a public API + without any authentication. + tags: + - sharing + - recipient_activation + description: >- + Retrieve access token with an activation url. This is a public API + without any authentication. + + + :param activation_url: str + The one time activation url. It also accepts activation token. + + :returns: :class:`RetrieveTokenResponse` + parameters: + - name: activation_url + in: path + required: true + schema: + type: string + description: The one time activation url. It also accepts activation token. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RetrieveTokenResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientActivationAPI + /api/2.0/data-sharing/recipients/{recipient_name}/federation-policies: + post: + operationId: recipient_federation_policies_create + summary: >- + Create a federation policy for an OIDC_FEDERATION recipient for sharing + data from Databricks to + tags: + - sharing + - recipient_federation_policies + description: >- + Create a federation policy for an OIDC_FEDERATION recipient for sharing + data from Databricks to + + non-Databricks recipients. The caller must be the owner of the + recipient. When sharing data from + + Databricks to non-Databricks clients, you can define a federation policy + to authenticate + + non-Databricks recipients. The federation policy validates OIDC claims + in federated tokens and is + + defined at the recipient level. This enables secretless sharing clients + to authenticate using OIDC + + tokens. + + + Supported scenarios for federation policies: 1. **User-to-Machine (U2M) + flow** (e.g., PowerBI): A user + + accesses a resource using their own identity. 2. **Machine-to-Machine + (M2M) flow** (e.g., OAuth App): + + An OAuth App accesses a resource using its own identity, typically for + tasks like running nightly + + jobs. + + + For an overview, refer to: - Blog post: Overview of feature: + + https://www.databricks.com/blog/announcing-oidc-token-federation-enhanced-delta-sharing-security + + + For detailed configuration guides based on your use case: - Creating a + Federation Policy as a + + provider: + https://docs.databricks.com/en/delta-sharing/create-recipient-oidc-fed - + Configuration and + + usage for Machine-to-Machine (M2M) applications (e.g., Python Delta + Sharing Client): + + https://docs.databricks.com/aws/en/delta-sharing/sharing-over-oidc-m2m - + Configuration and usage for + + User-to-Machine (U2M) applications (e.g., PowerBI): + + https://docs.databricks.com/aws/en/delta-sharing/sharing-over-oidc-u2m + + + :param recipient_name: str + Name of the recipient. This is the name of the recipient for which the policy is being created. + :param policy: :class:`FederationPolicy` + Name of the policy. This is the name of the policy to be created. + + :returns: :class:`FederationPolicy` + parameters: + - name: recipient_name + in: path + required: true + schema: + type: string + description: >- + Name of the recipient. This is the name of the recipient for which + the policy is being created. + requestBody: + content: + application/json: + schema: + type: object + properties: + policy: + type: string + description: >- + Name of the policy. This is the name of the policy to be + created. + required: + - policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientFederationPoliciesAPI + get: + operationId: recipient_federation_policies_list + summary: >- + Lists federation policies for an OIDC_FEDERATION recipient for sharing + data from Databricks to + tags: + - sharing + - recipient_federation_policies + description: >- + Lists federation policies for an OIDC_FEDERATION recipient for sharing + data from Databricks to + + non-Databricks recipients. The caller must have read access to the + recipient. + + + :param recipient_name: str + Name of the recipient. This is the name of the recipient for which the policies are being listed. + :param max_results: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`FederationPolicy` + parameters: + - name: recipient_name + in: path + required: true + schema: + type: string + description: >- + Name of the recipient. This is the name of the recipient for which + the policies are being listed. + - name: max_results + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListFederationPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientFederationPoliciesAPI + /api/2.0/data-sharing/recipients/{recipient_name}/federation-policies/{name}: + delete: + operationId: recipient_federation_policies_delete + summary: >- + Deletes an existing federation policy for an OIDC_FEDERATION recipient. + The caller must be the owner + tags: + - sharing + - recipient_federation_policies + description: >- + Deletes an existing federation policy for an OIDC_FEDERATION recipient. + The caller must be the owner + + of the recipient. + + + :param recipient_name: str + Name of the recipient. This is the name of the recipient for which the policy is being deleted. + :param name: str + Name of the policy. This is the name of the policy to be deleted. + parameters: + - name: recipient_name + in: path + required: true + schema: + type: string + description: >- + Name of the recipient. This is the name of the recipient for which + the policy is being deleted. + - name: name + in: path + required: true + schema: + type: string + description: Name of the policy. This is the name of the policy to be deleted. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientFederationPoliciesAPI + get: + operationId: recipient_federation_policies_get_federation_policy + summary: >- + Reads an existing federation policy for an OIDC_FEDERATION recipient for + sharing data from Databricks + tags: + - sharing + - recipient_federation_policies + description: >- + Reads an existing federation policy for an OIDC_FEDERATION recipient for + sharing data from Databricks + + to non-Databricks recipients. The caller must have read access to the + recipient. + + + :param recipient_name: str + Name of the recipient. This is the name of the recipient for which the policy is being retrieved. + :param name: str + Name of the policy. This is the name of the policy to be retrieved. + + :returns: :class:`FederationPolicy` + parameters: + - name: recipient_name + in: path + required: true + schema: + type: string + description: >- + Name of the recipient. This is the name of the recipient for which + the policy is being retrieved. + - name: name + in: path + required: true + schema: + type: string + description: Name of the policy. This is the name of the policy to be retrieved. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/FederationPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientFederationPoliciesAPI + /api/2.1/unity-catalog/recipients: + post: + operationId: recipients_create + summary: >- + Creates a new recipient with the delta sharing authentication type in + the metastore. The caller must + tags: + - sharing + - recipients + description: >- + Creates a new recipient with the delta sharing authentication type in + the metastore. The caller must + + be a metastore admin or have the **CREATE_RECIPIENT** privilege on the + metastore. + + + :param name: str + Name of Recipient. + :param authentication_type: :class:`AuthenticationType` + + :param comment: str (optional) + Description about the recipient. + :param data_recipient_global_metastore_id: str (optional) + The global Unity Catalog metastore id provided by the data recipient. This field is only present + when the __authentication_type__ is **DATABRICKS**. The identifier is of format + __cloud__:__region__:__metastore-uuid__. + :param expiration_time: int (optional) + Expiration timestamp of the token, in epoch milliseconds. + :param id: str (optional) + [Create,Update:IGN] common - id of the recipient + :param ip_access_list: :class:`IpAccessList` (optional) + IP Access List + :param owner: str (optional) + Username of the recipient owner. + :param properties_kvpairs: :class:`SecurablePropertiesKvPairs` + (optional) + Recipient properties as map of string key-value pairs. When provided in update request, the + specified properties will override the existing properties. To add and remove properties, one would + need to perform a read-modify-write. + :param sharing_code: str (optional) + The one-time sharing code provided by the data recipient. This field is only present when the + __authentication_type__ is **DATABRICKS**. + + :returns: :class:`RecipientInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of Recipient. + authentication_type: + type: string + description: >- + :param comment: str (optional) Description about the + recipient. + comment: + type: string + data_recipient_global_metastore_id: + type: string + description: >- + The global Unity Catalog metastore id provided by the data + recipient. This field is only present when the + __authentication_type__ is **DATABRICKS**. The identifier is + of format __cloud__:__region__:__metastore-uuid__. + expiration_time: + type: string + description: Expiration timestamp of the token, in epoch milliseconds. + id: + type: string + description: '[Create,Update:IGN] common - id of the recipient' + ip_access_list: + type: string + description: IP Access List + owner: + type: string + description: Username of the recipient owner. + properties_kvpairs: + type: string + description: >- + Recipient properties as map of string key-value pairs. When + provided in update request, the specified properties will + override the existing properties. To add and remove + properties, one would need to perform a read-modify-write. + sharing_code: + type: string + description: >- + The one-time sharing code provided by the data recipient. + This field is only present when the __authentication_type__ + is **DATABRICKS**. + required: + - name + - authentication_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RecipientInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientsAPI + get: + operationId: recipients_list + summary: >- + Gets an array of all share recipients within the current metastore + where: + tags: + - sharing + - recipients + description: >- + Gets an array of all share recipients within the current metastore + where: + + + * the caller is a metastore admin, or * the caller is the owner. There + is no guarantee of a specific + + ordering of the elements in the array. + + + :param data_recipient_global_metastore_id: str (optional) + If not provided, all recipients will be returned. If no recipients exist with this ID, no results + will be returned. + :param max_results: int (optional) + Maximum number of recipients to return. - when set to 0, the page length is set to a server + configured value (recommended); - when set to a value greater than 0, the page length is the minimum + of this value and a server configured value; - when set to a value less than 0, an invalid parameter + error is returned; - If not set, all valid recipients are returned (not recommended). - Note: The + number of returned recipients might be less than the specified max_results size, even zero. The only + definitive indication that no further recipients can be fetched is when the next_page_token is unset + from the response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`RecipientInfo` + parameters: + - name: data_recipient_global_metastore_id + in: query + required: false + schema: + type: string + description: >- + If not provided, all recipients will be returned. If no recipients + exist with this ID, no results will be returned. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of recipients to return. - when set to 0, the page + length is set to a server configured value (recommended); - when set + to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than + 0, an invalid parameter error is returned; - If not set, all valid + recipients are returned (not recommended). - Note: The number of + returned recipients might be less than the specified max_results + size, even zero. The only definitive indication that no further + recipients can be fetched is when the next_page_token is unset from + the response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListRecipientsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientsAPI + /api/2.1/unity-catalog/recipients/{name}: + delete: + operationId: recipients_delete + summary: >- + Deletes the specified recipient from the metastore. The caller must be + the owner of the recipient. + tags: + - sharing + - recipients + description: >- + Deletes the specified recipient from the metastore. The caller must be + the owner of the recipient. + + + :param name: str + Name of the recipient. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the recipient. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientsAPI + get: + operationId: recipients_get + summary: >- + Gets a share recipient from the metastore. The caller must be one of: * + A user with **USE_RECIPIENT** + tags: + - sharing + - recipients + description: >- + Gets a share recipient from the metastore. The caller must be one of: * + A user with **USE_RECIPIENT** + + privilege on the metastore * The owner of the share recipient * A + metastore admin + + + :param name: str + Name of the recipient. + + :returns: :class:`RecipientInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the recipient. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RecipientInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientsAPI + patch: + operationId: recipients_update + summary: >- + Updates an existing recipient in the metastore. The caller must be a + metastore admin or the owner of + tags: + - sharing + - recipients + description: >- + Updates an existing recipient in the metastore. The caller must be a + metastore admin or the owner of + + the recipient. If the recipient name will be updated, the user must be + both a metastore admin and the + + owner of the recipient. + + + :param name: str + Name of the recipient. + :param comment: str (optional) + Description about the recipient. + :param expiration_time: int (optional) + Expiration timestamp of the token, in epoch milliseconds. + :param id: str (optional) + [Create,Update:IGN] common - id of the recipient + :param ip_access_list: :class:`IpAccessList` (optional) + IP Access List + :param new_name: str (optional) + New name for the recipient. . + :param owner: str (optional) + Username of the recipient owner. + :param properties_kvpairs: :class:`SecurablePropertiesKvPairs` + (optional) + Recipient properties as map of string key-value pairs. When provided in update request, the + specified properties will override the existing properties. To add and remove properties, one would + need to perform a read-modify-write. + + :returns: :class:`RecipientInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Name of the recipient. + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: Description about the recipient. + expiration_time: + type: string + description: Expiration timestamp of the token, in epoch milliseconds. + id: + type: string + description: '[Create,Update:IGN] common - id of the recipient' + ip_access_list: + type: string + description: IP Access List + new_name: + type: string + description: New name for the recipient. . + owner: + type: string + description: Username of the recipient owner. + properties_kvpairs: + type: string + description: >- + Recipient properties as map of string key-value pairs. When + provided in update request, the specified properties will + override the existing properties. To add and remove + properties, one would need to perform a read-modify-write. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RecipientInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientsAPI + /api/2.1/unity-catalog/recipients/{name}/rotate-token: + post: + operationId: recipients_rotate_token + summary: >- + Refreshes the specified recipient's delta sharing authentication token + with the provided token info. + tags: + - sharing + - recipients + description: >- + Refreshes the specified recipient's delta sharing authentication token + with the provided token info. + + The caller must be the owner of the recipient. + + + :param name: str + The name of the Recipient. + :param existing_token_expire_in_seconds: int + The expiration time of the bearer token in ISO 8601 format. This will set the expiration_time of + existing token only to a smaller timestamp, it cannot extend the expiration_time. Use 0 to expire + the existing token immediately, negative number will return an error. + + :returns: :class:`RecipientInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the Recipient. + requestBody: + content: + application/json: + schema: + type: object + properties: + existing_token_expire_in_seconds: + type: integer + description: >- + The expiration time of the bearer token in ISO 8601 format. + This will set the expiration_time of existing token only to + a smaller timestamp, it cannot extend the expiration_time. + Use 0 to expire the existing token immediately, negative + number will return an error. + required: + - existing_token_expire_in_seconds + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RecipientInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientsAPI + /api/2.1/unity-catalog/recipients/{name}/share-permissions: + get: + operationId: recipients_share_permissions + summary: >- + Gets the share permissions for the specified Recipient. The caller must + have the **USE_RECIPIENT** + tags: + - sharing + - recipients + description: >- + Gets the share permissions for the specified Recipient. The caller must + have the **USE_RECIPIENT** + + privilege on the metastore or be the owner of the Recipient. + + + :param name: str + The name of the Recipient. + :param max_results: int (optional) + Maximum number of permissions to return. - when set to 0, the page length is set to a server + configured value (recommended); - when set to a value greater than 0, the page length is the minimum + of this value and a server configured value; - when set to a value less than 0, an invalid parameter + error is returned; - If not set, all valid permissions are returned (not recommended). - Note: The + number of returned permissions might be less than the specified max_results size, even zero. The + only definitive indication that no further permissions can be fetched is when the next_page_token is + unset from the response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: :class:`GetRecipientSharePermissionsResponse` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the Recipient. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of permissions to return. - when set to 0, the page + length is set to a server configured value (recommended); - when set + to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than + 0, an invalid parameter error is returned; - If not set, all valid + permissions are returned (not recommended). - Note: The number of + returned permissions might be less than the specified max_results + size, even zero. The only definitive indication that no further + permissions can be fetched is when the next_page_token is unset from + the response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetRecipientSharePermissionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RecipientsAPI + /api/2.1/unity-catalog/shares: + post: + operationId: shares_create + summary: >- + Creates a new share for data objects. Data objects can be added after + creation with **update**. The + tags: + - sharing + - shares + description: >- + Creates a new share for data objects. Data objects can be added after + creation with **update**. The + + caller must be a metastore admin or have the **CREATE_SHARE** privilege + on the metastore. + + + :param name: str + Name of the share. + :param comment: str (optional) + User-provided free-form text description. + :param storage_root: str (optional) + Storage root URL for the share. + + :returns: :class:`ShareInfo` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the share. + comment: + type: string + description: User-provided free-form text description. + storage_root: + type: string + description: Storage root URL for the share. + required: + - name + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ShareInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SharesAPI + get: + operationId: shares_list_shares + summary: >- + Gets an array of data object shares from the metastore. If the caller + has the USE_SHARE privilege on + tags: + - sharing + - shares + description: >- + Gets an array of data object shares from the metastore. If the caller + has the USE_SHARE privilege on + + the metastore, all shares are returned. Otherwise, only shares owned by + the caller are returned. There + + is no guarantee of a specific ordering of the elements in the array. + + + :param max_results: int (optional) + Maximum number of shares to return. - when set to 0, the page length is set to a server configured + value (recommended); - when set to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than 0, an invalid parameter error + is returned; - If not set, all valid shares are returned (not recommended). - Note: The number of + returned shares might be less than the specified max_results size, even zero. The only definitive + indication that no further shares can be fetched is when the next_page_token is unset from the + response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: Iterator over :class:`ShareInfo` + parameters: + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of shares to return. - when set to 0, the page length + is set to a server configured value (recommended); - when set to a + value greater than 0, the page length is the minimum of this value + and a server configured value; - when set to a value less than 0, an + invalid parameter error is returned; - If not set, all valid shares + are returned (not recommended). - Note: The number of returned + shares might be less than the specified max_results size, even zero. + The only definitive indication that no further shares can be fetched + is when the next_page_token is unset from the response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListSharesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SharesAPI + /api/2.1/unity-catalog/shares/{name}: + delete: + operationId: shares_delete + summary: >- + Deletes a data object share from the metastore. The caller must be an + owner of the share. + tags: + - sharing + - shares + description: >- + Deletes a data object share from the metastore. The caller must be an + owner of the share. + + + :param name: str + The name of the share. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the share. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SharesAPI + get: + operationId: shares_get + summary: >- + Gets a data object share from the metastore. The caller must have the + USE_SHARE privilege on the + tags: + - sharing + - shares + description: >- + Gets a data object share from the metastore. The caller must have the + USE_SHARE privilege on the + + metastore or be the owner of the share. + + + :param name: str + The name of the share. + :param include_shared_data: bool (optional) + Query for data to include in the share. + + :returns: :class:`ShareInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the share. + - name: include_shared_data + in: query + required: false + schema: + type: string + description: Query for data to include in the share. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ShareInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SharesAPI + patch: + operationId: shares_update + summary: >- + Updates the share with the changes and data objects in the request. The + caller must be the owner of + tags: + - sharing + - shares + description: >- + Updates the share with the changes and data objects in the request. The + caller must be the owner of + + the share or a metastore admin. + + + When the caller is a metastore admin, only the __owner__ field can be + updated. + + + In the case the share name is changed, **updateShare** requires that the + caller is the owner of the + + share and has the CREATE_SHARE privilege. + + + If there are notebook files in the share, the __storage_root__ field + cannot be updated. + + + For each table that is added through this method, the share owner must + also have **SELECT** privilege + + on the table. This privilege must be maintained indefinitely for + recipients to be able to access the + + table. Typically, you should use a group as the share owner. + + + Table removals through **update** do not require additional privileges. + + + :param name: str + The name of the share. + :param comment: str (optional) + User-provided free-form text description. + :param new_name: str (optional) + New name for the share. + :param owner: str (optional) + Username of current owner of share. + :param storage_root: str (optional) + Storage root URL for the share. + :param updates: List[:class:`SharedDataObjectUpdate`] (optional) + Array of shared data object updates. + + :returns: :class:`ShareInfo` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the share. + requestBody: + content: + application/json: + schema: + type: object + properties: + comment: + type: string + description: User-provided free-form text description. + new_name: + type: string + description: New name for the share. + owner: + type: string + description: Username of current owner of share. + storage_root: + type: string + description: Storage root URL for the share. + updates: + type: string + description: Array of shared data object updates. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ShareInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SharesAPI + /api/2.1/unity-catalog/shares/{name}/permissions: + get: + operationId: shares_share_permissions + summary: >- + Gets the permissions for a data share from the metastore. The caller + must have the USE_SHARE privilege + tags: + - sharing + - shares + description: >- + Gets the permissions for a data share from the metastore. The caller + must have the USE_SHARE privilege + + on the metastore or be the owner of the share. + + + :param name: str + The name of the Recipient. + :param max_results: int (optional) + Maximum number of permissions to return. - when set to 0, the page length is set to a server + configured value (recommended); - when set to a value greater than 0, the page length is the minimum + of this value and a server configured value; - when set to a value less than 0, an invalid parameter + error is returned; - If not set, all valid permissions are returned (not recommended). - Note: The + number of returned permissions might be less than the specified max_results size, even zero. The + only definitive indication that no further permissions can be fetched is when the next_page_token is + unset from the response. + :param page_token: str (optional) + Opaque pagination token to go to next page based on previous query. + + :returns: :class:`GetSharePermissionsResponse` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the Recipient. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Maximum number of permissions to return. - when set to 0, the page + length is set to a server configured value (recommended); - when set + to a value greater than 0, the page length is the minimum of this + value and a server configured value; - when set to a value less than + 0, an invalid parameter error is returned; - If not set, all valid + permissions are returned (not recommended). - Note: The number of + returned permissions might be less than the specified max_results + size, even zero. The only definitive indication that no further + permissions can be fetched is when the next_page_token is unset from + the response. + - name: page_token + in: query + required: false + schema: + type: string + description: Opaque pagination token to go to next page based on previous query. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetSharePermissionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SharesAPI + patch: + operationId: shares_update_permissions + summary: >- + Updates the permissions for a data share in the metastore. The caller + must have both the USE_SHARE and + tags: + - sharing + - shares + description: >- + Updates the permissions for a data share in the metastore. The caller + must have both the USE_SHARE and + + SET_SHARE_PERMISSION privileges on the metastore, or be the owner of the + share. + + + For new recipient grants, the user must also be the owner of the + recipients. recipient revocations do + + not require additional privileges. + + + :param name: str + The name of the share. + :param changes: List[:class:`PermissionsChange`] (optional) + Array of permissions change objects. + :param omit_permissions_list: bool (optional) + Optional. Whether to return the latest permissions list of the share in the response. + + :returns: :class:`UpdateSharePermissionsResponse` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: The name of the share. + requestBody: + content: + application/json: + schema: + type: object + properties: + changes: + type: string + description: Array of permissions change objects. + omit_permissions_list: + type: string + description: >- + Optional. Whether to return the latest permissions list of + the share in the response. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateSharePermissionsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SharesAPI +components: + schemas: + DeltaSharingDependency: + type: object + properties: + function: + $ref: '#/components/schemas/DeltaSharingFunctionDependency' + table: + $ref: '#/components/schemas/DeltaSharingTableDependency' + description: Represents a UC dependency. + DeltaSharingDependencyList: + type: object + properties: + dependencies: + type: array + items: + $ref: '#/components/schemas/DeltaSharingDependency' + description: An array of Dependency. + description: Represents a list of dependencies. + DeltaSharingFunction: + type: object + properties: + aliases: + type: array + items: + $ref: '#/components/schemas/RegisteredModelAlias' + comment: + type: string + description: The comment of the function. + data_type: + $ref: '#/components/schemas/ColumnTypeName' + description: The data type of the function. + dependency_list: + $ref: '#/components/schemas/DeltaSharingDependencyList' + description: The dependency list of the function. + full_data_type: + type: string + description: The full data type of the function. + id: + type: string + description: The id of the function. + input_params: + $ref: '#/components/schemas/FunctionParameterInfos' + description: The function parameter information. + name: + type: string + description: The name of the function. + properties: + type: string + description: The properties of the function. + routine_definition: + type: string + description: The routine definition of the function. + schema: + type: string + description: The name of the schema that the function belongs to. + securable_kind: + $ref: '#/components/schemas/SharedSecurableKind' + description: The securable kind of the function. + share: + type: string + description: The name of the share that the function belongs to. + share_id: + type: string + description: The id of the share that the function belongs to. + storage_location: + type: string + description: The storage location of the function. + tags: + type: string + description: The tags of the function. + DeltaSharingFunctionDependency: + type: object + properties: + function_name: + type: string + schema_name: + type: string + description: A Function in UC as a dependency. + DeltaSharingTableDependency: + type: object + properties: + schema_name: + type: string + table_name: + type: string + description: A Table in UC as a dependency. + FederationPolicy: + type: object + properties: + comment: + type: string + create_time: + type: string + description: System-generated timestamp indicating when the policy was created. + id: + type: string + description: >- + Unique, immutable system-generated identifier for the federation + policy. + name: + type: string + description: >- + Name of the federation policy. A recipient can have multiple + policies with different names. The name must contain only lowercase + alphanumeric characters, numbers, and hyphens. + oidc_policy: + $ref: '#/components/schemas/OidcFederationPolicy' + description: >- + Specifies the policy to use for validating OIDC claims in the + federated tokens. + update_time: + type: string + description: >- + System-generated timestamp indicating when the policy was last + updated. + FunctionParameterInfo: + type: object + properties: + comment: + type: string + description: The comment of the parameter. + name: + type: string + description: The name of the parameter. + parameter_default: + type: string + description: The default value of the parameter. + parameter_mode: + $ref: '#/components/schemas/FunctionParameterMode' + description: The mode of the function parameter. + parameter_type: + $ref: '#/components/schemas/FunctionParameterType' + description: The type of the function parameter. + position: + type: integer + description: The position of the parameter. + type_interval_type: + type: string + description: The interval type of the parameter type. + type_json: + type: string + description: The type of the parameter in JSON format. + type_name: + $ref: '#/components/schemas/ColumnTypeName' + description: The type of the parameter in Enum format. + type_precision: + type: integer + description: The precision of the parameter type. + type_scale: + type: integer + description: The scale of the parameter type. + type_text: + type: string + description: The type of the parameter in text format. + description: >- + Represents a parameter of a function. The same message is used for both + input and output + columns. + FunctionParameterInfos: + type: object + properties: + parameters: + type: array + items: + $ref: '#/components/schemas/FunctionParameterInfo' + GetActivationUrlInfoResponse: + type: object + properties: {} + GetRecipientSharePermissionsResponse: + type: object + properties: + next_page_token: + type: string + permissions_out: + type: array + items: + $ref: '#/components/schemas/ShareToPrivilegeAssignment' + description: An array of data share permissions for a recipient. + GetSharePermissionsResponse: + type: object + properties: + next_page_token: + type: string + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/PrivilegeAssignment' + description: The privileges assigned to each principal + IpAccessList: + type: object + properties: + allowed_ip_addresses: + type: array + items: + type: string + ListFederationPoliciesResponse: + type: object + properties: + next_page_token: + type: string + policies: + type: array + items: + $ref: '#/components/schemas/FederationPolicy' + ListProviderShareAssetsResponse: + type: object + properties: + functions: + type: array + items: + $ref: '#/components/schemas/DeltaSharingFunction' + description: The list of functions in the share. + notebooks: + type: array + items: + $ref: '#/components/schemas/NotebookFile' + description: The list of notebooks in the share. + share: + $ref: '#/components/schemas/Share' + description: The metadata of the share. + tables: + type: array + items: + $ref: '#/components/schemas/Table' + description: The list of tables in the share. + volumes: + type: array + items: + $ref: '#/components/schemas/Volume' + description: The list of volumes in the share. + description: >- + Response to ListProviderShareAssets, which contains the list of assets + of a share. + ListProviderSharesResponse: + type: object + properties: + next_page_token: + type: string + shares: + type: array + items: + $ref: '#/components/schemas/ProviderShare' + description: An array of provider shares. + ListProvidersResponse: + type: object + properties: + next_page_token: + type: string + providers: + type: array + items: + $ref: '#/components/schemas/ProviderInfo' + description: An array of provider information objects. + ListRecipientsResponse: + type: object + properties: + next_page_token: + type: string + recipients: + type: array + items: + $ref: '#/components/schemas/RecipientInfo' + description: An array of recipient information objects. + ListSharesResponse: + type: object + properties: + next_page_token: + type: string + shares: + type: array + items: + $ref: '#/components/schemas/ShareInfo' + description: An array of data share information objects. + NotebookFile: + type: object + properties: + comment: + type: string + id: + type: string + description: The id of the notebook file. + name: + type: string + description: Name of the notebook file. + share: + type: string + description: The name of the share that the notebook file belongs to. + share_id: + type: string + description: The id of the share that the notebook file belongs to. + tags: + type: string + description: The tags of the notebook file. + OidcFederationPolicy: + type: object + properties: + issuer: + type: string + description: >- + The required token issuer, as specified in the 'iss' claim of + federated tokens. + subject_claim: + type: string + description: >- + The claim that contains the subject of the token. Depending on the + identity provider and the use case (U2M or M2M), this can vary: - + For Entra ID (AAD): * U2M flow (group access): Use `groups`. * U2M + flow (user access): Use `oid`. * M2M flow (OAuth App access): Use + `azp`. - For other IdPs, refer to the specific IdP documentation. + Supported `subject_claim` values are: - `oid`: Object ID of the + user. - `azp`: Client ID of the OAuth app. - `groups`: Object ID of + the group. - `sub`: Subject identifier for other use cases. + subject: + type: string + description: >- + The required token subject, as specified in the subject claim of + federated tokens. The subject claim identifies the identity of the + user or machine accessing the resource. Examples for Entra ID (AAD): + - U2M flow (group access): If the subject claim is `groups`, this + must be the Object ID of the group in Entra ID. - U2M flow (user + access): If the subject claim is `oid`, this must be the Object ID + of the user in Entra ID. - M2M flow (OAuth App access): If the + subject claim is `azp`, this must be the client ID of the OAuth app + registered in Entra ID. + audiences: + type: array + items: + type: string + description: >- + The allowed token audiences, as specified in the 'aud' claim of + federated tokens. The audience identifier is intended to represent + the recipient of the token. Can be any non-empty string value. As + long as the audience in the token matches at least one audience in + the policy, + required: + - issuer + - subject_claim + - subject + description: >- + Specifies the policy to use for validating OIDC claims in your federated + tokens from Delta + Sharing Clients. Refer to https://docs.databricks.com/en/delta-sharing/create-recipient-oidc-fed + for more details. + Partition: + type: object + properties: + values: + type: array + items: + $ref: '#/components/schemas/PartitionValue' + PartitionValue: + type: object + properties: + name: + type: string + op: + $ref: '#/components/schemas/PartitionValueOp' + description: The operator to apply for the value. + recipient_property_key: + type: string + description: >- + The key of a Delta Sharing recipient's property. For example + "databricks-account-id". When this field is set, field `value` can + not be set. + value: + type: string + description: >- + The value of the partition column. When this value is not set, it + means `null` value. When this field is set, field + `recipient_property_key` can not be set. + PermissionsChange: + type: object + properties: + add: + type: array + items: + type: string + principal: + type: string + description: >- + The principal whose privileges we are changing. Only one of + principal or principal_id should be specified, never both at the + same time. + remove: + type: array + items: + type: string + description: The set of privileges to remove. + PrivilegeAssignment: + type: object + properties: + principal: + type: string + privileges: + type: array + items: + $ref: '#/components/schemas/Privilege' + description: The privileges assigned to the principal. + ProviderInfo: + type: object + properties: + authentication_type: + $ref: '#/components/schemas/AuthenticationType' + cloud: + type: string + description: >- + Cloud vendor of the provider's UC metastore. This field is only + present when the __authentication_type__ is **DATABRICKS**. + comment: + type: string + description: Description about the provider. + created_at: + type: integer + description: Time at which this Provider was created, in epoch milliseconds. + created_by: + type: string + description: Username of Provider creator. + data_provider_global_metastore_id: + type: string + description: >- + The global UC metastore id of the data provider. This field is only + present when the __authentication_type__ is **DATABRICKS**. The + identifier is of format __cloud__:__region__:__metastore-uuid__. + metastore_id: + type: string + description: >- + UUID of the provider's UC metastore. This field is only present when + the __authentication_type__ is **DATABRICKS**. + name: + type: string + description: The name of the Provider. + owner: + type: string + description: Username of Provider owner. + recipient_profile: + $ref: '#/components/schemas/RecipientProfile' + description: >- + The recipient profile. This field is only present when the + authentication_type is `TOKEN` or `OAUTH_CLIENT_CREDENTIALS`. + recipient_profile_str: + type: string + description: >- + This field is required when the __authentication_type__ is + **TOKEN**, **OAUTH_CLIENT_CREDENTIALS** or not provided. + region: + type: string + description: >- + Cloud region of the provider's UC metastore. This field is only + present when the __authentication_type__ is **DATABRICKS**. + updated_at: + type: integer + description: Time at which this Provider was created, in epoch milliseconds. + updated_by: + type: string + description: Username of user who last modified Provider. + ProviderShare: + type: object + properties: + name: + type: string + RecipientInfo: + type: object + properties: + activated: + type: boolean + activation_url: + type: string + description: >- + Full activation url to retrieve the access token. It will be empty + if the token is already retrieved. + authentication_type: + $ref: '#/components/schemas/AuthenticationType' + cloud: + type: string + description: >- + Cloud vendor of the recipient's Unity Catalog Metastore. This field + is only present when the __authentication_type__ is **DATABRICKS**. + comment: + type: string + description: Description about the recipient. + created_at: + type: integer + description: Time at which this recipient was created, in epoch milliseconds. + created_by: + type: string + description: Username of recipient creator. + data_recipient_global_metastore_id: + type: string + description: >- + The global Unity Catalog metastore id provided by the data + recipient. This field is only present when the + __authentication_type__ is **DATABRICKS**. The identifier is of + format __cloud__:__region__:__metastore-uuid__. + expiration_time: + type: integer + description: Expiration timestamp of the token, in epoch milliseconds. + id: + type: string + description: '[Create,Update:IGN] common - id of the recipient' + ip_access_list: + $ref: '#/components/schemas/IpAccessList' + description: IP Access List + metastore_id: + type: string + description: >- + Unique identifier of recipient's Unity Catalog Metastore. This field + is only present when the __authentication_type__ is **DATABRICKS**. + name: + type: string + description: Name of Recipient. + owner: + type: string + description: Username of the recipient owner. + properties_kvpairs: + $ref: '#/components/schemas/SecurablePropertiesKvPairs' + description: >- + Recipient properties as map of string key-value pairs. When provided + in update request, the specified properties will override the + existing properties. To add and remove properties, one would need to + perform a read-modify-write. + region: + type: string + description: >- + Cloud region of the recipient's Unity Catalog Metastore. This field + is only present when the __authentication_type__ is **DATABRICKS**. + sharing_code: + type: string + description: >- + The one-time sharing code provided by the data recipient. This field + is only present when the __authentication_type__ is **DATABRICKS**. + tokens: + type: array + items: + $ref: '#/components/schemas/RecipientTokenInfo' + description: >- + This field is only present when the __authentication_type__ is + **TOKEN**. + updated_at: + type: integer + description: Time at which the recipient was updated, in epoch milliseconds. + updated_by: + type: string + description: Username of recipient updater. + RecipientProfile: + type: object + properties: + bearer_token: + type: string + endpoint: + type: string + description: The endpoint for the share to be used by the recipient. + share_credentials_version: + type: integer + description: The version number of the recipient's credentials on a share. + RecipientTokenInfo: + type: object + properties: + activation_url: + type: string + created_at: + type: integer + description: >- + Time at which this recipient token was created, in epoch + milliseconds. + created_by: + type: string + description: Username of recipient token creator. + expiration_time: + type: integer + description: Expiration timestamp of the token in epoch milliseconds. + id: + type: string + description: Unique ID of the recipient token. + updated_at: + type: integer + description: >- + Time at which this recipient token was updated, in epoch + milliseconds. + updated_by: + type: string + description: Username of recipient token updater. + RegisteredModelAlias: + type: object + properties: + alias_name: + type: string + version_num: + type: integer + description: Numeric model version that alias will reference. + RetrieveTokenResponse: + type: object + properties: + bearerToken: + type: string + endpoint: + type: string + description: The endpoint for the share to be used by the recipient. + expirationTime: + type: string + description: Expiration timestamp of the token in epoch milliseconds. + shareCredentialsVersion: + type: integer + description: These field names must follow the delta sharing protocol. + SecurablePropertiesKvPairs: + type: object + properties: + properties: + type: object + description: A map of key-value properties attached to the securable. + required: + - properties + description: >- + An object with __properties__ containing map of key-value properties + attached to the securable. + Share: + type: object + properties: + id: + type: string + name: + type: string + ShareInfo: + type: object + properties: + comment: + type: string + created_at: + type: integer + description: Time at which this share was created, in epoch milliseconds. + created_by: + type: string + description: Username of share creator. + name: + type: string + description: Name of the share. + objects: + type: array + items: + $ref: '#/components/schemas/SharedDataObject' + description: A list of shared data objects within the share. + owner: + type: string + description: Username of current owner of share. + storage_location: + type: string + description: Storage Location URL (full path) for the share. + storage_root: + type: string + description: Storage root URL for the share. + updated_at: + type: integer + description: Time at which this share was updated, in epoch milliseconds. + updated_by: + type: string + description: Username of share updater. + ShareToPrivilegeAssignment: + type: object + properties: + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/PrivilegeAssignment' + share_name: + type: string + description: The share name. + SharedDataObject: + type: object + properties: + name: + type: string + added_at: + type: integer + description: >- + The time when this data object is added to the share, in epoch + milliseconds. + added_by: + type: string + description: Username of the sharer. + cdf_enabled: + type: boolean + description: >- + Whether to enable cdf or indicate if cdf is enabled on the shared + object. + comment: + type: string + description: A user-provided comment when adding the data object to the share. + content: + type: string + description: >- + The content of the notebook file when the data object type is + NOTEBOOK_FILE. This should be base64 encoded. Required for adding a + NOTEBOOK_FILE, optional for updating, ignored for other types. + data_object_type: + $ref: '#/components/schemas/SharedDataObjectDataObjectType' + description: The type of the data object. + history_data_sharing_status: + $ref: '#/components/schemas/SharedDataObjectHistoryDataSharingStatus' + description: >- + Whether to enable or disable sharing of data history. If not + specified, the default is **DISABLED**. + partitions: + type: array + items: + $ref: '#/components/schemas/Partition' + description: Array of partitions for the shared data. + shared_as: + type: string + description: >- + A user-provided alias name for table-like data objects within the + share. Use this field for table-like objects (for example: TABLE, + VIEW, MATERIALIZED_VIEW, STREAMING_TABLE, FOREIGN_TABLE). For + non-table objects (for example: VOLUME, MODEL, NOTEBOOK_FILE, + FUNCTION), use `string_shared_as` instead. Important: For non-table + objects, this field must be omitted entirely. Format: Must be a + 2-part name `.` (e.g., + "sales_schema.orders_table") - Both schema and table names must + contain only alphanumeric characters and underscores - No periods, + spaces, forward slashes, or control characters are allowed within + each part - Do not include the catalog name (use 2 parts, not 3) + Behavior: - If not provided, the service automatically generates the + alias as `.` from the object's original name - If you + don't want to specify this field, omit it entirely from the request + (do not pass an empty string) - The `shared_as` name must be unique + within the share Examples: - Valid: "analytics_schema.customer_view" + - Invalid: "catalog.analytics_schema.customer_view" (3 parts not + allowed) - Invalid: "analytics-schema.customer-view" (hyphens not + allowed) + start_version: + type: integer + description: >- + The start version associated with the object. This allows data + providers to control the lowest object version that is accessible by + clients. If specified, clients can query snapshots or changes for + versions >= start_version. If not specified, clients can only query + starting from the version of the object at the time it was added to + the share. NOTE: The start_version should be <= the `current` + version of the object. + status: + $ref: '#/components/schemas/SharedDataObjectStatus' + description: 'One of: **ACTIVE**, **PERMISSION_DENIED**.' + string_shared_as: + type: string + description: >- + A user-provided alias name for non-table data objects within the + share. Use this field for non-table objects (for example: VOLUME, + MODEL, NOTEBOOK_FILE, FUNCTION). For table-like objects (for + example: TABLE, VIEW, MATERIALIZED_VIEW, STREAMING_TABLE, + FOREIGN_TABLE), use `shared_as` instead. Important: For table-like + objects, this field must be omitted entirely. Format: - For VOLUME: + Must be a 2-part name `.` (e.g., + "data_schema.ml_models") - For FUNCTION: Must be a 2-part name + `.` (e.g., "udf_schema.calculate_tax") - + For MODEL: Must be a 2-part name `.` (e.g., + "models.prediction_model") - For NOTEBOOK_FILE: Should be the + notebook file name (e.g., "analysis_notebook.py") - All names must + contain only alphanumeric characters and underscores - No periods, + spaces, forward slashes, or control characters are allowed within + each part Behavior: - If not provided, the service automatically + generates the alias from the object's original name - If you don't + want to specify this field, omit it entirely from the request (do + not pass an empty string) - The `string_shared_as` name must be + unique for objects of the same type within the share Examples: - + Valid for VOLUME: "data_schema.training_data" - Valid for FUNCTION: + "analytics.calculate_revenue" - Invalid: + "catalog.data_schema.training_data" (3 parts not allowed for + volumes) - Invalid: "data-schema.training-data" (hyphens not + allowed) + required: + - name + SharedDataObjectUpdate: + type: object + properties: + action: + $ref: '#/components/schemas/SharedDataObjectUpdateAction' + data_object: + $ref: '#/components/schemas/SharedDataObject' + description: >- + The data object that is being added, removed, or updated. The + maximum number update data objects allowed is a 100. + Table: + type: object + properties: + comment: + type: string + id: + type: string + description: The id of the table. + internal_attributes: + $ref: '#/components/schemas/TableInternalAttributes' + description: >- + Internal information for D2D sharing that should not be disclosed to + external users. + materialization_namespace: + type: string + description: The catalog and schema of the materialized table + materialized_table_name: + type: string + description: The name of a materialized table. + name: + type: string + description: The name of the table. + schema: + type: string + description: The name of the schema that the table belongs to. + share: + type: string + description: The name of the share that the table belongs to. + share_id: + type: string + description: The id of the share that the table belongs to. + tags: + type: string + description: The Tags of the table. + TableInternalAttributes: + type: object + properties: + auxiliary_managed_location: + type: string + description: Managed Delta Metadata location for foreign iceberg tables. + dependency_storage_locations: + type: array + items: + type: string + description: >- + Storage locations of all table dependencies for shared views. Used + on the recipient side for SEG (Secure Egress Gateway) whitelisting. + has_delta_uniform_iceberg: + type: boolean + description: Whether the table has uniform enabled. + parent_storage_location: + type: string + description: >- + Will be populated in the reconciliation response for VIEW and + FOREIGN_TABLE, with the value of the parent UC entity's + storage_location, following the same logic as getManagedEntityPath + in CreateStagingTableHandler, which is used to store the + materialized table for a shared VIEW/FOREIGN_TABLE for D2O queries. + The value will be used on the recipient side to be whitelisted when + SEG is enabled on the workspace of the recipient, to allow the + recipient users to query this shared VIEW/FOREIGN_TABLE. + storage_location: + type: string + description: >- + The cloud storage location of a shard table with + DIRECTORY_BASED_TABLE type. + type: + $ref: '#/components/schemas/TableInternalAttributesSharedTableType' + description: The type of the shared table. + view_definition: + type: string + description: The view definition of a shared view. DEPRECATED. + description: >- + Internal information for D2D sharing that should not be disclosed to + external users. + UpdateSharePermissionsResponse: + type: object + properties: + privilege_assignments: + type: array + items: + $ref: '#/components/schemas/PrivilegeAssignment' + Volume: + type: object + properties: + comment: + type: string + id: + type: string + description: >- + This id maps to the shared_volume_id in database Recipient needs + shared_volume_id for recon to check if this volume is already in + recipient's DB or not. + internal_attributes: + $ref: '#/components/schemas/VolumeInternalAttributes' + description: >- + Internal attributes for D2D sharing that should not be disclosed to + external users. + name: + type: string + description: The name of the volume. + schema: + type: string + description: The name of the schema that the volume belongs to. + share: + type: string + description: The name of the share that the volume belongs to. + share_id: + type: string + description: / The id of the share that the volume belongs to. + tags: + type: string + description: The tags of the volume. + VolumeInternalAttributes: + type: object + properties: + storage_location: + type: string + description: The cloud storage location of the volume + type: + type: string + description: The type of the shared volume. + description: >- + Internal information for D2D sharing that should not be disclosed to + external users. + AuthenticationType: + type: string + enum: + - DATABRICKS + - OAUTH_CLIENT_CREDENTIALS + - OIDC_FEDERATION + - TOKEN + description: The delta sharing authentication type. + ColumnTypeName: + type: string + enum: + - ARRAY + - BINARY + - BOOLEAN + - BYTE + - CHAR + - DATE + - DECIMAL + - DOUBLE + - FLOAT + - INT + - INTERVAL + - LONG + - MAP + - 'NULL' + - SHORT + - STRING + - STRUCT + - TABLE_TYPE + - TIMESTAMP + - TIMESTAMP_NTZ + - USER_DEFINED_TYPE + - VARIANT + description: >- + UC supported column types Copied from + + https://src.dev.databricks.com/databricks/universe@23a85902bb58695ab9293adc9f327b0714b55e72/-/blob/managed-catalog/api/messages/table.proto?L68 + FunctionParameterMode: + type: string + enum: + - IN + - INOUT + - OUT + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + FunctionParameterType: + type: string + enum: + - COLUMN + - PARAM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PartitionValueOp: + type: string + enum: + - EQUAL + - LIKE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + Privilege: + type: string + enum: + - ACCESS + - ALL_PRIVILEGES + - APPLY_TAG + - CREATE + - CREATE_CATALOG + - CREATE_CONNECTION + - CREATE_EXTERNAL_LOCATION + - CREATE_EXTERNAL_TABLE + - CREATE_EXTERNAL_VOLUME + - CREATE_FOREIGN_CATALOG + - CREATE_FOREIGN_SECURABLE + - CREATE_FUNCTION + - CREATE_MANAGED_STORAGE + - CREATE_MATERIALIZED_VIEW + - CREATE_MODEL + - CREATE_PROVIDER + - CREATE_RECIPIENT + - CREATE_SCHEMA + - CREATE_SERVICE_CREDENTIAL + - CREATE_SHARE + - CREATE_STORAGE_CREDENTIAL + - CREATE_TABLE + - CREATE_VIEW + - CREATE_VOLUME + - EXECUTE + - MANAGE + - MANAGE_ALLOWLIST + - MODIFY + - READ_FILES + - READ_PRIVATE_FILES + - READ_VOLUME + - REFRESH + - SELECT + - SET_SHARE_PERMISSION + - USAGE + - USE_CATALOG + - USE_CONNECTION + - USE_MARKETPLACE_ASSETS + - USE_PROVIDER + - USE_RECIPIENT + - USE_SCHEMA + - USE_SHARE + - WRITE_FILES + - WRITE_PRIVATE_FILES + - WRITE_VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SharedDataObjectDataObjectType: + type: string + enum: + - FEATURE_SPEC + - FOREIGN_TABLE + - FUNCTION + - MATERIALIZED_VIEW + - MODEL + - NOTEBOOK_FILE + - SCHEMA + - STREAMING_TABLE + - TABLE + - VIEW + - VOLUME + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SharedDataObjectHistoryDataSharingStatus: + type: string + enum: + - DISABLED + - ENABLED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SharedDataObjectStatus: + type: string + enum: + - ACTIVE + - PERMISSION_DENIED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SharedDataObjectUpdateAction: + type: string + enum: + - ADD + - REMOVE + - UPDATE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SharedSecurableKind: + type: string + enum: + - FUNCTION_FEATURE_SPEC + - FUNCTION_REGISTERED_MODEL + - FUNCTION_STANDARD + description: The SecurableKind of a delta-shared object. + TableInternalAttributesSharedTableType: + type: string + enum: + - DELTA_ICEBERG_TABLE + - DIRECTORY_BASED_TABLE + - FILE_BASED_TABLE + - FOREIGN_ICEBERG_TABLE + - FOREIGN_TABLE + - MATERIALIZED_VIEW + - METRIC_VIEW + - STREAMING_TABLE + - VIEW + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + providers: + id: databricks_workspace.sharing.providers + name: providers + title: Providers + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1providers/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1providers/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.providers + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/providers/methods/get' + - $ref: '#/components/x-stackQL-resources/providers/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/providers/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/providers/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/providers/methods/delete' + replace: [] + provider_share_assets: + id: databricks_workspace.sharing.provider_share_assets + name: provider_share_assets + title: Provider Share Assets + methods: + list: + operation: + $ref: >- + #/paths/~1api~12.1~1data-sharing~1providers~1{provider_name}~1shares~1{share_name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/provider_share_assets/methods/list + insert: [] + update: [] + delete: [] + replace: [] + provider_shares: + id: databricks_workspace.sharing.provider_shares + name: provider_shares + title: Provider Shares + methods: + list_shares: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1providers~1{name}~1shares/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.shares + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/provider_shares/methods/list_shares + insert: [] + update: [] + delete: [] + replace: [] + recipient_activation: + id: databricks_workspace.sharing.recipient_activation + name: recipient_activation + title: Recipient Activation + methods: + get_url: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1public~1data_sharing_activation_info~1{activation_url}/get + response: + mediaType: application/json + openAPIDocKey: '200' + retrieve_token: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1public~1data_sharing_activation~1{activation_url}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: [] + update: [] + delete: [] + replace: [] + recipient_federation_policies: + id: databricks_workspace.sharing.recipient_federation_policies + name: recipient_federation_policies + title: Recipient Federation Policies + methods: + create: + operation: + $ref: >- + #/paths/~1api~12.0~1data-sharing~1recipients~1{recipient_name}~1federation-policies/post + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1data-sharing~1recipients~1{recipient_name}~1federation-policies/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.policies + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1data-sharing~1recipients~1{recipient_name}~1federation-policies~1{name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1data-sharing~1recipients~1{recipient_name}~1federation-policies~1{name}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/recipient_federation_policies/methods/get + - $ref: >- + #/components/x-stackQL-resources/recipient_federation_policies/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/recipient_federation_policies/methods/create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/recipient_federation_policies/methods/delete + replace: [] + recipients: + id: databricks_workspace.sharing.recipients + name: recipients + title: Recipients + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.recipients + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1recipients~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + rotate_token: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1recipients~1{name}~1rotate-token/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/recipients/methods/get' + - $ref: '#/components/x-stackQL-resources/recipients/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/recipients/methods/rotate_token' + - $ref: '#/components/x-stackQL-resources/recipients/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/recipients/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/recipients/methods/delete' + replace: [] + recipient_share_permissions: + id: databricks_workspace.sharing.recipient_share_permissions + name: recipient_share_permissions + title: Recipient Share Permissions + methods: + list: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1recipients~1{name}~1share-permissions/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/recipient_share_permissions/methods/list + insert: [] + update: [] + delete: [] + replace: [] + shares: + id: databricks_workspace.sharing.shares + name: shares + title: Shares + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1shares/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1shares/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.shares + delete: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/shares/methods/get' + - $ref: '#/components/x-stackQL-resources/shares/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/shares/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/shares/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/shares/methods/delete' + replace: [] + share_permissions: + id: databricks_workspace.sharing.share_permissions + name: share_permissions + title: Share Permissions + methods: + list: + operation: + $ref: '#/paths/~1api~12.1~1unity-catalog~1shares~1{name}~1permissions/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.1~1unity-catalog~1shares~1{name}~1permissions/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/share_permissions/methods/list' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/share_permissions/methods/update' + delete: [] + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/sql.yaml b/providers/src/databricks_workspace/v00.00.00000/services/sql.yaml new file mode 100644 index 00000000..e1fd7e08 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/sql.yaml @@ -0,0 +1,9728 @@ +openapi: 3.0.0 +info: + title: Databricks Sql API (workspace) + description: >- + OpenAPI specification for the Databricks sql service (workspace-level APIs), + generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.sql +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/sql/alerts: + post: + operationId: alerts_create + summary: Creates an alert. + tags: + - sql + - alerts + description: |- + Creates an alert. + + :param alert: :class:`CreateAlertRequestAlert` (optional) + :param auto_resolve_display_name: bool (optional) + If true, automatically resolve alert display name conflicts. Otherwise, fail the request if the + alert's display name conflicts with an existing alert's display name. + + :returns: :class:`Alert` + requestBody: + content: + application/json: + schema: + type: object + properties: + alert: + type: string + description: >- + :param auto_resolve_display_name: bool (optional) If true, + automatically resolve alert display name conflicts. + Otherwise, fail the request if the alert's display name + conflicts with an existing alert's display name. + auto_resolve_display_name: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Alert' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsAPI + get: + operationId: alerts_list + summary: >- + Gets a list of alerts accessible to the user, ordered by creation time. + **Warning:** Calling this API + tags: + - sql + - alerts + description: >- + Gets a list of alerts accessible to the user, ordered by creation time. + **Warning:** Calling this API + + concurrently 10 or more times could result in throttling, service + degradation, or a temporary ban. + + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`ListAlertsResponseAlert` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAlertsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsAPI + /api/2.0/sql/alerts/{id}: + delete: + operationId: alerts_delete + summary: >- + Moves an alert to the trash. Trashed alerts immediately disappear from + searches and list views, and + tags: + - sql + - alerts + description: >- + Moves an alert to the trash. Trashed alerts immediately disappear from + searches and list views, and + + can no longer trigger. You can restore a trashed alert through the UI. A + trashed alert is permanently + + deleted after 30 days. + + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsAPI + get: + operationId: alerts_get + summary: Gets an alert. + tags: + - sql + - alerts + description: |- + Gets an alert. + + :param id: str + + :returns: :class:`Alert` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`Alert`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Alert' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsAPI + patch: + operationId: alerts_update + summary: Updates an alert. + tags: + - sql + - alerts + description: |- + Updates an alert. + + :param id: str + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + :param alert: :class:`UpdateAlertRequestAlert` (optional) + :param auto_resolve_display_name: bool (optional) + If true, automatically resolve alert display name conflicts. Otherwise, fail the request if the + alert's display name conflicts with an existing alert's display name. + + :returns: :class:`Alert` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: >- + :param update_mask: str The field mask must be a single string, with + multiple fields separated by commas (no spaces). The field path is + relative to the resource object, using a dot (`.`) to navigate + sub-fields (e.g., `author.given_name`). Specification of elements in + sequence or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the resource + field names. A field mask of `*` indicates full replacement. It’s + recommended to always explicitly list the fields being updated and + avoid using `*` wildcards, as it can lead to unintended results if + the API changes in the future. + requestBody: + content: + application/json: + schema: + type: object + properties: + update_mask: + type: string + alert: + type: string + description: >- + :param auto_resolve_display_name: bool (optional) If true, + automatically resolve alert display name conflicts. + Otherwise, fail the request if the alert's display name + conflicts with an existing alert's display name. + auto_resolve_display_name: + type: string + required: + - update_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Alert' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsAPI + /api/2.0/preview/sql/alerts: + post: + operationId: alerts_legacy_create + summary: >- + Creates an alert. An alert is a Databricks SQL object that periodically + runs a query, evaluates a + tags: + - sql + - alerts_legacy + description: >- + Creates an alert. An alert is a Databricks SQL object that periodically + runs a query, evaluates a + + condition of its result, and notifies users or notification destinations + if the condition was met. + + + **Warning**: This API is deprecated. Please use :method:alerts/create + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param name: str + Name of the alert. + :param options: :class:`AlertOptions` + Alert configuration options. + :param query_id: str + Query ID. + :param parent: str (optional) + The identifier of the workspace folder containing the object. + :param rearm: int (optional) + Number of seconds after being triggered before the alert rearms itself and can be triggered again. + If `null`, alert will never be triggered again. + + :returns: :class:`LegacyAlert` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the alert. + options: + type: string + description: Alert configuration options. + query_id: + type: string + description: Query ID. + parent: + type: string + description: >- + The identifier of the workspace folder containing the + object. + rearm: + type: string + description: >- + Number of seconds after being triggered before the alert + rearms itself and can be triggered again. If `null`, alert + will never be triggered again. + required: + - name + - options + - query_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyAlert' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsLegacyAPI + get: + operationId: alerts_legacy_list + summary: Gets a list of alerts. + tags: + - sql + - alerts_legacy + description: >- + Gets a list of alerts. + + + **Warning**: This API is deprecated. Please use :method:alerts/list + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + + :returns: Iterator over :class:`LegacyAlert` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyAlert' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsLegacyAPI + /api/2.0/preview/sql/alerts/{alert_id}: + delete: + operationId: alerts_legacy_delete + summary: >- + Deletes an alert. Deleted alerts are no longer accessible and cannot be + restored. **Note**: Unlike + tags: + - sql + - alerts_legacy + description: >- + Deletes an alert. Deleted alerts are no longer accessible and cannot be + restored. **Note**: Unlike + + queries and dashboards, alerts cannot be moved to the trash. + + + **Warning**: This API is deprecated. Please use :method:alerts/delete + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param alert_id: str + parameters: + - name: alert_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsLegacyAPI + get: + operationId: alerts_legacy_get + summary: Gets an alert. + tags: + - sql + - alerts_legacy + description: >- + Gets an alert. + + + **Warning**: This API is deprecated. Please use :method:alerts/get + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param alert_id: str + + + :returns: :class:`LegacyAlert` + parameters: + - name: alert_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`LegacyAlert`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyAlert' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsLegacyAPI + put: + operationId: alerts_legacy_update + summary: Updates an alert. + tags: + - sql + - alerts_legacy + description: >- + Updates an alert. + + + **Warning**: This API is deprecated. Please use :method:alerts/update + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param alert_id: str + + :param name: str + Name of the alert. + :param options: :class:`AlertOptions` + Alert configuration options. + :param query_id: str + Query ID. + :param rearm: int (optional) + Number of seconds after being triggered before the alert rearms itself and can be triggered again. + If `null`, alert will never be triggered again. + parameters: + - name: alert_id + in: path + required: true + schema: + type: string + description: ':param name: str Name of the alert.' + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + options: + type: string + description: Alert configuration options. + query_id: + type: string + description: Query ID. + rearm: + type: string + description: >- + Number of seconds after being triggered before the alert + rearms itself and can be triggered again. If `null`, alert + will never be triggered again. + required: + - name + - options + - query_id + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsLegacyAPI + /api/2.0/alerts: + post: + operationId: alerts_v2_create_alert + summary: Create Alert + tags: + - sql + - alerts_v2 + description: |- + Create Alert + + :param alert: :class:`AlertV2` + + :returns: :class:`AlertV2` + requestBody: + content: + application/json: + schema: + type: object + properties: + alert: + type: string + description: ':returns: :class:`AlertV2`' + required: + - alert + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AlertV2' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsV2API + get: + operationId: alerts_v2_list_alerts + summary: Gets a list of alerts accessible to the user, ordered by creation time. + tags: + - sql + - alerts_v2 + description: |- + Gets a list of alerts accessible to the user, ordered by creation time. + + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`AlertV2` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAlertsV2Response' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsV2API + /api/2.0/alerts/{id}: + get: + operationId: alerts_v2_get_alert + summary: Gets an alert. + tags: + - sql + - alerts_v2 + description: |- + Gets an alert. + + :param id: str + + :returns: :class:`AlertV2` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`AlertV2`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AlertV2' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsV2API + delete: + operationId: alerts_v2_trash_alert + summary: >- + Moves an alert to the trash. Trashed alerts immediately disappear from + list views, and can no longer + tags: + - sql + - alerts_v2 + description: >- + Moves an alert to the trash. Trashed alerts immediately disappear from + list views, and can no longer + + trigger. You can restore a trashed alert through the UI. A trashed alert + is permanently deleted after + + 30 days. + + + :param id: str + + :param purge: bool (optional) + Whether to permanently delete the alert. If not set, the alert will only be soft deleted. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: >- + :param purge: bool (optional) Whether to permanently delete the + alert. If not set, the alert will only be soft deleted. + - name: purge + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsV2API + patch: + operationId: alerts_v2_update_alert + summary: Update alert + tags: + - sql + - alerts_v2 + description: |- + Update alert + + :param id: str + UUID identifying the alert. + :param alert: :class:`AlertV2` + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`AlertV2` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: UUID identifying the alert. + - name: update_mask + in: query + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + alert: + type: string + description: >- + :param update_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + required: + - alert + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AlertV2' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: AlertsV2API + /api/2.0/preview/sql/widgets: + post: + operationId: dashboard_widgets_create + summary: Adds a widget to a dashboard + tags: + - sql + - dashboard_widgets + description: |- + Adds a widget to a dashboard + + :param dashboard_id: str + Dashboard ID returned by :method:dashboards/create. + :param options: :class:`WidgetOptions` + :param width: int + Width of a widget + :param text: str (optional) + If this is a textbox widget, the application displays this text. This field is ignored if the widget + contains a visualization in the `visualization` field. + :param visualization_id: str (optional) + Query Vizualization ID returned by :method:queryvisualizations/create. + + :returns: :class:`Widget` + requestBody: + content: + application/json: + schema: + type: object + properties: + dashboard_id: + type: string + description: Dashboard ID returned by :method:dashboards/create. + options: + type: string + description: ':param width: int Width of a widget' + width: + type: integer + text: + type: string + description: >- + If this is a textbox widget, the application displays this + text. This field is ignored if the widget contains a + visualization in the `visualization` field. + visualization_id: + type: string + description: >- + Query Vizualization ID returned by + :method:queryvisualizations/create. + required: + - dashboard_id + - options + - width + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Widget' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardWidgetsAPI + /api/2.0/preview/sql/widgets/{id}: + delete: + operationId: dashboard_widgets_delete + summary: Removes a widget from a dashboard + tags: + - sql + - dashboard_widgets + description: |- + Removes a widget from a dashboard + + :param id: str + Widget ID returned by :method:dashboardwidgets/create + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Widget ID returned by :method:dashboardwidgets/create + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardWidgetsAPI + post: + operationId: dashboard_widgets_update + summary: Updates an existing widget + tags: + - sql + - dashboard_widgets + description: |- + Updates an existing widget + + :param id: str + Widget ID returned by :method:dashboardwidgets/create + :param dashboard_id: str + Dashboard ID returned by :method:dashboards/create. + :param options: :class:`WidgetOptions` + :param width: int + Width of a widget + :param text: str (optional) + If this is a textbox widget, the application displays this text. This field is ignored if the widget + contains a visualization in the `visualization` field. + :param visualization_id: str (optional) + Query Vizualization ID returned by :method:queryvisualizations/create. + + :returns: :class:`Widget` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Widget ID returned by :method:dashboardwidgets/create + requestBody: + content: + application/json: + schema: + type: object + properties: + dashboard_id: + type: string + description: Dashboard ID returned by :method:dashboards/create. + options: + type: string + description: ':param width: int Width of a widget' + width: + type: integer + text: + type: string + description: >- + If this is a textbox widget, the application displays this + text. This field is ignored if the widget contains a + visualization in the `visualization` field. + visualization_id: + type: string + description: >- + Query Vizualization ID returned by + :method:queryvisualizations/create. + required: + - dashboard_id + - options + - width + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Widget' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardWidgetsAPI + /api/2.0/preview/sql/dashboards/{dashboard_id}: + delete: + operationId: dashboards_delete + summary: >- + Moves a dashboard to the trash. Trashed dashboards do not appear in list + views or searches, and cannot + tags: + - sql + - dashboards + description: >- + Moves a dashboard to the trash. Trashed dashboards do not appear in list + views or searches, and cannot + + be shared. + + + :param dashboard_id: str + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardsAPI + get: + operationId: dashboards_get + summary: >- + Returns a JSON representation of a dashboard object, including its + visualization and query objects. + tags: + - sql + - dashboards + description: >- + Returns a JSON representation of a dashboard object, including its + visualization and query objects. + + + :param dashboard_id: str + + + :returns: :class:`Dashboard` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`Dashboard`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Dashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardsAPI + post: + operationId: dashboards_update + summary: >- + Modify this dashboard definition. This operation only affects attributes + of the dashboard object. It + tags: + - sql + - dashboards + description: >- + Modify this dashboard definition. This operation only affects attributes + of the dashboard object. It + + does not add, modify, or remove widgets. + + + **Note**: You cannot undo this operation. + + + :param dashboard_id: str + + :param name: str (optional) + The title of this dashboard that appears in list views and at the top of the dashboard page. + :param run_as_role: :class:`RunAsRole` (optional) + Sets the **Run as** role for the object. Must be set to one of `"viewer"` (signifying "run as + viewer" behavior) or `"owner"` (signifying "run as owner" behavior) + :param tags: List[str] (optional) + + + :returns: :class:`Dashboard` + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: >- + :param name: str (optional) The title of this dashboard that appears + in list views and at the top of the dashboard page. + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + run_as_role: + type: string + description: >- + Sets the **Run as** role for the object. Must be set to one + of `"viewer"` (signifying "run as viewer" behavior) or + `"owner"` (signifying "run as owner" behavior) + tags: + type: string + description: ':returns: :class:`Dashboard`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Dashboard' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardsAPI + /api/2.0/preview/sql/dashboards: + get: + operationId: dashboards_list + summary: Fetch a paginated list of dashboard objects. + tags: + - sql + - dashboards + description: >- + Fetch a paginated list of dashboard objects. + + + **Warning**: Calling this API concurrently 10 or more times could result + in throttling, service + + degradation, or a temporary ban. + + + :param order: :class:`ListOrder` (optional) + Name of dashboard attribute to order by. + :param page: int (optional) + Page number to retrieve. + :param page_size: int (optional) + Number of dashboards to return per page. + :param q: str (optional) + Full text search term. + + :returns: Iterator over :class:`Dashboard` + parameters: + - name: order + in: query + required: false + schema: + type: string + description: Name of dashboard attribute to order by. + - name: page + in: query + required: false + schema: + type: string + description: Page number to retrieve. + - name: page_size + in: query + required: false + schema: + type: string + description: Number of dashboards to return per page. + - name: q + in: query + required: false + schema: + type: string + description: Full text search term. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardsAPI + /api/2.0/preview/sql/dashboards/trash/{dashboard_id}: + post: + operationId: dashboards_restore + summary: >- + A restored dashboard appears in list views and searches and can be + shared. + tags: + - sql + - dashboards + description: >- + A restored dashboard appears in list views and searches and can be + shared. + + + :param dashboard_id: str + parameters: + - name: dashboard_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DashboardsAPI + /api/2.0/preview/sql/data_sources: + get: + operationId: data_sources_list + summary: >- + Retrieves a full list of SQL warehouses available in this workspace. All + fields that appear in this + tags: + - sql + - data_sources + description: >- + Retrieves a full list of SQL warehouses available in this workspace. All + fields that appear in this + + API response are enumerated for clarity. However, you need only a SQL + warehouse's `id` to create new + + queries against it. + + + **Warning**: This API is deprecated. Please use :method:warehouses/list + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + + :returns: Iterator over :class:`DataSource` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DataSource' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DataSourcesAPI + /api/2.0/preview/sql/permissions/{object_type.value}/{object_id}: + get: + operationId: dbsql_permissions_get + summary: >- + Gets a JSON representation of the access control list (ACL) for a + specified object. + tags: + - sql + - dbsql_permissions + description: >- + Gets a JSON representation of the access control list (ACL) for a + specified object. + + + **Warning**: This API is deprecated. Please use + :method:workspace/getpermissions instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param object_type: :class:`ObjectTypePlural` + The type of object permissions to check. + :param object_id: str + Object ID. An ACL is returned for the object with this UUID. + + :returns: :class:`GetResponse` + parameters: + - name: object_type.value + in: path + required: true + schema: + type: string + - name: object_id + in: path + required: true + schema: + type: string + description: Object ID. An ACL is returned for the object with this UUID. + - name: object_type + in: query + required: true + schema: + type: string + description: The type of object permissions to check. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbsqlPermissionsAPI + post: + operationId: dbsql_permissions_set + summary: >- + Sets the access control list (ACL) for a specified object. This + operation will complete rewrite the + tags: + - sql + - dbsql_permissions + description: >- + Sets the access control list (ACL) for a specified object. This + operation will complete rewrite the + + ACL. + + + **Warning**: This API is deprecated. Please use + :method:workspace/setpermissions instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param object_type: :class:`ObjectTypePlural` + The type of object permission to set. + :param object_id: str + Object ID. The ACL for the object with this UUID is overwritten by this request's POST content. + :param access_control_list: List[:class:`AccessControl`] (optional) + + + :returns: :class:`SetResponse` + parameters: + - name: object_type.value + in: path + required: true + schema: + type: string + - name: object_id + in: path + required: true + schema: + type: string + description: >- + Object ID. The ACL for the object with this UUID is overwritten by + this request's POST content. + requestBody: + content: + application/json: + schema: + type: object + properties: + object_type: + type: string + description: The type of object permission to set. + access_control_list: + type: string + description: ':returns: :class:`SetResponse`' + required: + - object_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/SetResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbsqlPermissionsAPI + /api/2.0/preview/sql/permissions/{object_type.value}/{object_id}/transfer: + post: + operationId: dbsql_permissions_transfer_ownership + summary: >- + Transfers ownership of a dashboard, query, or alert to an active user. + Requires an admin API key. + tags: + - sql + - dbsql_permissions + description: >- + Transfers ownership of a dashboard, query, or alert to an active user. + Requires an admin API key. + + + **Warning**: This API is deprecated. For queries and alerts, please use + :method:queries/update and + + :method:alerts/update respectively instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param object_type: :class:`OwnableObjectType` + The type of object on which to change ownership. + :param object_id: :class:`TransferOwnershipObjectId` + The ID of the object on which to change ownership. + :param new_owner: str (optional) + Email address for the new owner, who must exist in the workspace. + + :returns: :class:`Success` + parameters: + - name: object_type.value + in: path + required: true + schema: + type: string + - name: object_id + in: path + required: true + schema: + type: string + description: The ID of the object on which to change ownership. + requestBody: + content: + application/json: + schema: + type: object + properties: + object_type: + type: string + description: The type of object on which to change ownership. + new_owner: + type: string + description: >- + Email address for the new owner, who must exist in the + workspace. + required: + - object_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Success' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: DbsqlPermissionsAPI + /api/2.0/sql/queries: + post: + operationId: queries_create + summary: Creates a query. + tags: + - sql + - queries + description: |- + Creates a query. + + :param auto_resolve_display_name: bool (optional) + If true, automatically resolve query display name conflicts. Otherwise, fail the request if the + query's display name conflicts with an existing query's display name. + :param query: :class:`CreateQueryRequestQuery` (optional) + + :returns: :class:`Query` + requestBody: + content: + application/json: + schema: + type: object + properties: + auto_resolve_display_name: + type: string + description: >- + If true, automatically resolve query display name conflicts. + Otherwise, fail the request if the query's display name + conflicts with an existing query's display name. + query: + type: string + description: ':returns: :class:`Query`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Query' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesAPI + get: + operationId: queries_list + summary: >- + Gets a list of queries accessible to the user, ordered by creation time. + **Warning:** Calling this API + tags: + - sql + - queries + description: >- + Gets a list of queries accessible to the user, ordered by creation time. + **Warning:** Calling this API + + concurrently 10 or more times could result in throttling, service + degradation, or a temporary ban. + + + :param page_size: int (optional) + + :param page_token: str (optional) + + + :returns: Iterator over :class:`ListQueryObjectsResponseQuery` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: ':param page_token: str (optional)' + - name: page_token + in: query + required: false + schema: + type: string + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListQueryObjectsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesAPI + /api/2.0/sql/queries/{id}: + delete: + operationId: queries_delete + summary: >- + Moves a query to the trash. Trashed queries immediately disappear from + searches and list views, and + tags: + - sql + - queries + description: >- + Moves a query to the trash. Trashed queries immediately disappear from + searches and list views, and + + cannot be used for alerts. You can restore a trashed query through the + UI. A trashed query is + + permanently deleted after 30 days. + + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesAPI + get: + operationId: queries_get + summary: Gets a query. + tags: + - sql + - queries + description: |- + Gets a query. + + :param id: str + + :returns: :class:`Query` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':returns: :class:`Query`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Query' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesAPI + patch: + operationId: queries_update + summary: Updates a query. + tags: + - sql + - queries + description: |- + Updates a query. + + :param id: str + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + :param auto_resolve_display_name: bool (optional) + If true, automatically resolve alert display name conflicts. Otherwise, fail the request if the + alert's display name conflicts with an existing alert's display name. + :param query: :class:`UpdateQueryRequestQuery` (optional) + + :returns: :class:`Query` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: >- + :param update_mask: str The field mask must be a single string, with + multiple fields separated by commas (no spaces). The field path is + relative to the resource object, using a dot (`.`) to navigate + sub-fields (e.g., `author.given_name`). Specification of elements in + sequence or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the resource + field names. A field mask of `*` indicates full replacement. It’s + recommended to always explicitly list the fields being updated and + avoid using `*` wildcards, as it can lead to unintended results if + the API changes in the future. + requestBody: + content: + application/json: + schema: + type: object + properties: + update_mask: + type: string + auto_resolve_display_name: + type: string + description: >- + If true, automatically resolve alert display name conflicts. + Otherwise, fail the request if the alert's display name + conflicts with an existing alert's display name. + query: + type: string + description: ':returns: :class:`Query`' + required: + - update_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Query' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesAPI + /api/2.0/sql/queries/{id}/visualizations: + get: + operationId: queries_list_visualizations + summary: Gets a list of visualizations on a query. + tags: + - sql + - queries + description: |- + Gets a list of visualizations on a query. + + :param id: str + :param page_size: int (optional) + :param page_token: str (optional) + + :returns: Iterator over :class:`Visualization` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: ':param page_size: int (optional)' + - name: page_size + in: query + required: false + schema: + type: string + - name: page_token + in: query + required: false + schema: + type: string + description: ':returns: Iterator over :class:`Visualization`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListVisualizationsForQueryResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesAPI + /api/2.0/preview/sql/queries: + post: + operationId: queries_legacy_create + summary: >- + Creates a new query definition. Queries created with this endpoint + belong to the authenticated user + tags: + - sql + - queries_legacy + description: >- + Creates a new query definition. Queries created with this endpoint + belong to the authenticated user + + making the request. + + + The `data_source_id` field specifies the ID of the SQL warehouse to run + this query against. You can + + use the Data Sources API to see a complete list of available SQL + warehouses. Or you can copy the + + `data_source_id` from an existing query. + + + **Note**: You cannot add a visualization until you create the query. + + + **Warning**: This API is deprecated. Please use :method:queries/create + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param data_source_id: str (optional) + Data source ID maps to the ID of the data source used by the resource and is distinct from the + warehouse ID. [Learn more] + + [Learn more]: https://docs.databricks.com/api/workspace/datasources/list + :param description: str (optional) + General description that conveys additional information about this query such as usage notes. + :param name: str (optional) + The title of this query that appears in list views, widget headings, and on the query page. + :param options: Any (optional) + Exclusively used for storing a list parameter definitions. A parameter is an object with `title`, + `name`, `type`, and `value` properties. The `value` field here is the default value. It can be + overridden at runtime. + :param parent: str (optional) + The identifier of the workspace folder containing the object. + :param query: str (optional) + The text of the query to be run. + :param run_as_role: :class:`RunAsRole` (optional) + Sets the **Run as** role for the object. Must be set to one of `"viewer"` (signifying "run as + viewer" behavior) or `"owner"` (signifying "run as owner" behavior) + :param tags: List[str] (optional) + + + :returns: :class:`LegacyQuery` + requestBody: + content: + application/json: + schema: + type: object + properties: + data_source_id: + type: string + description: >- + Data source ID maps to the ID of the data source used by the + resource and is distinct from the warehouse ID. [Learn more] + [Learn more]: + https://docs.databricks.com/api/workspace/datasources/list + description: + type: string + description: >- + General description that conveys additional information + about this query such as usage notes. + name: + type: string + description: >- + The title of this query that appears in list views, widget + headings, and on the query page. + options: + type: string + description: >- + Exclusively used for storing a list parameter definitions. A + parameter is an object with `title`, `name`, `type`, and + `value` properties. The `value` field here is the default + value. It can be overridden at runtime. + parent: + type: string + description: >- + The identifier of the workspace folder containing the + object. + query: + type: string + description: The text of the query to be run. + run_as_role: + type: string + description: >- + Sets the **Run as** role for the object. Must be set to one + of `"viewer"` (signifying "run as viewer" behavior) or + `"owner"` (signifying "run as owner" behavior) + tags: + type: string + description: ':returns: :class:`LegacyQuery`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyQuery' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesLegacyAPI + get: + operationId: queries_legacy_list + summary: >- + Gets a list of queries. Optionally, this list can be filtered by a + search term. + tags: + - sql + - queries_legacy + description: >- + Gets a list of queries. Optionally, this list can be filtered by a + search term. + + + **Warning**: Calling this API concurrently 10 or more times could result + in throttling, service + + degradation, or a temporary ban. + + + **Warning**: This API is deprecated. Please use :method:queries/list + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param order: str (optional) + Name of query attribute to order by. Default sort order is ascending. Append a dash (`-`) to order + descending instead. + + - `name`: The name of the query. + + - `created_at`: The timestamp the query was created. + + - `runtime`: The time it took to run this query. This is blank for parameterized queries. A blank + value is treated as the highest value for sorting. + + - `executed_at`: The timestamp when the query was last run. + + - `created_by`: The user name of the user that created the query. + :param page: int (optional) + Page number to retrieve. + :param page_size: int (optional) + Number of queries to return per page. + :param q: str (optional) + Full text search term + + :returns: Iterator over :class:`LegacyQuery` + parameters: + - name: order + in: query + required: false + schema: + type: string + description: >- + Name of query attribute to order by. Default sort order is + ascending. Append a dash (`-`) to order descending instead. - + `name`: The name of the query. - `created_at`: The timestamp the + query was created. - `runtime`: The time it took to run this query. + This is blank for parameterized queries. A blank value is treated as + the highest value for sorting. - `executed_at`: The timestamp when + the query was last run. - `created_by`: The user name of the user + that created the query. + - name: page + in: query + required: false + schema: + type: string + description: Page number to retrieve. + - name: page_size + in: query + required: false + schema: + type: string + description: Number of queries to return per page. + - name: q + in: query + required: false + schema: + type: string + description: Full text search term + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/QueryList' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesLegacyAPI + /api/2.0/preview/sql/queries/{query_id}: + delete: + operationId: queries_legacy_delete + summary: >- + Moves a query to the trash. Trashed queries immediately disappear from + searches and list views, and + tags: + - sql + - queries_legacy + description: >- + Moves a query to the trash. Trashed queries immediately disappear from + searches and list views, and + + they cannot be used for alerts. The trash is deleted after 30 days. + + + **Warning**: This API is deprecated. Please use :method:queries/delete + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param query_id: str + parameters: + - name: query_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesLegacyAPI + get: + operationId: queries_legacy_get + summary: >- + Retrieve a query object definition along with contextual permissions + information about the currently + tags: + - sql + - queries_legacy + description: >- + Retrieve a query object definition along with contextual permissions + information about the currently + + authenticated user. + + + **Warning**: This API is deprecated. Please use :method:queries/get + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param query_id: str + + + :returns: :class:`LegacyQuery` + parameters: + - name: query_id + in: path + required: true + schema: + type: string + description: ':returns: :class:`LegacyQuery`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyQuery' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesLegacyAPI + post: + operationId: queries_legacy_update + summary: Modify this query definition. + tags: + - sql + - queries_legacy + description: >- + Modify this query definition. + + + **Note**: You cannot undo this operation. + + + **Warning**: This API is deprecated. Please use :method:queries/update + instead. [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param query_id: str + + :param data_source_id: str (optional) + Data source ID maps to the ID of the data source used by the resource and is distinct from the + warehouse ID. [Learn more] + + [Learn more]: https://docs.databricks.com/api/workspace/datasources/list + :param description: str (optional) + General description that conveys additional information about this query such as usage notes. + :param name: str (optional) + The title of this query that appears in list views, widget headings, and on the query page. + :param options: Any (optional) + Exclusively used for storing a list parameter definitions. A parameter is an object with `title`, + `name`, `type`, and `value` properties. The `value` field here is the default value. It can be + overridden at runtime. + :param query: str (optional) + The text of the query to be run. + :param run_as_role: :class:`RunAsRole` (optional) + Sets the **Run as** role for the object. Must be set to one of `"viewer"` (signifying "run as + viewer" behavior) or `"owner"` (signifying "run as owner" behavior) + :param tags: List[str] (optional) + + + :returns: :class:`LegacyQuery` + parameters: + - name: query_id + in: path + required: true + schema: + type: string + description: >- + :param data_source_id: str (optional) Data source ID maps to the ID + of the data source used by the resource and is distinct from the + warehouse ID. [Learn more] [Learn more]: + https://docs.databricks.com/api/workspace/datasources/list + requestBody: + content: + application/json: + schema: + type: object + properties: + data_source_id: + type: string + description: + type: string + description: >- + General description that conveys additional information + about this query such as usage notes. + name: + type: string + description: >- + The title of this query that appears in list views, widget + headings, and on the query page. + options: + type: string + description: >- + Exclusively used for storing a list parameter definitions. A + parameter is an object with `title`, `name`, `type`, and + `value` properties. The `value` field here is the default + value. It can be overridden at runtime. + query: + type: string + description: The text of the query to be run. + run_as_role: + type: string + description: >- + Sets the **Run as** role for the object. Must be set to one + of `"viewer"` (signifying "run as viewer" behavior) or + `"owner"` (signifying "run as owner" behavior) + tags: + type: string + description: ':returns: :class:`LegacyQuery`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyQuery' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesLegacyAPI + /api/2.0/preview/sql/queries/trash/{query_id}: + post: + operationId: queries_legacy_restore + summary: >- + Restore a query that has been moved to the trash. A restored query + appears in list views and searches. + tags: + - sql + - queries_legacy + description: >- + Restore a query that has been moved to the trash. A restored query + appears in list views and searches. + + You can use restored queries for alerts. + + + **Warning**: This API is deprecated. Please see the latest version. + [Learn more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param query_id: str + parameters: + - name: query_id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueriesLegacyAPI + /api/2.0/sql/history/queries: + get: + operationId: query_history_list + summary: >- + List the history of queries through SQL warehouses, and serverless + compute. + tags: + - sql + - query_history + description: >- + List the history of queries through SQL warehouses, and serverless + compute. + + + You can filter by user ID, warehouse ID, status, and time range. Most + recently started queries are + + returned first (up to max_results in request). The pagination token + returned in response can be used + + to list subsequent query statuses. + + + :param filter_by: :class:`QueryFilter` (optional) + An optional filter object to limit query history results. Accepts parameters such as user IDs, + endpoint IDs, and statuses to narrow the returned data. In a URL, the parameters of this filter are + specified with dot notation. For example: `filter_by.statement_ids`. + :param include_metrics: bool (optional) + Whether to include the query metrics with each query. Only use this for a small subset of queries + (max_results). Defaults to false. + :param max_results: int (optional) + Limit the number of results returned in one page. Must be less than 1000 and the default is 100. + :param page_token: str (optional) + A token that can be used to get the next page of results. The token can contains characters that + need to be encoded before using it in a URL. For example, the character '+' needs to be replaced by + %2B. This field is optional. + + :returns: :class:`ListQueriesResponse` + parameters: + - name: filter_by + in: query + required: false + schema: + type: string + description: >- + An optional filter object to limit query history results. Accepts + parameters such as user IDs, endpoint IDs, and statuses to narrow + the returned data. In a URL, the parameters of this filter are + specified with dot notation. For example: `filter_by.statement_ids`. + - name: include_metrics + in: query + required: false + schema: + type: string + description: >- + Whether to include the query metrics with each query. Only use this + for a small subset of queries (max_results). Defaults to false. + - name: max_results + in: query + required: false + schema: + type: string + description: >- + Limit the number of results returned in one page. Must be less than + 1000 and the default is 100. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A token that can be used to get the next page of results. The token + can contains characters that need to be encoded before using it in a + URL. For example, the character '+' needs to be replaced by %2B. + This field is optional. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListQueriesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueryHistoryAPI + /api/2.0/sql/visualizations: + post: + operationId: query_visualizations_create + summary: Adds a visualization to a query. + tags: + - sql + - query_visualizations + description: >- + Adds a visualization to a query. + + + :param visualization: :class:`CreateVisualizationRequestVisualization` + (optional) + + + :returns: :class:`Visualization` + requestBody: + content: + application/json: + schema: + type: object + properties: + visualization: + type: string + description: ':returns: :class:`Visualization`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Visualization' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueryVisualizationsAPI + /api/2.0/sql/visualizations/{id}: + delete: + operationId: query_visualizations_delete + summary: Removes a visualization. + tags: + - sql + - query_visualizations + description: |- + Removes a visualization. + + :param id: str + parameters: + - name: id + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueryVisualizationsAPI + patch: + operationId: query_visualizations_update + summary: Updates a visualization. + tags: + - sql + - query_visualizations + description: >- + Updates a visualization. + + + :param id: str + + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + :param visualization: :class:`UpdateVisualizationRequestVisualization` + (optional) + + + :returns: :class:`Visualization` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: >- + :param update_mask: str The field mask must be a single string, with + multiple fields separated by commas (no spaces). The field path is + relative to the resource object, using a dot (`.`) to navigate + sub-fields (e.g., `author.given_name`). Specification of elements in + sequence or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the resource + field names. A field mask of `*` indicates full replacement. It’s + recommended to always explicitly list the fields being updated and + avoid using `*` wildcards, as it can lead to unintended results if + the API changes in the future. + requestBody: + content: + application/json: + schema: + type: object + properties: + update_mask: + type: string + visualization: + type: string + description: ':returns: :class:`Visualization`' + required: + - update_mask + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/Visualization' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueryVisualizationsAPI + /api/2.0/preview/sql/visualizations: + post: + operationId: query_visualizations_legacy_create + summary: Creates visualization in the query. + tags: + - sql + - query_visualizations_legacy + description: >- + Creates visualization in the query. + + + **Warning**: This API is deprecated. Please use + :method:queryvisualizations/create instead. [Learn + + more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param options: Any + The options object varies widely from one visualization type to the next and is unsupported. + Databricks does not recommend modifying visualization settings in JSON. + :param query_id: str + The identifier returned by :method:queries/create + :param type: str + The type of visualization: chart, table, pivot table, and so on. + :param description: str (optional) + A short description of this visualization. This is not displayed in the UI. + :param name: str (optional) + The name of the visualization that appears on dashboards and the query screen. + + :returns: :class:`LegacyVisualization` + requestBody: + content: + application/json: + schema: + type: object + properties: + options: + type: string + description: >- + The options object varies widely from one visualization type + to the next and is unsupported. Databricks does not + recommend modifying visualization settings in JSON. + query_id: + type: string + description: The identifier returned by :method:queries/create + type: + type: string + description: >- + The type of visualization: chart, table, pivot table, and so + on. + description: + type: string + description: >- + A short description of this visualization. This is not + displayed in the UI. + name: + type: string + description: >- + The name of the visualization that appears on dashboards and + the query screen. + required: + - options + - query_id + - type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyVisualization' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueryVisualizationsLegacyAPI + /api/2.0/preview/sql/visualizations/{id}: + delete: + operationId: query_visualizations_legacy_delete + summary: Removes a visualization from the query. + tags: + - sql + - query_visualizations_legacy + description: >- + Removes a visualization from the query. + + + **Warning**: This API is deprecated. Please use + :method:queryvisualizations/delete instead. [Learn + + more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param id: str + Widget ID returned by :method:queryvisualizations/create + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Widget ID returned by :method:queryvisualizations/create + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueryVisualizationsLegacyAPI + post: + operationId: query_visualizations_legacy_update + summary: Updates visualization in the query. + tags: + - sql + - query_visualizations_legacy + description: >- + Updates visualization in the query. + + + **Warning**: This API is deprecated. Please use + :method:queryvisualizations/update instead. [Learn + + more] + + + [Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html + + + :param created_at: str (optional) + + :param description: str (optional) + A short description of this visualization. This is not displayed in the UI. + :param id: str (optional) + The UUID for this visualization. + :param name: str (optional) + The name of the visualization that appears on dashboards and the query screen. + :param options: Any (optional) + The options object varies widely from one visualization type to the next and is unsupported. + Databricks does not recommend modifying visualization settings in JSON. + :param query: :class:`LegacyQuery` (optional) + + :param type: str (optional) + The type of visualization: chart, table, pivot table, and so on. + :param updated_at: str (optional) + + + :returns: :class:`LegacyVisualization` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: The UUID for this visualization. + requestBody: + content: + application/json: + schema: + type: object + properties: + created_at: + type: string + description: >- + :param description: str (optional) A short description of + this visualization. This is not displayed in the UI. + description: + type: string + name: + type: string + description: >- + The name of the visualization that appears on dashboards and + the query screen. + options: + type: string + description: >- + The options object varies widely from one visualization type + to the next and is unsupported. Databricks does not + recommend modifying visualization settings in JSON. + query: + type: string + description: >- + :param type: str (optional) The type of visualization: + chart, table, pivot table, and so on. + type: + type: string + updated_at: + type: string + description: ':returns: :class:`LegacyVisualization`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/LegacyVisualization' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: QueryVisualizationsLegacyAPI + /api/2.0/redash-v2/config: + get: + operationId: redash_config_get_config + summary: Read workspace configuration for Redash-v2. + tags: + - sql + - redash_config + description: |- + Read workspace configuration for Redash-v2. + + + :returns: :class:`ClientConfig` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ClientConfig' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: RedashConfigAPI + /api/2.0/sql/statements/{statement_id}/cancel: + post: + operationId: statement_execution_cancel_execution + summary: >- + Requests that an executing statement be canceled. Callers must poll for + status to see the terminal + tags: + - sql + - statement_execution + description: >- + Requests that an executing statement be canceled. Callers must poll for + status to see the terminal + + state. Cancel response is empty; receiving response indicates successful + receipt. + + + :param statement_id: str + The statement ID is returned upon successfully submitting a SQL statement, and is a required + reference for all subsequent calls. + parameters: + - name: statement_id + in: path + required: true + schema: + type: string + description: >- + The statement ID is returned upon successfully submitting a SQL + statement, and is a required reference for all subsequent calls. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StatementExecutionAPI + /api/2.0/sql/statements: + post: + operationId: statement_execution_execute_statement + summary: >- + Execute a SQL statement and optionally await its results for a specified + time. + tags: + - sql + - statement_execution + description: >- + Execute a SQL statement and optionally await its results for a specified + time. + + + **Use case: small result sets with INLINE + JSON_ARRAY** + + + For flows that generate small and predictable result sets (<= 25 MiB), + `INLINE` responses of + + `JSON_ARRAY` result data are typically the simplest way to execute and + fetch result data. + + + **Use case: large result sets with EXTERNAL_LINKS** + + + Using `EXTERNAL_LINKS` to fetch result data allows you to fetch large + result sets efficiently. The + + main differences from using `INLINE` disposition are that the result + data is accessed with URLs, and + + that there are 3 supported formats: `JSON_ARRAY`, `ARROW_STREAM` and + `CSV` compared to only + + `JSON_ARRAY` with `INLINE`. + + + ** URLs** + + + External links point to data stored within your workspace's internal + storage, in the form of a URL. + + The URLs are valid for only a short period, <= 15 minutes. Alongside + each `external_link` is an + + expiration field indicating the time at which the URL is no longer + valid. In `EXTERNAL_LINKS` mode, + + chunks can be resolved and fetched multiple times and in parallel. + + + ---- + + + ### **Warning: Databricks strongly recommends that you protect the URLs + that are returned by the + + `EXTERNAL_LINKS` disposition.** + + + When you use the `EXTERNAL_LINKS` disposition, a short-lived, URL is + generated, which can be used to + + download the results directly from . As a short-lived is embedded in + this URL, you should protect the + + URL. + + + Because URLs are already generated with embedded temporary s, you must + not set an `Authorization` + + header in the download requests. + + + The `EXTERNAL_LINKS` disposition can be disabled upon request by + creating a support case. + + + See also [Security best practices]. + + + ---- + + + StatementResponse contains `statement_id` and `status`; other fields + might be absent or present + + depending on context. If the SQL warehouse fails to execute the provided + statement, a 200 response is + + returned with `status.state` set to `FAILED` (in contrast to a failure + when accepting the request, + + which results in a non-200 response). Details of the error can be found + at `status.error` in case of + + execution failures. + + + [Security best practices]: + https://docs.databricks.com/sql/admin/sql-execution-tutorial.html#security-best-practices + + + :param statement: str + The SQL statement to execute. The statement can optionally be parameterized, see `parameters`. The + maximum query text size is 16 MiB. + :param warehouse_id: str + Warehouse upon which to execute a statement. See also [What are SQL warehouses?] + + [What are SQL warehouses?]: https://docs.databricks.com/sql/admin/warehouse-type.html + :param byte_limit: int (optional) + Applies the given byte limit to the statement's result size. Byte counts are based on internal data + representations and might not match the final size in the requested `format`. If the result was + truncated due to the byte limit, then `truncated` in the response is set to `true`. When using + `EXTERNAL_LINKS` disposition, a default `byte_limit` of 100 GiB is applied if `byte_limit` is not + explicitly set. + :param catalog: str (optional) + Sets default catalog for statement execution, similar to [`USE CATALOG`] in SQL. + + [`USE CATALOG`]: https://docs.databricks.com/sql/language-manual/sql-ref-syntax-ddl-use-catalog.html + :param disposition: :class:`Disposition` (optional) + The fetch disposition provides two modes of fetching results: `INLINE` and `EXTERNAL_LINKS`. + + Statements executed with `INLINE` disposition will return result data inline, in `JSON_ARRAY` + format, in a series of chunks. If a given statement produces a result set with a size larger than 25 + MiB, that statement execution is aborted, and no result set will be available. + + **NOTE** Byte limits are computed based upon internal representations of the result set data, and + might not match the sizes visible in JSON responses. + + Statements executed with `EXTERNAL_LINKS` disposition will return result data as external links: + URLs that point to cloud storage internal to the workspace. Using `EXTERNAL_LINKS` disposition + allows statements to generate arbitrarily sized result sets for fetching up to 100 GiB. The + resulting links have two important properties: + + 1. They point to resources _external_ to the Databricks compute; therefore any associated + authentication information (typically a personal access token, OAuth token, or similar) _must be + removed_ when fetching from these links. + + 2. These are URLs with a specific expiration, indicated in the response. The behavior when + attempting to use an expired link is cloud specific. + :param format: :class:`Format` (optional) + Statement execution supports three result formats: `JSON_ARRAY` (default), `ARROW_STREAM`, and + `CSV`. + + Important: The formats `ARROW_STREAM` and `CSV` are supported only with `EXTERNAL_LINKS` + disposition. `JSON_ARRAY` is supported in `INLINE` and `EXTERNAL_LINKS` disposition. + + When specifying `format=JSON_ARRAY`, result data will be formatted as an array of arrays of values, + where each value is either the *string representation* of a value, or `null`. For example, the + output of `SELECT concat('id-', id) AS strCol, id AS intCol, null AS nullCol FROM range(3)` would + look like this: + + ``` [ [ "id-1", "1", null ], [ "id-2", "2", null ], [ "id-3", "3", null ], ] ``` + + When specifying `format=JSON_ARRAY` and `disposition=EXTERNAL_LINKS`, each chunk in the result + contains compact JSON with no indentation or extra whitespace. + + When specifying `format=ARROW_STREAM` and `disposition=EXTERNAL_LINKS`, each chunk in the result + will be formatted as Apache Arrow Stream. See the [Apache Arrow streaming format]. + + When specifying `format=CSV` and `disposition=EXTERNAL_LINKS`, each chunk in the result will be a + CSV according to [RFC 4180] standard. All the columns values will have *string representation* + similar to the `JSON_ARRAY` format, and `null` values will be encoded as “null”. Only the first + chunk in the result would contain a header row with column names. For example, the output of `SELECT + concat('id-', id) AS strCol, id AS intCol, null as nullCol FROM range(3)` would look like this: + + ``` strCol,intCol,nullCol id-1,1,null id-2,2,null id-3,3,null ``` + + [Apache Arrow streaming format]: https://arrow.apache.org/docs/format/Columnar.html#ipc-streaming-format + [RFC 4180]: https://www.rfc-editor.org/rfc/rfc4180 + :param on_wait_timeout: :class:`ExecuteStatementRequestOnWaitTimeout` + (optional) + When `wait_timeout > 0s`, the call will block up to the specified time. If the statement execution + doesn't finish within this time, `on_wait_timeout` determines whether the execution should continue + or be canceled. When set to `CONTINUE`, the statement execution continues asynchronously and the + call returns a statement ID which can be used for polling with + :method:statementexecution/getStatement. When set to `CANCEL`, the statement execution is canceled + and the call returns with a `CANCELED` state. + :param parameters: List[:class:`StatementParameterListItem`] (optional) + A list of parameters to pass into a SQL statement containing parameter markers. A parameter consists + of a name, a value, and optionally a type. To represent a NULL value, the `value` field may be + omitted or set to `null` explicitly. If the `type` field is omitted, the value is interpreted as a + string. + + If the type is given, parameters will be checked for type correctness according to the given type. A + value is correct if the provided string can be converted to the requested type using the `cast` + function. The exact semantics are described in the section [`cast` function] of the SQL language + reference. + + For example, the following statement contains two parameters, `my_name` and `my_date`: + + ``` SELECT * FROM my_table WHERE name = :my_name AND date = :my_date ``` + + The parameters can be passed in the request body as follows: + + ` { ..., "statement": "SELECT * FROM my_table WHERE name = :my_name AND date = :my_date", + "parameters": [ { "name": "my_name", "value": "the name" }, { "name": "my_date", "value": + "2020-01-01", "type": "DATE" } ] } ` + + Currently, positional parameters denoted by a `?` marker are not supported by the Databricks SQL + Statement Execution API. + + Also see the section [Parameter markers] of the SQL language reference. + + [Parameter markers]: https://docs.databricks.com/sql/language-manual/sql-ref-parameter-marker.html + [`cast` function]: https://docs.databricks.com/sql/language-manual/functions/cast.html + :param query_tags: List[:class:`QueryTag`] (optional) + An array of query tags to annotate a SQL statement. A query tag consists of a non-empty key and, + optionally, a value. To represent a NULL value, either omit the `value` field or manually set it to + `null` or white space. Refer to the SQL language reference for the format specification of query + tags. There's no significance to the order of tags. Only one value per key will be recorded. A + sequence in excess of 20 query tags will be coerced to 20. Example: + + { ..., "query_tags": [ { "key": "team", "value": "eng" }, { "key": "some key only tag" } ] } + :param row_limit: int (optional) + Applies the given row limit to the statement's result set, but unlike the `LIMIT` clause in SQL, it + also sets the `truncated` field in the response to indicate whether the result was trimmed due to + the limit or not. + :param schema: str (optional) + Sets default schema for statement execution, similar to [`USE SCHEMA`] in SQL. + + [`USE SCHEMA`]: https://docs.databricks.com/sql/language-manual/sql-ref-syntax-ddl-use-schema.html + :param wait_timeout: str (optional) + The time in seconds the call will wait for the statement's result set as `Ns`, where `N` can be set + to 0 or to a value between 5 and 50. + + When set to `0s`, the statement will execute in asynchronous mode and the call will not wait for the + execution to finish. In this case, the call returns directly with `PENDING` state and a statement ID + which can be used for polling with :method:statementexecution/getStatement. + + When set between 5 and 50 seconds, the call will behave synchronously up to this timeout and wait + for the statement execution to finish. If the execution finishes within this time, the call returns + immediately with a manifest and result data (or a `FAILED` state in case of an execution error). If + the statement takes longer to execute, `on_wait_timeout` determines what should happen after the + timeout is reached. + + :returns: :class:`StatementResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + statement: + type: string + description: >- + The SQL statement to execute. The statement can optionally + be parameterized, see `parameters`. The maximum query text + size is 16 MiB. + warehouse_id: + type: string + description: >- + Warehouse upon which to execute a statement. See also [What + are SQL warehouses?] [What are SQL warehouses?]: + https://docs.databricks.com/sql/admin/warehouse-type.html + byte_limit: + type: string + description: >- + Applies the given byte limit to the statement's result size. + Byte counts are based on internal data representations and + might not match the final size in the requested `format`. If + the result was truncated due to the byte limit, then + `truncated` in the response is set to `true`. When using + `EXTERNAL_LINKS` disposition, a default `byte_limit` of 100 + GiB is applied if `byte_limit` is not explicitly set. + catalog: + type: string + description: >- + Sets default catalog for statement execution, similar to + [`USE CATALOG`] in SQL. [`USE CATALOG`]: + https://docs.databricks.com/sql/language-manual/sql-ref-syntax-ddl-use-catalog.html + disposition: + type: string + description: >- + The fetch disposition provides two modes of fetching + results: `INLINE` and `EXTERNAL_LINKS`. Statements executed + with `INLINE` disposition will return result data inline, in + `JSON_ARRAY` format, in a series of chunks. If a given + statement produces a result set with a size larger than 25 + MiB, that statement execution is aborted, and no result set + will be available. **NOTE** Byte limits are computed based + upon internal representations of the result set data, and + might not match the sizes visible in JSON responses. + Statements executed with `EXTERNAL_LINKS` disposition will + return result data as external links: URLs that point to + cloud storage internal to the workspace. Using + `EXTERNAL_LINKS` disposition allows statements to generate + arbitrarily sized result sets for fetching up to 100 GiB. + The resulting links have two important properties: 1. They + point to resources _external_ to the Databricks compute; + therefore any associated authentication information + (typically a personal access token, OAuth token, or similar) + _must be removed_ when fetching from these links. 2. These + are URLs with a specific expiration, indicated in the + response. The behavior when attempting to use an expired + link is cloud specific. + format: + type: string + description: >- + Statement execution supports three result formats: + `JSON_ARRAY` (default), `ARROW_STREAM`, and `CSV`. + Important: The formats `ARROW_STREAM` and `CSV` are + supported only with `EXTERNAL_LINKS` disposition. + `JSON_ARRAY` is supported in `INLINE` and `EXTERNAL_LINKS` + disposition. When specifying `format=JSON_ARRAY`, result + data will be formatted as an array of arrays of values, + where each value is either the *string representation* of a + value, or `null`. For example, the output of `SELECT + concat('id-', id) AS strCol, id AS intCol, null AS nullCol + FROM range(3)` would look like this: ``` [ [ "id-1", "1", + null ], [ "id-2", "2", null ], [ "id-3", "3", null ], ] ``` + When specifying `format=JSON_ARRAY` and + `disposition=EXTERNAL_LINKS`, each chunk in the result + contains compact JSON with no indentation or extra + whitespace. When specifying `format=ARROW_STREAM` and + `disposition=EXTERNAL_LINKS`, each chunk in the result will + be formatted as Apache Arrow Stream. See the [Apache Arrow + streaming format]. When specifying `format=CSV` and + `disposition=EXTERNAL_LINKS`, each chunk in the result will + be a CSV according to [RFC 4180] standard. All the columns + values will have *string representation* similar to the + `JSON_ARRAY` format, and `null` values will be encoded as + “null”. Only the first chunk in the result would contain a + header row with column names. For example, the output of + `SELECT concat('id-', id) AS strCol, id AS intCol, null as + nullCol FROM range(3)` would look like this: ``` + strCol,intCol,nullCol id-1,1,null id-2,2,null id-3,3,null + ``` [Apache Arrow streaming format]: + https://arrow.apache.org/docs/format/Columnar.html#ipc-streaming-format + [RFC 4180]: https://www.rfc-editor.org/rfc/rfc4180 + on_wait_timeout: + type: string + description: >- + When `wait_timeout > 0s`, the call will block up to the + specified time. If the statement execution doesn't finish + within this time, `on_wait_timeout` determines whether the + execution should continue or be canceled. When set to + `CONTINUE`, the statement execution continues asynchronously + and the call returns a statement ID which can be used for + polling with :method:statementexecution/getStatement. When + set to `CANCEL`, the statement execution is canceled and the + call returns with a `CANCELED` state. + parameters: + type: string + description: >- + A list of parameters to pass into a SQL statement containing + parameter markers. A parameter consists of a name, a value, + and optionally a type. To represent a NULL value, the + `value` field may be omitted or set to `null` explicitly. If + the `type` field is omitted, the value is interpreted as a + string. If the type is given, parameters will be checked for + type correctness according to the given type. A value is + correct if the provided string can be converted to the + requested type using the `cast` function. The exact + semantics are described in the section [`cast` function] of + the SQL language reference. For example, the following + statement contains two parameters, `my_name` and `my_date`: + ``` SELECT * FROM my_table WHERE name = :my_name AND date = + :my_date ``` The parameters can be passed in the request + body as follows: ` { ..., "statement": "SELECT * FROM + my_table WHERE name = :my_name AND date = :my_date", + "parameters": [ { "name": "my_name", "value": "the name" }, + { "name": "my_date", "value": "2020-01-01", "type": "DATE" } + ] } ` Currently, positional parameters denoted by a `?` + marker are not supported by the Databricks SQL Statement + Execution API. Also see the section [Parameter markers] of + the SQL language reference. [Parameter markers]: + https://docs.databricks.com/sql/language-manual/sql-ref-parameter-marker.html + [`cast` function]: + https://docs.databricks.com/sql/language-manual/functions/cast.html + query_tags: + type: string + description: >- + An array of query tags to annotate a SQL statement. A query + tag consists of a non-empty key and, optionally, a value. To + represent a NULL value, either omit the `value` field or + manually set it to `null` or white space. Refer to the SQL + language reference for the format specification of query + tags. There's no significance to the order of tags. Only one + value per key will be recorded. A sequence in excess of 20 + query tags will be coerced to 20. Example: { ..., + "query_tags": [ { "key": "team", "value": "eng" }, { "key": + "some key only tag" } ] } + row_limit: + type: string + description: >- + Applies the given row limit to the statement's result set, + but unlike the `LIMIT` clause in SQL, it also sets the + `truncated` field in the response to indicate whether the + result was trimmed due to the limit or not. + schema: + type: string + description: >- + Sets default schema for statement execution, similar to + [`USE SCHEMA`] in SQL. [`USE SCHEMA`]: + https://docs.databricks.com/sql/language-manual/sql-ref-syntax-ddl-use-schema.html + wait_timeout: + type: string + description: >- + The time in seconds the call will wait for the statement's + result set as `Ns`, where `N` can be set to 0 or to a value + between 5 and 50. When set to `0s`, the statement will + execute in asynchronous mode and the call will not wait for + the execution to finish. In this case, the call returns + directly with `PENDING` state and a statement ID which can + be used for polling with + :method:statementexecution/getStatement. When set between 5 + and 50 seconds, the call will behave synchronously up to + this timeout and wait for the statement execution to finish. + If the execution finishes within this time, the call returns + immediately with a manifest and result data (or a `FAILED` + state in case of an execution error). If the statement takes + longer to execute, `on_wait_timeout` determines what should + happen after the timeout is reached. + required: + - statement + - warehouse_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StatementResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StatementExecutionAPI + /api/2.0/sql/statements/{statement_id}: + get: + operationId: statement_execution_get_statement + summary: >- + This request can be used to poll for the statement's status. + StatementResponse contains `statement_id` + tags: + - sql + - statement_execution + description: >- + This request can be used to poll for the statement's status. + StatementResponse contains `statement_id` + + and `status`; other fields might be absent or present depending on + context. When the `status.state` + + field is `SUCCEEDED` it will also return the result manifest and the + first chunk of the result data. + + When the statement is in the terminal states `CANCELED`, `CLOSED` or + `FAILED`, it returns HTTP 200 + + with the state set. After at least 12 hours in terminal state, the + statement is removed from the + + warehouse and further calls will receive an HTTP 404 response. + + + **NOTE** This call currently might take up to 5 seconds to get the + latest status and result. + + + :param statement_id: str + The statement ID is returned upon successfully submitting a SQL statement, and is a required + reference for all subsequent calls. + + :returns: :class:`StatementResponse` + parameters: + - name: statement_id + in: path + required: true + schema: + type: string + description: >- + The statement ID is returned upon successfully submitting a SQL + statement, and is a required reference for all subsequent calls. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/StatementResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StatementExecutionAPI + /api/2.0/sql/statements/{statement_id}/result/chunks/{chunk_index}: + get: + operationId: statement_execution_get_statement_result_chunk_n + summary: >- + After the statement execution has `SUCCEEDED`, this request can be used + to fetch any chunk by index. + tags: + - sql + - statement_execution + description: >- + After the statement execution has `SUCCEEDED`, this request can be used + to fetch any chunk by index. + + Whereas the first chunk with `chunk_index=0` is typically fetched with + + :method:statementexecution/executeStatement or + :method:statementexecution/getStatement, this request + + can be used to fetch subsequent chunks. The response structure is + identical to the nested `result` + + element described in the :method:statementexecution/getStatement + request, and similarly includes the + + `next_chunk_index` and `next_chunk_internal_link` fields for simple + iteration through the result set. + + Depending on `disposition`, the response returns chunks of data either + inline, or as links. + + + :param statement_id: str + The statement ID is returned upon successfully submitting a SQL statement, and is a required + reference for all subsequent calls. + :param chunk_index: int + + + :returns: :class:`ResultData` + parameters: + - name: statement_id + in: path + required: true + schema: + type: string + description: >- + The statement ID is returned upon successfully submitting a SQL + statement, and is a required reference for all subsequent calls. + - name: chunk_index + in: path + required: true + schema: + type: integer + description: ':returns: :class:`ResultData`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ResultData' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: StatementExecutionAPI + /api/2.0/sql/warehouses: + post: + operationId: warehouses_create + summary: Creates a new SQL warehouse. + tags: + - sql + - warehouses + description: >- + Creates a new SQL warehouse. + + + :param auto_stop_mins: int (optional) + The amount of time in minutes that a SQL warehouse must be idle (i.e., no RUNNING queries) before it + is automatically stopped. + + Supported values: - Must be == 0 or >= 10 mins - 0 indicates no autostop. + + Defaults to 120 mins + :param channel: :class:`Channel` (optional) + Channel Details + :param cluster_size: str (optional) + Size of the clusters allocated for this warehouse. Increasing the size of a spark cluster allows you + to run larger queries on it. If you want to increase the number of concurrent queries, please tune + max_num_clusters. + + Supported values: - 2X-Small - X-Small - Small - Medium - Large - X-Large - 2X-Large - 3X-Large - + 4X-Large + :param creator_name: str (optional) + warehouse creator name + :param enable_photon: bool (optional) + Configures whether the warehouse should use Photon optimized clusters. + + Defaults to false. + :param enable_serverless_compute: bool (optional) + Configures whether the warehouse should use serverless compute + :param instance_profile_arn: str (optional) + Deprecated. Instance profile used to pass IAM role to the cluster + :param max_num_clusters: int (optional) + Maximum number of clusters that the autoscaler will create to handle concurrent queries. + + Supported values: - Must be >= min_num_clusters - Must be <= 40. + + Defaults to min_clusters if unset. + :param min_num_clusters: int (optional) + Minimum number of available clusters that will be maintained for this SQL warehouse. Increasing this + will ensure that a larger number of clusters are always running and therefore may reduce the cold + start time for new queries. This is similar to reserved vs. revocable cores in a resource manager. + + Supported values: - Must be > 0 - Must be <= min(max_num_clusters, 30) + + Defaults to 1 + :param name: str (optional) + Logical name for the cluster. + + Supported values: - Must be unique within an org. - Must be less than 100 characters. + :param spot_instance_policy: :class:`SpotInstancePolicy` (optional) + Configurations whether the endpoint should use spot instances. + :param tags: :class:`EndpointTags` (optional) + A set of key-value pairs that will be tagged on all resources (e.g., AWS instances and EBS volumes) + associated with this SQL warehouse. + + Supported values: - Number of tags < 45. + :param warehouse_type: :class:`CreateWarehouseRequestWarehouseType` + (optional) + Warehouse type: `PRO` or `CLASSIC`. If you want to use serverless compute, you must set to `PRO` and + also set the field `enable_serverless_compute` to `true`. + + :returns: + Long-running operation waiter for :class:`GetWarehouseResponse`. + See :method:wait_get_warehouse_running for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + auto_stop_mins: + type: string + description: >- + The amount of time in minutes that a SQL warehouse must be + idle (i.e., no RUNNING queries) before it is automatically + stopped. Supported values: - Must be == 0 or >= 10 mins - 0 + indicates no autostop. Defaults to 120 mins + channel: + type: string + description: Channel Details + cluster_size: + type: string + description: >- + Size of the clusters allocated for this warehouse. + Increasing the size of a spark cluster allows you to run + larger queries on it. If you want to increase the number of + concurrent queries, please tune max_num_clusters. Supported + values: - 2X-Small - X-Small - Small - Medium - Large - + X-Large - 2X-Large - 3X-Large - 4X-Large + creator_name: + type: string + description: warehouse creator name + enable_photon: + type: string + description: >- + Configures whether the warehouse should use Photon optimized + clusters. Defaults to false. + enable_serverless_compute: + type: string + description: >- + Configures whether the warehouse should use serverless + compute + instance_profile_arn: + type: string + description: >- + Deprecated. Instance profile used to pass IAM role to the + cluster + max_num_clusters: + type: string + description: >- + Maximum number of clusters that the autoscaler will create + to handle concurrent queries. Supported values: - Must be >= + min_num_clusters - Must be <= 40. Defaults to min_clusters + if unset. + min_num_clusters: + type: string + description: >- + Minimum number of available clusters that will be maintained + for this SQL warehouse. Increasing this will ensure that a + larger number of clusters are always running and therefore + may reduce the cold start time for new queries. This is + similar to reserved vs. revocable cores in a resource + manager. Supported values: - Must be > 0 - Must be <= + min(max_num_clusters, 30) Defaults to 1 + name: + type: string + description: >- + Logical name for the cluster. Supported values: - Must be + unique within an org. - Must be less than 100 characters. + spot_instance_policy: + type: string + description: >- + Configurations whether the endpoint should use spot + instances. + tags: + type: string + description: >- + A set of key-value pairs that will be tagged on all + resources (e.g., AWS instances and EBS volumes) associated + with this SQL warehouse. Supported values: - Number of tags + < 45. + warehouse_type: + type: string + description: >- + Warehouse type: `PRO` or `CLASSIC`. If you want to use + serverless compute, you must set to `PRO` and also set the + field `enable_serverless_compute` to `true`. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWarehouseResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + get: + operationId: warehouses_list + summary: Lists all SQL warehouses that a user has access to. + tags: + - sql + - warehouses + description: |- + Lists all SQL warehouses that a user has access to. + + :param page_size: int (optional) + The max number of warehouses to return. + :param page_token: str (optional) + A page token, received from a previous `ListWarehouses` call. Provide this to retrieve the + subsequent page; otherwise the first will be retrieved. + + When paginating, all other parameters provided to `ListWarehouses` must match the call that provided + the page token. + :param run_as_user_id: int (optional) + Service Principal which will be used to fetch the list of endpoints. If not specified, SQL Gateway + will use the user from the session header. + + :returns: Iterator over :class:`EndpointInfo` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: The max number of warehouses to return. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous `ListWarehouses` call. + Provide this to retrieve the subsequent page; otherwise the first + will be retrieved. When paginating, all other parameters provided to + `ListWarehouses` must match the call that provided the page token. + - name: run_as_user_id + in: query + required: false + schema: + type: string + description: >- + Service Principal which will be used to fetch the list of endpoints. + If not specified, SQL Gateway will use the user from the session + header. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListWarehousesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/warehouses/v1/default-warehouse-overrides: + post: + operationId: warehouses_create_default_warehouse_override + summary: >- + Creates a new default warehouse override for a user. Users can create + their own override. Admins can + tags: + - sql + - warehouses + description: >- + Creates a new default warehouse override for a user. Users can create + their own override. Admins can + + create overrides for any user. + + + :param default_warehouse_override: :class:`DefaultWarehouseOverride` + Required. The default warehouse override to create. + :param default_warehouse_override_id: str + Required. The ID to use for the override, which will become the final component of the override's + resource name. Can be a numeric user ID or the literal string "me" for the current user. + + :returns: :class:`DefaultWarehouseOverride` + parameters: + - name: default_warehouse_override_id + in: query + required: true + schema: + type: string + description: >- + Required. The ID to use for the override, which will become the + final component of the override's resource name. Can be a numeric + user ID or the literal string "me" for the current user. + requestBody: + content: + application/json: + schema: + type: object + properties: + default_warehouse_override: + type: string + description: Required. The default warehouse override to create. + required: + - default_warehouse_override + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DefaultWarehouseOverride' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + get: + operationId: warehouses_list_default_warehouse_overrides + summary: >- + Lists all default warehouse overrides in the workspace. Only workspace + administrators can list all + tags: + - sql + - warehouses + description: >- + Lists all default warehouse overrides in the workspace. Only workspace + administrators can list all + + overrides. + + + :param page_size: int (optional) + The maximum number of overrides to return. The service may return fewer than this value. If + unspecified, at most 100 overrides will be returned. The maximum value is 1000; values above 1000 + will be coerced to 1000. + :param page_token: str (optional) + A page token, received from a previous `ListDefaultWarehouseOverrides` call. Provide this to + retrieve the subsequent page. + + When paginating, all other parameters provided to `ListDefaultWarehouseOverrides` must match the + call that provided the page token. + + :returns: Iterator over :class:`DefaultWarehouseOverride` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of overrides to return. The service may return + fewer than this value. If unspecified, at most 100 overrides will be + returned. The maximum value is 1000; values above 1000 will be + coerced to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + A page token, received from a previous + `ListDefaultWarehouseOverrides` call. Provide this to retrieve the + subsequent page. When paginating, all other parameters provided to + `ListDefaultWarehouseOverrides` must match the call that provided + the page token. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListDefaultWarehouseOverridesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/2.0/sql/warehouses/{id}: + delete: + operationId: warehouses_delete + summary: Deletes a SQL warehouse. + tags: + - sql + - warehouses + description: |- + Deletes a SQL warehouse. + + :param id: str + Required. Id of the SQL warehouse. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Required. Id of the SQL warehouse. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + get: + operationId: warehouses_get + summary: Gets the information for a single SQL warehouse. + tags: + - sql + - warehouses + description: |- + Gets the information for a single SQL warehouse. + + :param id: str + Required. Id of the SQL warehouse. + + :returns: :class:`GetWarehouseResponse` + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Required. Id of the SQL warehouse. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWarehouseResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/warehouses/v1/{name}: + delete: + operationId: warehouses_delete_default_warehouse_override + summary: >- + Deletes the default warehouse override for a user. Users can delete + their own override. Admins can + tags: + - sql + - warehouses + description: >- + Deletes the default warehouse override for a user. Users can delete + their own override. Admins can + + delete overrides for any user. After deletion, the workspace default + warehouse will be used. + + + :param name: str + Required. The resource name of the default warehouse override to delete. Format: + default-warehouse-overrides/{default_warehouse_override_id} The default_warehouse_override_id can be + a numeric user ID or the literal string "me" for the current user. + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + Required. The resource name of the default warehouse override to + delete. Format: + default-warehouse-overrides/{default_warehouse_override_id} The + default_warehouse_override_id can be a numeric user ID or the + literal string "me" for the current user. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + get: + operationId: warehouses_get_default_warehouse_override + summary: >- + Returns the default warehouse override for a user. Users can fetch their + own override. Admins can + tags: + - sql + - warehouses + description: >- + Returns the default warehouse override for a user. Users can fetch their + own override. Admins can + + fetch overrides for any user. If no override exists, the UI will + fallback to the workspace default + + warehouse. + + + :param name: str + Required. The resource name of the default warehouse override to retrieve. Format: + default-warehouse-overrides/{default_warehouse_override_id} The default_warehouse_override_id can be + a numeric user ID or the literal string "me" for the current user. + + :returns: :class:`DefaultWarehouseOverride` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + Required. The resource name of the default warehouse override to + retrieve. Format: + default-warehouse-overrides/{default_warehouse_override_id} The + default_warehouse_override_id can be a numeric user ID or the + literal string "me" for the current user. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DefaultWarehouseOverride' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + patch: + operationId: warehouses_update_default_warehouse_override + summary: >- + Updates an existing default warehouse override for a user. Users can + update their own override. Admins + tags: + - sql + - warehouses + description: >- + Updates an existing default warehouse override for a user. Users can + update their own override. Admins + + can update overrides for any user. + + + :param name: str + The resource name of the default warehouse override. Format: + default-warehouse-overrides/{default_warehouse_override_id} + :param default_warehouse_override: :class:`DefaultWarehouseOverride` + Required. The default warehouse override to update. The name field must be set in the format: + default-warehouse-overrides/{default_warehouse_override_id} The default_warehouse_override_id can be + a numeric user ID or the literal string "me" for the current user. + :param update_mask: FieldMask + Required. Field mask specifying which fields to update. Only the fields specified in the mask will + be updated. Use "*" to update all fields. When allow_missing is true, this field is ignored and all + fields are applied. + :param allow_missing: bool (optional) + If set to true, and the override is not found, a new override will be created. In this situation, + `update_mask` is ignored and all fields are applied. Defaults to false. + + :returns: :class:`DefaultWarehouseOverride` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: >- + The resource name of the default warehouse override. Format: + default-warehouse-overrides/{default_warehouse_override_id} + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + Required. Field mask specifying which fields to update. Only the + fields specified in the mask will be updated. Use "*" to update all + fields. When allow_missing is true, this field is ignored and all + fields are applied. + - name: allow_missing + in: query + required: false + schema: + type: string + description: >- + If set to true, and the override is not found, a new override will + be created. In this situation, `update_mask` is ignored and all + fields are applied. Defaults to false. + requestBody: + content: + application/json: + schema: + type: object + properties: + default_warehouse_override: + type: string + description: >- + Required. The default warehouse override to update. The name + field must be set in the format: + default-warehouse-overrides/{default_warehouse_override_id} + The default_warehouse_override_id can be a numeric user ID + or the literal string "me" for the current user. + required: + - default_warehouse_override + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DefaultWarehouseOverride' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/2.0/sql/warehouses/{id}/edit: + post: + operationId: warehouses_edit + summary: Updates the configuration for a SQL warehouse. + tags: + - sql + - warehouses + description: >- + Updates the configuration for a SQL warehouse. + + + :param id: str + Required. Id of the warehouse to configure. + :param auto_stop_mins: int (optional) + The amount of time in minutes that a SQL warehouse must be idle (i.e., no RUNNING queries) before it + is automatically stopped. + + Supported values: - Must be == 0 or >= 10 mins - 0 indicates no autostop. + + Defaults to 120 mins + :param channel: :class:`Channel` (optional) + Channel Details + :param cluster_size: str (optional) + Size of the clusters allocated for this warehouse. Increasing the size of a spark cluster allows you + to run larger queries on it. If you want to increase the number of concurrent queries, please tune + max_num_clusters. + + Supported values: - 2X-Small - X-Small - Small - Medium - Large - X-Large - 2X-Large - 3X-Large - + 4X-Large + :param creator_name: str (optional) + warehouse creator name + :param enable_photon: bool (optional) + Configures whether the warehouse should use Photon optimized clusters. + + Defaults to false. + :param enable_serverless_compute: bool (optional) + Configures whether the warehouse should use serverless compute + :param instance_profile_arn: str (optional) + Deprecated. Instance profile used to pass IAM role to the cluster + :param max_num_clusters: int (optional) + Maximum number of clusters that the autoscaler will create to handle concurrent queries. + + Supported values: - Must be >= min_num_clusters - Must be <= 40. + + Defaults to min_clusters if unset. + :param min_num_clusters: int (optional) + Minimum number of available clusters that will be maintained for this SQL warehouse. Increasing this + will ensure that a larger number of clusters are always running and therefore may reduce the cold + start time for new queries. This is similar to reserved vs. revocable cores in a resource manager. + + Supported values: - Must be > 0 - Must be <= min(max_num_clusters, 30) + + Defaults to 1 + :param name: str (optional) + Logical name for the cluster. + + Supported values: - Must be unique within an org. - Must be less than 100 characters. + :param spot_instance_policy: :class:`SpotInstancePolicy` (optional) + Configurations whether the endpoint should use spot instances. + :param tags: :class:`EndpointTags` (optional) + A set of key-value pairs that will be tagged on all resources (e.g., AWS instances and EBS volumes) + associated with this SQL warehouse. + + Supported values: - Number of tags < 45. + :param warehouse_type: :class:`EditWarehouseRequestWarehouseType` + (optional) + Warehouse type: `PRO` or `CLASSIC`. If you want to use serverless compute, you must set to `PRO` and + also set the field `enable_serverless_compute` to `true`. + + :returns: + Long-running operation waiter for :class:`GetWarehouseResponse`. + See :method:wait_get_warehouse_running for more details. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Required. Id of the warehouse to configure. + requestBody: + content: + application/json: + schema: + type: object + properties: + auto_stop_mins: + type: string + description: >- + The amount of time in minutes that a SQL warehouse must be + idle (i.e., no RUNNING queries) before it is automatically + stopped. Supported values: - Must be == 0 or >= 10 mins - 0 + indicates no autostop. Defaults to 120 mins + channel: + type: string + description: Channel Details + cluster_size: + type: string + description: >- + Size of the clusters allocated for this warehouse. + Increasing the size of a spark cluster allows you to run + larger queries on it. If you want to increase the number of + concurrent queries, please tune max_num_clusters. Supported + values: - 2X-Small - X-Small - Small - Medium - Large - + X-Large - 2X-Large - 3X-Large - 4X-Large + creator_name: + type: string + description: warehouse creator name + enable_photon: + type: string + description: >- + Configures whether the warehouse should use Photon optimized + clusters. Defaults to false. + enable_serverless_compute: + type: string + description: >- + Configures whether the warehouse should use serverless + compute + instance_profile_arn: + type: string + description: >- + Deprecated. Instance profile used to pass IAM role to the + cluster + max_num_clusters: + type: string + description: >- + Maximum number of clusters that the autoscaler will create + to handle concurrent queries. Supported values: - Must be >= + min_num_clusters - Must be <= 40. Defaults to min_clusters + if unset. + min_num_clusters: + type: string + description: >- + Minimum number of available clusters that will be maintained + for this SQL warehouse. Increasing this will ensure that a + larger number of clusters are always running and therefore + may reduce the cold start time for new queries. This is + similar to reserved vs. revocable cores in a resource + manager. Supported values: - Must be > 0 - Must be <= + min(max_num_clusters, 30) Defaults to 1 + name: + type: string + description: >- + Logical name for the cluster. Supported values: - Must be + unique within an org. - Must be less than 100 characters. + spot_instance_policy: + type: string + description: >- + Configurations whether the endpoint should use spot + instances. + tags: + type: string + description: >- + A set of key-value pairs that will be tagged on all + resources (e.g., AWS instances and EBS volumes) associated + with this SQL warehouse. Supported values: - Number of tags + < 45. + warehouse_type: + type: string + description: >- + Warehouse type: `PRO` or `CLASSIC`. If you want to use + serverless compute, you must set to `PRO` and also set the + field `enable_serverless_compute` to `true`. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWarehouseResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/2.0/permissions/warehouses/{warehouse_id}/permissionLevels: + get: + operationId: warehouses_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - sql + - warehouses + description: |- + Gets the permission levels that a user can have on an object. + + :param warehouse_id: str + The SQL warehouse for which to get or manage permissions. + + :returns: :class:`GetWarehousePermissionLevelsResponse` + parameters: + - name: warehouse_id + in: path + required: true + schema: + type: string + description: The SQL warehouse for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWarehousePermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/2.0/permissions/warehouses/{warehouse_id}: + get: + operationId: warehouses_get_permissions + summary: >- + Gets the permissions of a SQL warehouse. SQL warehouses can inherit + permissions from their root + tags: + - sql + - warehouses + description: >- + Gets the permissions of a SQL warehouse. SQL warehouses can inherit + permissions from their root + + object. + + + :param warehouse_id: str + The SQL warehouse for which to get or manage permissions. + + :returns: :class:`WarehousePermissions` + parameters: + - name: warehouse_id + in: path + required: true + schema: + type: string + description: The SQL warehouse for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WarehousePermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + put: + operationId: warehouses_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - sql + - warehouses + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param warehouse_id: str + The SQL warehouse for which to get or manage permissions. + :param access_control_list: List[:class:`WarehouseAccessControlRequest`] + (optional) + + + :returns: :class:`WarehousePermissions` + parameters: + - name: warehouse_id + in: path + required: true + schema: + type: string + description: The SQL warehouse for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`WarehousePermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WarehousePermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + patch: + operationId: warehouses_update_permissions + summary: >- + Updates the permissions on a SQL warehouse. SQL warehouses can inherit + permissions from their root + tags: + - sql + - warehouses + description: >- + Updates the permissions on a SQL warehouse. SQL warehouses can inherit + permissions from their root + + object. + + + :param warehouse_id: str + The SQL warehouse for which to get or manage permissions. + :param access_control_list: List[:class:`WarehouseAccessControlRequest`] + (optional) + + + :returns: :class:`WarehousePermissions` + parameters: + - name: warehouse_id + in: path + required: true + schema: + type: string + description: The SQL warehouse for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`WarehousePermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WarehousePermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/2.0/sql/config/warehouses: + get: + operationId: warehouses_get_workspace_warehouse_config + summary: >- + Gets the workspace level configuration that is shared by all SQL + warehouses in a workspace. + tags: + - sql + - warehouses + description: >- + Gets the workspace level configuration that is shared by all SQL + warehouses in a workspace. + + + + :returns: :class:`GetWorkspaceWarehouseConfigResponse` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWorkspaceWarehouseConfigResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + put: + operationId: warehouses_set_workspace_warehouse_config + summary: >- + Sets the workspace level configuration that is shared by all SQL + warehouses in a workspace. + tags: + - sql + - warehouses + description: >- + Sets the workspace level configuration that is shared by all SQL + warehouses in a workspace. + + + :param channel: :class:`Channel` (optional) + Optional: Channel selection details + :param config_param: :class:`RepeatedEndpointConfPairs` (optional) + Deprecated: Use sql_configuration_parameters + :param data_access_config: List[:class:`EndpointConfPair`] (optional) + Spark confs for external hive metastore configuration JSON serialized size must be less than <= 512K + :param enable_serverless_compute: bool (optional) + Enable Serverless compute for SQL warehouses + :param enabled_warehouse_types: List[:class:`WarehouseTypePair`] + (optional) + List of Warehouse Types allowed in this workspace (limits allowed value of the type field in + CreateWarehouse and EditWarehouse). Note: Some types cannot be disabled, they don't need to be + specified in SetWorkspaceWarehouseConfig. Note: Disabling a type may cause existing warehouses to be + converted to another type. Used by frontend to save specific type availability in the warehouse + create and edit form UI. + :param global_param: :class:`RepeatedEndpointConfPairs` (optional) + Deprecated: Use sql_configuration_parameters + :param google_service_account: str (optional) + GCP only: Google Service Account used to pass to cluster to access Google Cloud Storage + :param instance_profile_arn: str (optional) + AWS Only: The instance profile used to pass an IAM role to the SQL warehouses. This configuration is + also applied to the workspace's serverless compute for notebooks and jobs. + :param security_policy: + :class:`SetWorkspaceWarehouseConfigRequestSecurityPolicy` (optional) + Security policy for warehouses + :param sql_configuration_parameters: :class:`RepeatedEndpointConfPairs` + (optional) + SQL configuration parameters + requestBody: + content: + application/json: + schema: + type: object + properties: + channel: + type: string + description: 'Optional: Channel selection details' + config_param: + type: string + description: 'Deprecated: Use sql_configuration_parameters' + data_access_config: + type: string + description: >- + Spark confs for external hive metastore configuration JSON + serialized size must be less than <= 512K + enable_serverless_compute: + type: string + description: Enable Serverless compute for SQL warehouses + enabled_warehouse_types: + type: string + description: >- + List of Warehouse Types allowed in this workspace (limits + allowed value of the type field in CreateWarehouse and + EditWarehouse). Note: Some types cannot be disabled, they + don't need to be specified in SetWorkspaceWarehouseConfig. + Note: Disabling a type may cause existing warehouses to be + converted to another type. Used by frontend to save specific + type availability in the warehouse create and edit form UI. + global_param: + type: string + description: 'Deprecated: Use sql_configuration_parameters' + google_service_account: + type: string + description: >- + GCP only: Google Service Account used to pass to cluster to + access Google Cloud Storage + instance_profile_arn: + type: string + description: >- + AWS Only: The instance profile used to pass an IAM role to + the SQL warehouses. This configuration is also applied to + the workspace's serverless compute for notebooks and jobs. + security_policy: + type: string + description: Security policy for warehouses + sql_configuration_parameters: + type: string + description: SQL configuration parameters + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/2.0/sql/warehouses/{id}/start: + post: + operationId: warehouses_start + summary: Starts a SQL warehouse. + tags: + - sql + - warehouses + description: |- + Starts a SQL warehouse. + + :param id: str + Required. Id of the SQL warehouse. + + :returns: + Long-running operation waiter for :class:`GetWarehouseResponse`. + See :method:wait_get_warehouse_running for more details. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Required. Id of the SQL warehouse. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWarehouseResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI + /api/2.0/sql/warehouses/{id}/stop: + post: + operationId: warehouses_stop + summary: Stops a SQL warehouse. + tags: + - sql + - warehouses + description: |- + Stops a SQL warehouse. + + :param id: str + Required. Id of the SQL warehouse. + + :returns: + Long-running operation waiter for :class:`GetWarehouseResponse`. + See :method:wait_get_warehouse_stopped for more details. + parameters: + - name: id + in: path + required: true + schema: + type: string + description: Required. Id of the SQL warehouse. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetWarehouseResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WarehousesAPI +components: + schemas: + AccessControl: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/PermissionLevel' + description: >- + * `CAN_VIEW`: Can view the query * `CAN_RUN`: Can run the query * + `CAN_EDIT`: Can edit the query * `CAN_MANAGE`: Can manage the query + user_name: + type: string + Alert: + type: object + properties: + condition: + $ref: '#/components/schemas/AlertCondition' + create_time: + type: string + description: The timestamp indicating when the alert was created. + custom_body: + type: string + description: >- + Custom body of alert notification, if it exists. See [here] for + custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + custom_subject: + type: string + description: >- + Custom subject of alert notification, if it exists. This can include + email subject entries and Slack notification headers, for example. + See [here] for custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + display_name: + type: string + description: The display name of the alert. + id: + type: string + description: UUID identifying the alert. + lifecycle_state: + $ref: '#/components/schemas/LifecycleState' + description: The workspace state of the alert. Used for tracking trashed status. + notify_on_ok: + type: boolean + description: >- + Whether to notify alert subscribers when alert returns back to + normal. + owner_user_name: + type: string + description: >- + The owner's username. This field is set to "Unavailable" if the user + has been deleted. + parent_path: + type: string + description: The workspace path of the folder containing the alert. + query_id: + type: string + description: UUID of the query attached to the alert. + seconds_to_retrigger: + type: integer + description: >- + Number of seconds an alert must wait after being triggered to rearm + itself. After rearming, it can be triggered again. If 0 or not + specified, the alert will not be triggered again. + state: + $ref: '#/components/schemas/AlertState' + description: >- + Current state of the alert's trigger status. This field is set to + UNKNOWN if the alert has not yet been evaluated or ran into an error + during the last evaluation. + trigger_time: + type: string + description: >- + Timestamp when the alert was last triggered, if the alert has been + triggered before. + update_time: + type: string + description: The timestamp indicating when the alert was updated. + AlertCondition: + type: object + properties: + empty_result_state: + $ref: '#/components/schemas/AlertState' + op: + $ref: '#/components/schemas/AlertOperator' + description: Operator used for comparison in alert evaluation. + operand: + $ref: '#/components/schemas/AlertConditionOperand' + description: >- + Name of the column from the query result to use for comparison in + alert evaluation. + threshold: + $ref: '#/components/schemas/AlertConditionThreshold' + description: Threshold value used for comparison in alert evaluation. + AlertConditionOperand: + type: object + properties: + column: + $ref: '#/components/schemas/AlertOperandColumn' + AlertConditionThreshold: + type: object + properties: + value: + $ref: '#/components/schemas/AlertOperandValue' + AlertOperandColumn: + type: object + properties: + name: + type: string + AlertOperandValue: + type: object + properties: + bool_value: + type: boolean + double_value: + type: number + string_value: + type: string + AlertOptions: + type: object + properties: + column: + type: string + description: Name of column in the query result to compare in alert evaluation. + op: + type: string + description: >- + Operator used to compare in alert evaluation: `>`, `>=`, `<`, `<=`, + `==`, `!=` + value: + type: object + description: >- + Value used to compare in alert evaluation. Supported types include + strings (eg. 'foobar'), floats (eg. 123.4), and booleans (true). + custom_body: + type: string + description: >- + Custom body of alert notification, if it exists. See [here] for + custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + custom_subject: + type: string + description: >- + Custom subject of alert notification, if it exists. This includes + email subject, Slack notification header, etc. See [here] for custom + templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + empty_result_state: + $ref: '#/components/schemas/AlertOptionsEmptyResultState' + description: State that alert evaluates to when query result is empty. + muted: + type: boolean + description: >- + Whether or not the alert is muted. If an alert is muted, it will not + notify users and notification destinations when triggered. + required: + - column + - op + - value + description: Alert configuration options. + AlertQuery: + type: object + properties: + created_at: + type: string + data_source_id: + type: string + description: >- + Data source ID maps to the ID of the data source used by the + resource and is distinct from the warehouse ID. [Learn more] [Learn + more]: https://docs.databricks.com/api/workspace/datasources/list + description: + type: string + description: >- + General description that conveys additional information about this + query such as usage notes. + id: + type: string + description: Query ID. + is_archived: + type: boolean + description: >- + Indicates whether the query is trashed. Trashed queries can't be + used in dashboards, or appear in search results. If this boolean is + `true`, the `options` property for this query includes a + `moved_to_trash_at` timestamp. Trashed queries are permanently + deleted after 30 days. + is_draft: + type: boolean + description: >- + Whether the query is a draft. Draft queries only appear in list + views for their owners. Visualizations from draft queries cannot + appear on dashboards. + is_safe: + type: boolean + description: >- + Text parameter types are not safe from SQL injection for all types + of data source. Set this Boolean parameter to `true` if a query + either does not use any text type parameters or uses a data source + type where text type parameters are handled safely. + name: + type: string + description: >- + The title of this query that appears in list views, widget headings, + and on the query page. + options: + $ref: '#/components/schemas/QueryOptions' + query: + type: string + description: The text of the query to be run. + tags: + type: array + items: + type: string + updated_at: + type: string + description: The timestamp at which this query was last updated. + user_id: + type: integer + description: The ID of the user who owns the query. + AlertV2: + type: object + properties: + display_name: + type: string + query_text: + type: string + description: Text of the query to be run. + warehouse_id: + type: string + description: ID of the SQL warehouse attached to the alert. + evaluation: + $ref: '#/components/schemas/AlertV2Evaluation' + schedule: + $ref: '#/components/schemas/CronSchedule' + create_time: + type: string + description: The timestamp indicating when the alert was created. + custom_description: + type: string + description: Custom description for the alert. support mustache template. + custom_summary: + type: string + description: Custom summary for the alert. support mustache template. + effective_run_as: + $ref: '#/components/schemas/AlertV2RunAs' + description: >- + The actual identity that will be used to execute the alert. This is + an output-only field that shows the resolved run-as identity after + applying permissions and defaults. + id: + type: string + description: UUID identifying the alert. + lifecycle_state: + $ref: '#/components/schemas/AlertLifecycleState' + description: Indicates whether the query is trashed. + owner_user_name: + type: string + description: >- + The owner's username. This field is set to "Unavailable" if the user + has been deleted. + parent_path: + type: string + description: >- + The workspace path of the folder containing the alert. Can only be + set on create, and cannot be updated. + run_as: + $ref: '#/components/schemas/AlertV2RunAs' + description: >- + Specifies the identity that will be used to run the alert. This + field allows you to configure alerts to run as a specific user or + service principal. - For user identity: Set `user_name` to the email + of an active workspace user. Users can only set this to their own + email. - For service principal: Set `service_principal_name` to the + application ID. Requires the `servicePrincipal/user` role. If not + specified, the alert will run as the request user. + run_as_user_name: + type: string + description: >- + The run as username or application ID of service principal. On + Create and Update, this field can be set to application ID of an + active service principal. Setting this field requires the + servicePrincipal/user role. Deprecated: Use `run_as` field instead. + This field will be removed in a future release. + update_time: + type: string + description: The timestamp indicating when the alert was updated. + required: + - display_name + - query_text + - warehouse_id + - evaluation + - schedule + AlertV2Evaluation: + type: object + properties: + source: + $ref: '#/components/schemas/AlertV2OperandColumn' + comparison_operator: + $ref: '#/components/schemas/ComparisonOperator' + description: Operator used for comparison in alert evaluation. + empty_result_state: + $ref: '#/components/schemas/AlertEvaluationState' + description: >- + Alert state if result is empty. Please avoid setting this field to + be `UNKNOWN` because `UNKNOWN` state is planned to be deprecated. + last_evaluated_at: + type: string + description: Timestamp of the last evaluation. + notification: + $ref: '#/components/schemas/AlertV2Notification' + description: User or Notification Destination to notify when alert is triggered. + state: + $ref: '#/components/schemas/AlertEvaluationState' + description: Latest state of alert evaluation. + threshold: + $ref: '#/components/schemas/AlertV2Operand' + description: Threshold to user for alert evaluation, can be a column or a value. + required: + - source + - comparison_operator + AlertV2Notification: + type: object + properties: + notify_on_ok: + type: boolean + retrigger_seconds: + type: integer + description: >- + Number of seconds an alert waits after being triggered before it is + allowed to send another notification. If set to 0 or omitted, the + alert will not send any further notifications after the first + trigger Setting this value to 1 allows the alert to send a + notification on every evaluation where the condition is met, + effectively making it always retrigger for notification purposes. + subscriptions: + type: array + items: + $ref: '#/components/schemas/AlertV2Subscription' + AlertV2Operand: + type: object + properties: + column: + $ref: '#/components/schemas/AlertV2OperandColumn' + value: + $ref: '#/components/schemas/AlertV2OperandValue' + AlertV2OperandColumn: + type: object + properties: + name: + type: string + aggregation: + $ref: '#/components/schemas/Aggregation' + description: >- + If not set, the behavior is equivalent to using `First row` in the + UI. + display: + type: string + required: + - name + AlertV2OperandValue: + type: object + properties: + bool_value: + type: boolean + double_value: + type: number + string_value: + type: string + AlertV2RunAs: + type: object + properties: + service_principal_name: + type: string + user_name: + type: string + description: >- + The email of an active workspace user. Can only set this field to + their own email. + AlertV2Subscription: + type: object + properties: + destination_id: + type: string + user_email: + type: string + BaseChunkInfo: + type: object + properties: + byte_count: + type: integer + chunk_index: + type: integer + description: The position within the sequence of result set chunks. + row_count: + type: integer + description: The number of rows within the result chunk. + row_offset: + type: integer + description: The starting row offset within the result set. + Channel: + type: object + properties: + dbsql_version: + type: string + name: + $ref: '#/components/schemas/ChannelName' + description: >- + Configures the channel name and DBSQL version of the warehouse. + CHANNEL_NAME_CUSTOM should be + chosen only when `dbsql_version` is specified. + ChannelInfo: + type: object + properties: + dbsql_version: + type: string + description: DB SQL Version the Channel is mapped to. + name: + $ref: '#/components/schemas/ChannelName' + description: Name of the channel + description: Details about a Channel. + ClientConfig: + type: object + properties: + allow_custom_js_visualizations: + type: boolean + allow_downloads: + type: boolean + allow_external_shares: + type: boolean + allow_subscriptions: + type: boolean + date_format: + type: string + date_time_format: + type: string + disable_publish: + type: boolean + enable_legacy_autodetect_types: + type: boolean + feature_show_permissions_control: + type: boolean + hide_plotly_mode_bar: + type: boolean + ColumnInfo: + type: object + properties: + name: + type: string + position: + type: integer + description: The ordinal position of the column (starting at position 0). + type_interval_type: + type: string + description: The format of the interval type. + type_name: + $ref: '#/components/schemas/ColumnInfoTypeName' + description: >- + The name of the base data type. This doesn't include details for + complex types such as STRUCT, MAP or ARRAY. + type_precision: + type: integer + description: >- + Specifies the number of digits in a number. This applies to the + DECIMAL type. + type_scale: + type: integer + description: >- + Specifies the number of digits to the right of the decimal point in + a number. This applies to the DECIMAL type. + type_text: + type: string + description: The full SQL type specification. + CreateAlertRequestAlert: + type: object + properties: + condition: + $ref: '#/components/schemas/AlertCondition' + custom_body: + type: string + description: >- + Custom body of alert notification, if it exists. See [here] for + custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + custom_subject: + type: string + description: >- + Custom subject of alert notification, if it exists. This can include + email subject entries and Slack notification headers, for example. + See [here] for custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + display_name: + type: string + description: The display name of the alert. + notify_on_ok: + type: boolean + description: >- + Whether to notify alert subscribers when alert returns back to + normal. + parent_path: + type: string + description: The workspace path of the folder containing the alert. + query_id: + type: string + description: UUID of the query attached to the alert. + seconds_to_retrigger: + type: integer + description: >- + Number of seconds an alert must wait after being triggered to rearm + itself. After rearming, it can be triggered again. If 0 or not + specified, the alert will not be triggered again. + CreateQueryRequestQuery: + type: object + properties: + apply_auto_limit: + type: boolean + catalog: + type: string + description: Name of the catalog where this query will be executed. + description: + type: string + description: >- + General description that conveys additional information about this + query such as usage notes. + display_name: + type: string + description: >- + Display name of the query that appears in list views, widget + headings, and on the query page. + parameters: + type: array + items: + $ref: '#/components/schemas/QueryParameter' + description: List of query parameter definitions. + parent_path: + type: string + description: Workspace path of the workspace folder containing the object. + query_text: + type: string + description: Text of the query to be run. + run_as_mode: + $ref: '#/components/schemas/RunAsMode' + description: Sets the "Run as" role for the object. + schema: + type: string + description: Name of the schema where this query will be executed. + tags: + type: array + items: + type: string + warehouse_id: + type: string + description: ID of the SQL warehouse attached to the query. + CreateVisualizationRequestVisualization: + type: object + properties: + display_name: + type: string + query_id: + type: string + description: UUID of the query that the visualization is attached to. + serialized_options: + type: string + description: >- + The visualization options varies widely from one visualization type + to the next and is unsupported. Databricks does not recommend + modifying visualization options directly. + serialized_query_plan: + type: string + description: >- + The visualization query plan varies widely from one visualization + type to the next and is unsupported. Databricks does not recommend + modifying the visualization query plan directly. + type: + type: string + description: 'The type of visualization: counter, table, funnel, and so on.' + CreateWarehouseResponse: + type: object + properties: + id: + type: string + CronSchedule: + type: object + properties: + quartz_cron_schedule: + type: string + timezone_id: + type: string + description: >- + A Java timezone id. The schedule will be resolved using this + timezone. This will be combined with the quartz_cron_schedule to + determine the schedule. See + https://docs.databricks.com/sql/language-manual/sql-ref-syntax-aux-conf-mgmt-set-timezone.html + for details. + pause_status: + $ref: '#/components/schemas/SchedulePauseStatus' + description: Indicate whether this schedule is paused or not. + required: + - quartz_cron_schedule + - timezone_id + Dashboard: + type: object + properties: + can_edit: + type: boolean + description: Whether the authenticated user can edit the query definition. + created_at: + type: string + description: Timestamp when this dashboard was created. + dashboard_filters_enabled: + type: boolean + description: >- + In the web application, query filters that share a name are coupled + to a single selection box if this value is `true`. + id: + type: string + description: The ID for this dashboard. + is_archived: + type: boolean + description: >- + Indicates whether a dashboard is trashed. Trashed dashboards won't + appear in list views. If this boolean is `true`, the `options` + property for this dashboard includes a `moved_to_trash_at` + timestamp. Items in trash are permanently deleted after 30 days. + is_draft: + type: boolean + description: >- + Whether a dashboard is a draft. Draft dashboards only appear in list + views for their owners. + is_favorite: + type: boolean + description: >- + Indicates whether this query object appears in the current user's + favorites list. This flag determines whether the star icon for + favorites is selected. + name: + type: string + description: >- + The title of the dashboard that appears in list views and at the top + of the dashboard page. + options: + $ref: '#/components/schemas/DashboardOptions' + parent: + type: string + description: The identifier of the workspace folder containing the object. + permission_tier: + $ref: '#/components/schemas/PermissionLevel' + description: >- + * `CAN_VIEW`: Can view the query * `CAN_RUN`: Can run the query * + `CAN_EDIT`: Can edit the query * `CAN_MANAGE`: Can manage the query + slug: + type: string + description: >- + URL slug. Usually mirrors the query name with dashes (`-`) instead + of spaces. Appears in the URL for this query. + tags: + type: array + items: + type: string + updated_at: + type: string + description: Timestamp when this dashboard was last updated. + user: + $ref: '#/components/schemas/User' + user_id: + type: integer + description: The ID of the user who owns the dashboard. + widgets: + type: array + items: + $ref: '#/components/schemas/Widget' + description: >- + A JSON representing a dashboard containing widgets of visualizations and + text boxes. + DashboardOptions: + type: object + properties: + moved_to_trash_at: + type: string + DataSource: + type: object + properties: + id: + type: string + description: >- + Data source ID maps to the ID of the data source used by the + resource and is distinct from the warehouse ID. [Learn more] [Learn + more]: https://docs.databricks.com/api/workspace/datasources/list + name: + type: string + description: >- + The string name of this data source / SQL warehouse as it appears in + the Databricks SQL web application. + pause_reason: + type: string + description: Reserved for internal use. + paused: + type: integer + description: Reserved for internal use. + supports_auto_limit: + type: boolean + description: Reserved for internal use. + syntax: + type: string + description: Reserved for internal use. + type: + type: string + description: >- + The type of data source. For SQL warehouses, this will be + `databricks_internal`. + view_only: + type: boolean + description: Reserved for internal use. + warehouse_id: + type: string + description: >- + The ID of the associated SQL warehouse, if this data source is + backed by a SQL warehouse. + description: A JSON object representing a DBSQL data source / SQL warehouse. + DateRange: + type: object + properties: + start: + type: string + end: + type: string + required: + - start + - end + DateRangeValue: + type: object + properties: + date_range_value: + $ref: '#/components/schemas/DateRange' + dynamic_date_range_value: + $ref: '#/components/schemas/DateRangeValueDynamicDateRange' + description: Dynamic date-time range value based on current date-time. + precision: + $ref: '#/components/schemas/DatePrecision' + description: >- + Date-time precision to format the value into when the query is run. + Defaults to DAY_PRECISION (YYYY-MM-DD). + start_day_of_week: + type: integer + DateValue: + type: object + properties: + date_value: + type: string + dynamic_date_value: + $ref: '#/components/schemas/DateValueDynamicDate' + description: Dynamic date-time value based on current date-time. + precision: + $ref: '#/components/schemas/DatePrecision' + description: >- + Date-time precision to format the value into when the query is run. + Defaults to DAY_PRECISION (YYYY-MM-DD). + DefaultWarehouseOverride: + type: object + properties: + type: + $ref: '#/components/schemas/DefaultWarehouseOverrideType' + description: The type of override behavior. + default_warehouse_override_id: + type: string + description: The ID component of the resource name (user ID). + name: + type: string + description: >- + The resource name of the default warehouse override. Format: + default-warehouse-overrides/{default_warehouse_override_id} + warehouse_id: + type: string + description: >- + The specific warehouse ID when type is CUSTOM. Not set for + LAST_SELECTED type. + required: + - type + description: >- + Represents a per-user default warehouse override configuration. This + resource allows users or + administrators to customize how a user's default warehouse is selected for SQL operations. If no + override exists for a user, the workspace default warehouse will be used. + DeleteResponse: + type: object + properties: {} + DeleteWarehouseResponse: + type: object + properties: {} + EditWarehouseResponse: + type: object + properties: {} + Empty: + type: object + properties: {} + description: >- + Represents an empty message, similar to google.protobuf.Empty, which is + not available in the + firm right now. + EndpointConfPair: + type: object + properties: + key: + type: string + value: + type: string + EndpointHealth: + type: object + properties: + details: + type: string + failure_reason: + $ref: '#/components/schemas/TerminationReason' + description: >- + The reason for failure to bring up clusters for this warehouse. This + is available when status is 'FAILED' and sometimes when it is + DEGRADED. + message: + type: string + description: Deprecated. split into summary and details for security + status: + $ref: '#/components/schemas/Status' + description: Health status of the endpoint. + summary: + type: string + description: >- + A short summary of the health status in case of degraded/failed + warehouses. + EndpointInfo: + type: object + properties: + auto_stop_mins: + type: integer + channel: + $ref: '#/components/schemas/Channel' + description: Channel Details + cluster_size: + type: string + description: >- + Size of the clusters allocated for this warehouse. Increasing the + size of a spark cluster allows you to run larger queries on it. If + you want to increase the number of concurrent queries, please tune + max_num_clusters. Supported values: - 2X-Small - X-Small - Small - + Medium - Large - X-Large - 2X-Large - 3X-Large - 4X-Large + creator_name: + type: string + description: warehouse creator name + enable_photon: + type: boolean + description: >- + Configures whether the warehouse should use Photon optimized + clusters. Defaults to false. + enable_serverless_compute: + type: boolean + description: Configures whether the warehouse should use serverless compute + health: + $ref: '#/components/schemas/EndpointHealth' + description: >- + Optional health status. Assume the warehouse is healthy if this + field is not set. + id: + type: string + description: unique identifier for warehouse + instance_profile_arn: + type: string + description: Deprecated. Instance profile used to pass IAM role to the cluster + jdbc_url: + type: string + description: the jdbc connection string for this warehouse + max_num_clusters: + type: integer + description: >- + Maximum number of clusters that the autoscaler will create to handle + concurrent queries. Supported values: - Must be >= min_num_clusters + - Must be <= 40. Defaults to min_clusters if unset. + min_num_clusters: + type: integer + description: >- + Minimum number of available clusters that will be maintained for + this SQL warehouse. Increasing this will ensure that a larger number + of clusters are always running and therefore may reduce the cold + start time for new queries. This is similar to reserved vs. + revocable cores in a resource manager. Supported values: - Must be > + 0 - Must be <= min(max_num_clusters, 30) Defaults to 1 + name: + type: string + description: >- + Logical name for the cluster. Supported values: - Must be unique + within an org. - Must be less than 100 characters. + num_active_sessions: + type: integer + description: Deprecated. current number of active sessions for the warehouse + num_clusters: + type: integer + description: current number of clusters running for the service + odbc_params: + $ref: '#/components/schemas/OdbcParams' + description: ODBC parameters for the SQL warehouse + spot_instance_policy: + $ref: '#/components/schemas/SpotInstancePolicy' + description: Configurations whether the endpoint should use spot instances. + state: + $ref: '#/components/schemas/State' + description: state of the endpoint + tags: + $ref: '#/components/schemas/EndpointTags' + description: >- + A set of key-value pairs that will be tagged on all resources (e.g., + AWS instances and EBS volumes) associated with this SQL warehouse. + Supported values: - Number of tags < 45. + warehouse_type: + $ref: '#/components/schemas/EndpointInfoWarehouseType' + description: >- + Warehouse type: `PRO` or `CLASSIC`. If you want to use serverless + compute, you must set to `PRO` and also set the field + `enable_serverless_compute` to `true`. + EndpointTagPair: + type: object + properties: + key: + type: string + value: + type: string + EndpointTags: + type: object + properties: + custom_tags: + type: array + items: + $ref: '#/components/schemas/EndpointTagPair' + EnumValue: + type: object + properties: + enum_options: + type: string + multi_values_options: + $ref: '#/components/schemas/MultiValuesOptions' + description: >- + If specified, allows multiple values to be selected for this + parameter. + values: + type: array + items: + type: string + description: List of selected query parameter values. + ExternalLink: + type: object + properties: + byte_count: + type: integer + chunk_index: + type: integer + description: The position within the sequence of result set chunks. + expiration: + type: string + description: >- + Indicates the date-time that the given external link will expire and + becomes invalid, after which point a new `external_link` must be + requested. + external_link: + type: string + description: >- + A URL pointing to a chunk of result data, hosted by an external + service, with a short expiration time (<= 15 minutes). As this URL + contains a temporary credential, it should be considered sensitive + and the client should not expose this URL in a log. + http_headers: + type: object + description: >- + HTTP headers that must be included with a GET request to the + `external_link`. Each header is provided as a key-value pair. + Headers are typically used to pass a decryption key to the external + service. The values of these headers should be considered sensitive + and the client should not expose these values in a log. + next_chunk_index: + type: integer + description: >- + When fetching, provides the `chunk_index` for the _next_ chunk. If + absent, indicates there are no more chunks. The next chunk can be + fetched with a :method:statementexecution/getstatementresultchunkn + request. + next_chunk_internal_link: + type: string + description: >- + When fetching, provides a link to fetch the _next_ chunk. If absent, + indicates there are no more chunks. This link is an absolute `path` + to be joined with your `$DATABRICKS_HOST`, and should be treated as + an opaque link. This is an alternative to using `next_chunk_index`. + row_count: + type: integer + description: The number of rows within the result chunk. + row_offset: + type: integer + description: The starting row offset within the result set. + ExternalQuerySource: + type: object + properties: + alert_id: + type: string + dashboard_id: + type: string + description: The canonical identifier for this Lakeview dashboard + genie_space_id: + type: string + description: The canonical identifier for this Genie space + job_info: + $ref: '#/components/schemas/ExternalQuerySourceJobInfo' + legacy_dashboard_id: + type: string + description: The canonical identifier for this legacy dashboard + notebook_id: + type: string + description: The canonical identifier for this notebook + sql_query_id: + type: string + description: The canonical identifier for this SQL query + ExternalQuerySourceJobInfo: + type: object + properties: + job_id: + type: string + job_run_id: + type: string + description: >- + The canonical identifier of the run. This ID is unique across all + runs of all jobs. + job_task_run_id: + type: string + description: The canonical identifier of the task run. + GetResponse: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/AccessControl' + object_id: + type: string + description: >- + An object's type and UUID, separated by a forward slash (/) + character. + object_type: + $ref: '#/components/schemas/ObjectType' + description: A singular noun object type. + GetWarehousePermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/WarehousePermissionsDescription' + GetWarehouseResponse: + type: object + properties: + auto_stop_mins: + type: integer + channel: + $ref: '#/components/schemas/Channel' + description: Channel Details + cluster_size: + type: string + description: >- + Size of the clusters allocated for this warehouse. Increasing the + size of a spark cluster allows you to run larger queries on it. If + you want to increase the number of concurrent queries, please tune + max_num_clusters. Supported values: - 2X-Small - X-Small - Small - + Medium - Large - X-Large - 2X-Large - 3X-Large - 4X-Large + creator_name: + type: string + description: warehouse creator name + enable_photon: + type: boolean + description: >- + Configures whether the warehouse should use Photon optimized + clusters. Defaults to false. + enable_serverless_compute: + type: boolean + description: Configures whether the warehouse should use serverless compute + health: + $ref: '#/components/schemas/EndpointHealth' + description: >- + Optional health status. Assume the warehouse is healthy if this + field is not set. + id: + type: string + description: unique identifier for warehouse + instance_profile_arn: + type: string + description: Deprecated. Instance profile used to pass IAM role to the cluster + jdbc_url: + type: string + description: the jdbc connection string for this warehouse + max_num_clusters: + type: integer + description: >- + Maximum number of clusters that the autoscaler will create to handle + concurrent queries. Supported values: - Must be >= min_num_clusters + - Must be <= 40. Defaults to min_clusters if unset. + min_num_clusters: + type: integer + description: >- + Minimum number of available clusters that will be maintained for + this SQL warehouse. Increasing this will ensure that a larger number + of clusters are always running and therefore may reduce the cold + start time for new queries. This is similar to reserved vs. + revocable cores in a resource manager. Supported values: - Must be > + 0 - Must be <= min(max_num_clusters, 30) Defaults to 1 + name: + type: string + description: >- + Logical name for the cluster. Supported values: - Must be unique + within an org. - Must be less than 100 characters. + num_active_sessions: + type: integer + description: Deprecated. current number of active sessions for the warehouse + num_clusters: + type: integer + description: current number of clusters running for the service + odbc_params: + $ref: '#/components/schemas/OdbcParams' + description: ODBC parameters for the SQL warehouse + spot_instance_policy: + $ref: '#/components/schemas/SpotInstancePolicy' + description: Configurations whether the endpoint should use spot instances. + state: + $ref: '#/components/schemas/State' + description: state of the endpoint + tags: + $ref: '#/components/schemas/EndpointTags' + description: >- + A set of key-value pairs that will be tagged on all resources (e.g., + AWS instances and EBS volumes) associated with this SQL warehouse. + Supported values: - Number of tags < 45. + warehouse_type: + $ref: '#/components/schemas/GetWarehouseResponseWarehouseType' + description: >- + Warehouse type: `PRO` or `CLASSIC`. If you want to use serverless + compute, you must set to `PRO` and also set the field + `enable_serverless_compute` to `true`. + GetWorkspaceWarehouseConfigResponse: + type: object + properties: + channel: + $ref: '#/components/schemas/Channel' + config_param: + $ref: '#/components/schemas/RepeatedEndpointConfPairs' + description: 'Deprecated: Use sql_configuration_parameters' + data_access_config: + type: array + items: + $ref: '#/components/schemas/EndpointConfPair' + description: >- + Spark confs for external hive metastore configuration JSON + serialized size must be less than <= 512K + enable_serverless_compute: + type: boolean + description: Enable Serverless compute for SQL warehouses + enabled_warehouse_types: + type: array + items: + $ref: '#/components/schemas/WarehouseTypePair' + description: >- + List of Warehouse Types allowed in this workspace (limits allowed + value of the type field in CreateWarehouse and EditWarehouse). Note: + Some types cannot be disabled, they don't need to be specified in + SetWorkspaceWarehouseConfig. Note: Disabling a type may cause + existing warehouses to be converted to another type. Used by + frontend to save specific type availability in the warehouse create + and edit form UI. + global_param: + $ref: '#/components/schemas/RepeatedEndpointConfPairs' + description: 'Deprecated: Use sql_configuration_parameters' + google_service_account: + type: string + description: >- + GCP only: Google Service Account used to pass to cluster to access + Google Cloud Storage + instance_profile_arn: + type: string + description: >- + AWS Only: The instance profile used to pass an IAM role to the SQL + warehouses. This configuration is also applied to the workspace's + serverless compute for notebooks and jobs. + security_policy: + $ref: >- + #/components/schemas/GetWorkspaceWarehouseConfigResponseSecurityPolicy + description: Security policy for warehouses + sql_configuration_parameters: + $ref: '#/components/schemas/RepeatedEndpointConfPairs' + description: SQL configuration parameters + LegacyAlert: + type: object + properties: + created_at: + type: string + id: + type: string + description: Alert ID. + last_triggered_at: + type: string + description: Timestamp when the alert was last triggered. + name: + type: string + description: Name of the alert. + options: + $ref: '#/components/schemas/AlertOptions' + description: Alert configuration options. + parent: + type: string + description: The identifier of the workspace folder containing the object. + query: + $ref: '#/components/schemas/AlertQuery' + rearm: + type: integer + description: >- + Number of seconds after being triggered before the alert rearms + itself and can be triggered again. If `null`, alert will never be + triggered again. + state: + $ref: '#/components/schemas/LegacyAlertState' + description: >- + State of the alert. Possible values are: `unknown` (yet to be + evaluated), `triggered` (evaluated and fulfilled trigger + conditions), or `ok` (evaluated and did not fulfill trigger + conditions). + updated_at: + type: string + description: Timestamp when the alert was last updated. + user: + $ref: '#/components/schemas/User' + LegacyQuery: + type: object + properties: + can_edit: + type: boolean + created_at: + type: string + description: The timestamp when this query was created. + data_source_id: + type: string + description: >- + Data source ID maps to the ID of the data source used by the + resource and is distinct from the warehouse ID. [Learn more] [Learn + more]: https://docs.databricks.com/api/workspace/datasources/list + description: + type: string + description: >- + General description that conveys additional information about this + query such as usage notes. + id: + type: string + description: Query ID. + is_archived: + type: boolean + description: >- + Indicates whether the query is trashed. Trashed queries can't be + used in dashboards, or appear in search results. If this boolean is + `true`, the `options` property for this query includes a + `moved_to_trash_at` timestamp. Trashed queries are permanently + deleted after 30 days. + is_draft: + type: boolean + description: >- + Whether the query is a draft. Draft queries only appear in list + views for their owners. Visualizations from draft queries cannot + appear on dashboards. + is_favorite: + type: boolean + description: >- + Whether this query object appears in the current user's favorites + list. This flag determines whether the star icon for favorites is + selected. + is_safe: + type: boolean + description: >- + Text parameter types are not safe from SQL injection for all types + of data source. Set this Boolean parameter to `true` if a query + either does not use any text type parameters or uses a data source + type where text type parameters are handled safely. + last_modified_by: + $ref: '#/components/schemas/User' + last_modified_by_id: + type: integer + description: The ID of the user who last saved changes to this query. + latest_query_data_id: + type: string + description: >- + If there is a cached result for this query and user, this field + includes the query result ID. If this query uses parameters, this + field is always null. + name: + type: string + description: >- + The title of this query that appears in list views, widget headings, + and on the query page. + options: + $ref: '#/components/schemas/QueryOptions' + parent: + type: string + description: The identifier of the workspace folder containing the object. + permission_tier: + $ref: '#/components/schemas/PermissionLevel' + description: >- + * `CAN_VIEW`: Can view the query * `CAN_RUN`: Can run the query * + `CAN_EDIT`: Can edit the query * `CAN_MANAGE`: Can manage the query + query: + type: string + description: The text of the query to be run. + query_hash: + type: string + description: >- + A SHA-256 hash of the query text along with the authenticated user + ID. + run_as_role: + $ref: '#/components/schemas/RunAsRole' + description: >- + Sets the **Run as** role for the object. Must be set to one of + `"viewer"` (signifying "run as viewer" behavior) or `"owner"` + (signifying "run as owner" behavior) + tags: + type: array + items: + type: string + updated_at: + type: string + description: The timestamp at which this query was last updated. + user: + $ref: '#/components/schemas/User' + user_id: + type: integer + description: The ID of the user who owns the query. + visualizations: + type: array + items: + $ref: '#/components/schemas/LegacyVisualization' + LegacyVisualization: + type: object + properties: + created_at: + type: string + description: + type: string + description: >- + A short description of this visualization. This is not displayed in + the UI. + id: + type: string + description: The UUID for this visualization. + name: + type: string + description: >- + The name of the visualization that appears on dashboards and the + query screen. + options: + type: object + description: >- + The options object varies widely from one visualization type to the + next and is unsupported. Databricks does not recommend modifying + visualization settings in JSON. + query: + $ref: '#/components/schemas/LegacyQuery' + type: + type: string + description: 'The type of visualization: chart, table, pivot table, and so on.' + updated_at: + type: string + description: >- + The visualization description API changes frequently and is unsupported. + You can duplicate a + visualization by copying description objects received _from the API_ and then using them to + create a new one with a POST request to the same endpoint. Databricks does not recommend + constructing ad-hoc visualizations entirely in JSON. + ListAlertsResponse: + type: object + properties: + next_page_token: + type: string + results: + type: array + items: + $ref: '#/components/schemas/ListAlertsResponseAlert' + ListAlertsResponseAlert: + type: object + properties: + condition: + $ref: '#/components/schemas/AlertCondition' + create_time: + type: string + description: The timestamp indicating when the alert was created. + custom_body: + type: string + description: >- + Custom body of alert notification, if it exists. See [here] for + custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + custom_subject: + type: string + description: >- + Custom subject of alert notification, if it exists. This can include + email subject entries and Slack notification headers, for example. + See [here] for custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + display_name: + type: string + description: The display name of the alert. + id: + type: string + description: UUID identifying the alert. + lifecycle_state: + $ref: '#/components/schemas/LifecycleState' + description: The workspace state of the alert. Used for tracking trashed status. + notify_on_ok: + type: boolean + description: >- + Whether to notify alert subscribers when alert returns back to + normal. + owner_user_name: + type: string + description: >- + The owner's username. This field is set to "Unavailable" if the user + has been deleted. + query_id: + type: string + description: UUID of the query attached to the alert. + seconds_to_retrigger: + type: integer + description: >- + Number of seconds an alert must wait after being triggered to rearm + itself. After rearming, it can be triggered again. If 0 or not + specified, the alert will not be triggered again. + state: + $ref: '#/components/schemas/AlertState' + description: >- + Current state of the alert's trigger status. This field is set to + UNKNOWN if the alert has not yet been evaluated or ran into an error + during the last evaluation. + trigger_time: + type: string + description: >- + Timestamp when the alert was last triggered, if the alert has been + triggered before. + update_time: + type: string + description: The timestamp indicating when the alert was updated. + ListAlertsV2Response: + type: object + properties: + alerts: + type: array + items: + $ref: '#/components/schemas/AlertV2' + next_page_token: + type: string + ListDefaultWarehouseOverridesResponse: + type: object + properties: + default_warehouse_overrides: + type: array + items: + $ref: '#/components/schemas/DefaultWarehouseOverride' + description: The default warehouse overrides in the workspace. + next_page_token: + type: string + description: >- + A token, which can be sent as `page_token` to retrieve the next + page. If this field is omitted, there are no subsequent pages. + description: Response message for ListDefaultWarehouseOverrides. + ListQueriesResponse: + type: object + properties: + has_next_page: + type: boolean + next_page_token: + type: string + description: A token that can be used to get the next page of results. + res: + type: array + items: + $ref: '#/components/schemas/QueryInfo' + ListQueryObjectsResponse: + type: object + properties: + next_page_token: + type: string + results: + type: array + items: + $ref: '#/components/schemas/ListQueryObjectsResponseQuery' + ListQueryObjectsResponseQuery: + type: object + properties: + apply_auto_limit: + type: boolean + catalog: + type: string + description: Name of the catalog where this query will be executed. + create_time: + type: string + description: Timestamp when this query was created. + description: + type: string + description: >- + General description that conveys additional information about this + query such as usage notes. + display_name: + type: string + description: >- + Display name of the query that appears in list views, widget + headings, and on the query page. + id: + type: string + description: UUID identifying the query. + last_modifier_user_name: + type: string + description: Username of the user who last saved changes to this query. + lifecycle_state: + $ref: '#/components/schemas/LifecycleState' + description: Indicates whether the query is trashed. + owner_user_name: + type: string + description: Username of the user that owns the query. + parameters: + type: array + items: + $ref: '#/components/schemas/QueryParameter' + description: List of query parameter definitions. + query_text: + type: string + description: Text of the query to be run. + run_as_mode: + $ref: '#/components/schemas/RunAsMode' + description: Sets the "Run as" role for the object. + schema: + type: string + description: Name of the schema where this query will be executed. + tags: + type: array + items: + type: string + update_time: + type: string + description: Timestamp when this query was last updated. + warehouse_id: + type: string + description: ID of the SQL warehouse attached to the query. + ListResponse: + type: object + properties: + count: + type: integer + page: + type: integer + description: The current page being displayed. + page_size: + type: integer + description: The number of dashboards per page. + results: + type: array + items: + $ref: '#/components/schemas/Dashboard' + description: List of dashboards returned. + ListVisualizationsForQueryResponse: + type: object + properties: + next_page_token: + type: string + results: + type: array + items: + $ref: '#/components/schemas/Visualization' + ListWarehousesResponse: + type: object + properties: + next_page_token: + type: string + warehouses: + type: array + items: + $ref: '#/components/schemas/EndpointInfo' + description: A list of warehouses and their configurations. + MultiValuesOptions: + type: object + properties: + prefix: + type: string + separator: + type: string + description: >- + Character that separates each selected parameter value. Defaults to + a comma. + suffix: + type: string + description: Character that suffixes each selected parameter value. + NumericValue: + type: object + properties: + value: + type: number + OdbcParams: + type: object + properties: + hostname: + type: string + path: + type: string + port: + type: integer + protocol: + type: string + Parameter: + type: object + properties: + enumOptions: + type: string + multiValuesOptions: + $ref: '#/components/schemas/MultiValuesOptions' + description: >- + If specified, allows multiple values to be selected for this + parameter. Only applies to dropdown list and query-based dropdown + list parameters. + name: + type: string + description: >- + The literal parameter marker that appears between double curly + braces in the query text. + queryId: + type: string + description: >- + The UUID of the query that provides the parameter values. Only + applies for query-based dropdown list parameters. + title: + type: string + description: The text displayed in a parameter picking widget. + type: + $ref: '#/components/schemas/ParameterType' + description: Parameters can have several different types. + value: + type: object + description: The default value for this parameter. + Query: + type: object + properties: + apply_auto_limit: + type: boolean + catalog: + type: string + description: Name of the catalog where this query will be executed. + create_time: + type: string + description: Timestamp when this query was created. + description: + type: string + description: >- + General description that conveys additional information about this + query such as usage notes. + display_name: + type: string + description: >- + Display name of the query that appears in list views, widget + headings, and on the query page. + id: + type: string + description: UUID identifying the query. + last_modifier_user_name: + type: string + description: Username of the user who last saved changes to this query. + lifecycle_state: + $ref: '#/components/schemas/LifecycleState' + description: Indicates whether the query is trashed. + owner_user_name: + type: string + description: Username of the user that owns the query. + parameters: + type: array + items: + $ref: '#/components/schemas/QueryParameter' + description: List of query parameter definitions. + parent_path: + type: string + description: Workspace path of the workspace folder containing the object. + query_text: + type: string + description: Text of the query to be run. + run_as_mode: + $ref: '#/components/schemas/RunAsMode' + description: Sets the "Run as" role for the object. + schema: + type: string + description: Name of the schema where this query will be executed. + tags: + type: array + items: + type: string + update_time: + type: string + description: Timestamp when this query was last updated. + warehouse_id: + type: string + description: ID of the SQL warehouse attached to the query. + QueryBackedValue: + type: object + properties: + multi_values_options: + $ref: '#/components/schemas/MultiValuesOptions' + query_id: + type: string + description: UUID of the query that provides the parameter values. + values: + type: array + items: + type: string + description: List of selected query parameter values. + QueryFilter: + type: object + properties: + query_start_time_range: + $ref: '#/components/schemas/TimeRange' + statement_ids: + type: array + items: + type: string + description: A list of statement IDs. + statuses: + type: array + items: + $ref: '#/components/schemas/QueryStatus' + description: >- + A list of statuses (QUEUED, RUNNING, CANCELED, FAILED, FINISHED) to + match query results. Corresponds to the `status` field in the + response. Filtering for multiple statuses is not recommended. + Instead, opt to filter by a single status multiple times and then + combine the results. + user_ids: + type: array + items: + type: integer + description: A list of user IDs who ran the queries. + warehouse_ids: + type: array + items: + type: string + description: A list of warehouse IDs. + QueryInfo: + type: object + properties: + cache_query_id: + type: string + channel_used: + $ref: '#/components/schemas/ChannelInfo' + description: SQL Warehouse channel information at the time of query execution + client_application: + type: string + description: >- + Client application that ran the statement. For example: Databricks + SQL Editor, Tableau, and Power BI. This field is derived from + information provided by client applications. While values are + expected to remain static over time, this cannot be guaranteed. + duration: + type: integer + description: >- + Total time of the statement execution. This value does not include + the time taken to retrieve the results, which can result in a + discrepancy between this value and the start-to-finish wall-clock + time. + endpoint_id: + type: string + description: Alias for `warehouse_id`. + error_message: + type: string + description: Message describing why the query could not complete. + executed_as_user_id: + type: integer + description: The ID of the user whose credentials were used to run the query. + executed_as_user_name: + type: string + description: >- + The email address or username of the user whose credentials were + used to run the query. + execution_end_time_ms: + type: integer + description: The time execution of the query ended. + is_final: + type: boolean + description: Whether more updates for the query are expected. + lookup_key: + type: string + description: A key that can be used to look up query details. + metrics: + $ref: '#/components/schemas/QueryMetrics' + description: Metrics about query execution. + plans_state: + $ref: '#/components/schemas/PlansState' + description: >- + Whether plans exist for the execution, or the reason why they are + missing + query_end_time_ms: + type: integer + description: The time the query ended. + query_id: + type: string + description: The query ID. + query_source: + $ref: '#/components/schemas/ExternalQuerySource' + description: >- + A struct that contains key-value pairs representing Databricks + entities that were involved in the execution of this statement, such + as jobs, notebooks, or dashboards. This field only records + Databricks entities. + query_start_time_ms: + type: integer + description: The time the query started. + query_tags: + type: array + items: + $ref: '#/components/schemas/QueryTag' + description: A query execution can be optionally annotated with query tags + query_text: + type: string + description: The text of the query. + rows_produced: + type: integer + description: The number of results returned by the query. + session_id: + type: string + description: >- + The spark session UUID that query ran on. This is either the Spark + Connect, DBSQL, or SDP session ID. + spark_ui_url: + type: string + description: URL to the Spark UI query plan. + statement_type: + $ref: '#/components/schemas/QueryStatementType' + description: Type of statement for this query + status: + $ref: '#/components/schemas/QueryStatus' + description: >- + Query status with one the following values: - `QUEUED`: Query has + been received and queued. - `RUNNING`: Query has started. - + `CANCELED`: Query has been cancelled by the user. - `FAILED`: Query + has failed. - `FINISHED`: Query has completed. + user_id: + type: integer + description: The ID of the user who ran the query. + user_name: + type: string + description: The email address or username of the user who ran the query. + warehouse_id: + type: string + description: Warehouse ID. + QueryList: + type: object + properties: + count: + type: integer + page: + type: integer + description: The page number that is currently displayed. + page_size: + type: integer + description: The number of queries per page. + results: + type: array + items: + $ref: '#/components/schemas/LegacyQuery' + description: List of queries returned. + QueryMetrics: + type: object + properties: + compilation_time_ms: + type: integer + description: >- + Time spent loading metadata and optimizing the query, in + milliseconds. + execution_time_ms: + type: integer + description: Time spent executing the query, in milliseconds. + network_sent_bytes: + type: integer + description: >- + Total amount of data sent over the network between executor nodes + during shuffle, in bytes. + overloading_queue_start_timestamp: + type: integer + description: >- + Timestamp of when the query was enqueued waiting while the warehouse + was at max load. This field is optional and will not appear if the + query skipped the overloading queue. + photon_total_time_ms: + type: integer + description: >- + Total execution time for all individual Photon query engine tasks in + the query, in milliseconds. + projected_remaining_task_total_time_ms: + type: integer + description: >- + projected remaining work to be done aggregated across all stages in + the query, in milliseconds + projected_remaining_wallclock_time_ms: + type: integer + description: >- + projected lower bound on remaining total task time based on + projected_remaining_task_total_time_ms / maximum concurrency + provisioning_queue_start_timestamp: + type: integer + description: >- + Timestamp of when the query was enqueued waiting for a cluster to be + provisioned for the warehouse. This field is optional and will not + appear if the query skipped the provisioning queue. + pruned_bytes: + type: integer + description: Total number of file bytes in all tables not read due to pruning + pruned_files_count: + type: integer + description: Total number of files from all tables not read due to pruning + query_compilation_start_timestamp: + type: integer + description: >- + Timestamp of when the underlying compute started compilation of the + query. + read_bytes: + type: integer + description: Total size of data read by the query, in bytes. + read_cache_bytes: + type: integer + description: Size of persistent data read from the cache, in bytes. + read_files_bytes: + type: integer + description: Total number of file bytes in all tables read + read_files_count: + type: integer + description: Number of files read after pruning + read_partitions_count: + type: integer + description: Number of partitions read after pruning. + read_remote_bytes: + type: integer + description: >- + Size of persistent data read from cloud object storage on your cloud + tenant, in bytes. + remaining_task_count: + type: integer + description: >- + number of remaining tasks to complete this is based on the current + status and could be bigger or smaller in the future based on future + updates + result_fetch_time_ms: + type: integer + description: >- + Time spent fetching the query results after the execution finished, + in milliseconds. + result_from_cache: + type: boolean + description: >- + `true` if the query result was fetched from cache, `false` + otherwise. + rows_produced_count: + type: integer + description: Total number of rows returned by the query. + rows_read_count: + type: integer + description: Total number of rows read by the query. + runnable_tasks: + type: integer + description: >- + number of remaining tasks to complete, calculated by autoscaler + StatementAnalysis.scala deprecated: use remaining_task_count instead + spill_to_disk_bytes: + type: integer + description: >- + Size of data temporarily written to disk while executing the query, + in bytes. + task_time_over_time_range: + $ref: '#/components/schemas/TaskTimeOverRange' + description: >- + sum of task times completed in a range of wall clock time, + approximated to a configurable number of points aggregated over all + stages and jobs in the query (based on task_total_time_ms) + task_total_time_ms: + type: integer + description: Sum of execution time for all of the query’s tasks, in milliseconds. + total_time_ms: + type: integer + description: >- + Total execution time of the query from the client’s point of view, + in milliseconds. + work_to_be_done: + type: integer + description: >- + remaining work to be done across all stages in the query, calculated + by autoscaler StatementAnalysis.scala, in milliseconds deprecated: + using projected_remaining_task_total_time_ms instead + write_remote_bytes: + type: integer + description: >- + Size pf persistent data written to cloud object storage in your + cloud tenant, in bytes. + description: >- + A query metric that encapsulates a set of measurements for a single + query. Metrics come from the + driver and are stored in the history service database. + QueryOptions: + type: object + properties: + catalog: + type: string + moved_to_trash_at: + type: string + description: >- + The timestamp when this query was moved to trash. Only present when + the `is_archived` property is `true`. Trashed items are deleted + after thirty days. + parameters: + type: array + items: + $ref: '#/components/schemas/Parameter' + schema: + type: string + description: The name of the schema to execute this query in. + QueryParameter: + type: object + properties: + date_range_value: + $ref: '#/components/schemas/DateRangeValue' + date_value: + $ref: '#/components/schemas/DateValue' + description: >- + Date query parameter value. Can only specify one of + `dynamic_date_value` or `date_value`. + enum_value: + $ref: '#/components/schemas/EnumValue' + description: Dropdown query parameter value. + name: + type: string + description: >- + Literal parameter marker that appears between double curly braces in + the query text. + numeric_value: + $ref: '#/components/schemas/NumericValue' + description: Numeric query parameter value. + query_backed_value: + $ref: '#/components/schemas/QueryBackedValue' + description: Query-based dropdown query parameter value. + text_value: + $ref: '#/components/schemas/TextValue' + description: Text query parameter value. + title: + type: string + description: Text displayed in the user-facing parameter widget in the UI. + QueryTag: + type: object + properties: + key: + type: string + value: + type: string + required: + - key + description: >- + * A query execution can be annotated with an optional key-value pair to + allow users to attribute + the executions by key and optional value to filter by. QueryTag is the user-facing + representation. + RepeatedEndpointConfPairs: + type: object + properties: + config_pair: + type: array + items: + $ref: '#/components/schemas/EndpointConfPair' + configuration_pairs: + type: array + items: + $ref: '#/components/schemas/EndpointConfPair' + RestoreResponse: + type: object + properties: {} + ResultData: + type: object + properties: + byte_count: + type: integer + description: >- + The number of bytes in the result chunk. This field is not available + when using `INLINE` disposition. + chunk_index: + type: integer + description: The position within the sequence of result set chunks. + data_array: + type: array + items: + type: array + items: + type: string + description: >- + The `JSON_ARRAY` format is an array of arrays of values, where each + non-null value is formatted as a string. Null values are encoded as + JSON `null`. + external_links: + type: array + items: + $ref: '#/components/schemas/ExternalLink' + next_chunk_index: + type: integer + description: >- + When fetching, provides the `chunk_index` for the _next_ chunk. If + absent, indicates there are no more chunks. The next chunk can be + fetched with a :method:statementexecution/getstatementresultchunkn + request. + next_chunk_internal_link: + type: string + description: >- + When fetching, provides a link to fetch the _next_ chunk. If absent, + indicates there are no more chunks. This link is an absolute `path` + to be joined with your `$DATABRICKS_HOST`, and should be treated as + an opaque link. This is an alternative to using `next_chunk_index`. + row_count: + type: integer + description: The number of rows within the result chunk. + row_offset: + type: integer + description: The starting row offset within the result set. + description: >- + Contains the result data of a single chunk when using `INLINE` + disposition. When using + `EXTERNAL_LINKS` disposition, the array `external_links` is used instead to provide URLs to the + result data in cloud storage. Exactly one of these alternatives is used. (While the + `external_links` array prepares the API to return multiple links in a single response. Currently + only a single link is returned.) + ResultManifest: + type: object + properties: + chunks: + type: array + items: + $ref: '#/components/schemas/BaseChunkInfo' + description: Array of result set chunk metadata. + format: + $ref: '#/components/schemas/Format' + schema: + $ref: '#/components/schemas/ResultSchema' + total_byte_count: + type: integer + description: >- + The total number of bytes in the result set. This field is not + available when using `INLINE` disposition. + total_chunk_count: + type: integer + description: >- + The total number of chunks that the result set has been divided + into. + total_row_count: + type: integer + description: The total number of rows in the result set. + truncated: + type: boolean + description: >- + Indicates whether the result is truncated due to `row_limit` or + `byte_limit`. + description: The result manifest provides schema and metadata for the result set. + ResultSchema: + type: object + properties: + column_count: + type: integer + columns: + type: array + items: + $ref: '#/components/schemas/ColumnInfo' + description: The schema is an ordered list of column descriptions. + ServiceError: + type: object + properties: + error_code: + $ref: '#/components/schemas/ServiceErrorCode' + message: + type: string + description: A brief summary of the error condition. + SetResponse: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/AccessControl' + object_id: + type: string + description: >- + An object's type and UUID, separated by a forward slash (/) + character. + object_type: + $ref: '#/components/schemas/ObjectType' + description: A singular noun object type. + SetWorkspaceWarehouseConfigResponse: + type: object + properties: {} + StartWarehouseResponse: + type: object + properties: {} + StatementParameterListItem: + type: object + properties: + name: + type: string + type: + type: string + description: >- + The data type, given as a string. For example: `INT`, `STRING`, + `DECIMAL(10,2)`. If no type is given the type is assumed to be + `STRING`. Complex types, such as `ARRAY`, `MAP`, and `STRUCT` are + not supported. For valid types, refer to the section [Data types] of + the SQL language reference. [Data types]: + https://docs.databricks.com/sql/language-manual/functions/cast.html + value: + type: string + description: >- + The value to substitute, represented as a string. If omitted, the + value is interpreted as NULL. + required: + - name + StatementResponse: + type: object + properties: + manifest: + $ref: '#/components/schemas/ResultManifest' + result: + $ref: '#/components/schemas/ResultData' + statement_id: + type: string + description: >- + The statement ID is returned upon successfully submitting a SQL + statement, and is a required reference for all subsequent calls. + status: + $ref: '#/components/schemas/StatementStatus' + StatementStatus: + type: object + properties: + error: + $ref: '#/components/schemas/ServiceError' + state: + $ref: '#/components/schemas/StatementState' + description: >- + Statement execution state: - `PENDING`: waiting for warehouse - + `RUNNING`: running - `SUCCEEDED`: execution was successful, result + data available for fetch - `FAILED`: execution failed; reason for + failure described in accompanying error message - `CANCELED`: user + canceled; can come from explicit cancel call, or timeout with + `on_wait_timeout=CANCEL` - `CLOSED`: execution successful, and + statement closed; result no longer available for fetch + description: >- + The status response includes execution state and if relevant, error + information. + StopWarehouseResponse: + type: object + properties: {} + Success: + type: object + properties: + message: + $ref: '#/components/schemas/SuccessMessage' + TaskTimeOverRange: + type: object + properties: + entries: + type: array + items: + $ref: '#/components/schemas/TaskTimeOverRangeEntry' + interval: + type: integer + description: >- + interval length for all entries (difference in start time and end + time of an entry range) the same for all entries start time of first + interval is query_start_time_ms + TaskTimeOverRangeEntry: + type: object + properties: + task_completed_time_ms: + type: integer + TerminationReason: + type: object + properties: + code: + $ref: '#/components/schemas/TerminationReasonCode' + parameters: + type: object + description: >- + list of parameters that provide additional information about why the + cluster was terminated + type: + $ref: '#/components/schemas/TerminationReasonType' + description: type of the termination + TextValue: + type: object + properties: + value: + type: string + TimeRange: + type: object + properties: + end_time_ms: + type: integer + start_time_ms: + type: integer + description: The start time in milliseconds. + TransferOwnershipObjectId: + type: object + properties: + new_owner: + type: string + UpdateAlertRequestAlert: + type: object + properties: + condition: + $ref: '#/components/schemas/AlertCondition' + custom_body: + type: string + description: >- + Custom body of alert notification, if it exists. See [here] for + custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + custom_subject: + type: string + description: >- + Custom subject of alert notification, if it exists. This can include + email subject entries and Slack notification headers, for example. + See [here] for custom templating instructions. [here]: + https://docs.databricks.com/sql/user/alerts/index.html + display_name: + type: string + description: The display name of the alert. + notify_on_ok: + type: boolean + description: >- + Whether to notify alert subscribers when alert returns back to + normal. + owner_user_name: + type: string + description: >- + The owner's username. This field is set to "Unavailable" if the user + has been deleted. + query_id: + type: string + description: UUID of the query attached to the alert. + seconds_to_retrigger: + type: integer + description: >- + Number of seconds an alert must wait after being triggered to rearm + itself. After rearming, it can be triggered again. If 0 or not + specified, the alert will not be triggered again. + UpdateQueryRequestQuery: + type: object + properties: + apply_auto_limit: + type: boolean + catalog: + type: string + description: Name of the catalog where this query will be executed. + description: + type: string + description: >- + General description that conveys additional information about this + query such as usage notes. + display_name: + type: string + description: >- + Display name of the query that appears in list views, widget + headings, and on the query page. + owner_user_name: + type: string + description: Username of the user that owns the query. + parameters: + type: array + items: + $ref: '#/components/schemas/QueryParameter' + description: List of query parameter definitions. + query_text: + type: string + description: Text of the query to be run. + run_as_mode: + $ref: '#/components/schemas/RunAsMode' + description: Sets the "Run as" role for the object. + schema: + type: string + description: Name of the schema where this query will be executed. + tags: + type: array + items: + type: string + warehouse_id: + type: string + description: ID of the SQL warehouse attached to the query. + UpdateResponse: + type: object + properties: {} + UpdateVisualizationRequestVisualization: + type: object + properties: + display_name: + type: string + serialized_options: + type: string + description: >- + The visualization options varies widely from one visualization type + to the next and is unsupported. Databricks does not recommend + modifying visualization options directly. + serialized_query_plan: + type: string + description: >- + The visualization query plan varies widely from one visualization + type to the next and is unsupported. Databricks does not recommend + modifying the visualization query plan directly. + type: + type: string + description: 'The type of visualization: counter, table, funnel, and so on.' + User: + type: object + properties: + email: + type: string + id: + type: integer + name: + type: string + Visualization: + type: object + properties: + create_time: + type: string + display_name: + type: string + description: The display name of the visualization. + id: + type: string + description: UUID identifying the visualization. + query_id: + type: string + description: UUID of the query that the visualization is attached to. + serialized_options: + type: string + description: >- + The visualization options varies widely from one visualization type + to the next and is unsupported. Databricks does not recommend + modifying visualization options directly. + serialized_query_plan: + type: string + description: >- + The visualization query plan varies widely from one visualization + type to the next and is unsupported. Databricks does not recommend + modifying the visualization query plan directly. + type: + type: string + description: 'The type of visualization: counter, table, funnel, and so on.' + update_time: + type: string + description: The timestamp indicating when the visualization was updated. + WarehouseAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/WarehousePermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + WarehouseAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/WarehousePermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + WarehousePermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/WarehousePermissionLevel' + WarehousePermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/WarehouseAccessControlResponse' + object_id: + type: string + object_type: + type: string + WarehousePermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/WarehousePermissionLevel' + WarehouseTypePair: + type: object + properties: + enabled: + type: boolean + description: >- + If set to false the specific warehouse type will not be be allowed + as a value for warehouse_type in CreateWarehouse and EditWarehouse + warehouse_type: + $ref: '#/components/schemas/WarehouseTypePairWarehouseType' + description: >- + * Configuration values to enable or disable the access to specific + warehouse types in the + workspace. + Widget: + type: object + properties: + id: + type: string + options: + $ref: '#/components/schemas/WidgetOptions' + visualization: + $ref: '#/components/schemas/LegacyVisualization' + description: >- + The visualization description API changes frequently and is + unsupported. You can duplicate a visualization by copying + description objects received _from the API_ and then using them to + create a new one with a POST request to the same endpoint. + Databricks does not recommend constructing ad-hoc visualizations + entirely in JSON. + width: + type: integer + description: Unused field. + WidgetOptions: + type: object + properties: + created_at: + type: string + description: + type: string + description: Custom description of the widget + isHidden: + type: boolean + description: Whether this widget is hidden on the dashboard. + parameterMappings: + type: object + description: >- + How parameters used by the visualization in this widget relate to + other widgets on the dashboard. Databricks does not recommend + modifying this definition in JSON. + position: + $ref: '#/components/schemas/WidgetPosition' + description: >- + Coordinates of this widget on a dashboard. This portion of the API + changes frequently and is unsupported. + title: + type: string + description: Custom title of the widget + updated_at: + type: string + description: Timestamp of the last time this object was updated. + WidgetPosition: + type: object + properties: + autoHeight: + type: boolean + description: reserved for internal use + col: + type: integer + description: column in the dashboard grid. Values start with 0 + row: + type: integer + description: row in the dashboard grid. Values start with 0 + sizeX: + type: integer + description: width of the widget measured in dashboard grid cells + sizeY: + type: integer + description: height of the widget measured in dashboard grid cells + description: >- + Coordinates of this widget on a dashboard. This portion of the API + changes frequently and is + unsupported. + Aggregation: + type: string + enum: + - AVG + - COUNT + - COUNT_DISTINCT + - MAX + - MEDIAN + - MIN + - STDDEV + - SUM + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AlertEvaluationState: + type: string + enum: + - ERROR + - OK + - TRIGGERED + - UNKNOWN + description: >- + UNSPECIFIED - default unspecify value for proto enum, do not use it in + the code UNKNOWN - alert + + not yet evaluated TRIGGERED - alert is triggered OK - alert is not + triggered ERROR - alert + + evaluation failed + AlertLifecycleState: + type: string + enum: + - ACTIVE + - DELETED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AlertOperator: + type: string + enum: + - EQUAL + - GREATER_THAN + - GREATER_THAN_OR_EQUAL + - IS_NULL + - LESS_THAN + - LESS_THAN_OR_EQUAL + - NOT_EQUAL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + AlertOptionsEmptyResultState: + type: string + enum: + - ok + - triggered + - unknown + description: State that alert evaluates to when query result is empty. + AlertState: + type: string + enum: + - OK + - TRIGGERED + - UNKNOWN + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ChannelName: + type: string + enum: + - CHANNEL_NAME_CURRENT + - CHANNEL_NAME_CUSTOM + - CHANNEL_NAME_PREVIEW + - CHANNEL_NAME_PREVIOUS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ColumnInfoTypeName: + type: string + enum: + - ARRAY + - BINARY + - BOOLEAN + - BYTE + - CHAR + - DATE + - DECIMAL + - DOUBLE + - FLOAT + - INT + - INTERVAL + - LONG + - MAP + - 'NULL' + - SHORT + - STRING + - STRUCT + - TIMESTAMP + - USER_DEFINED_TYPE + description: >- + The name of the base data type. This doesn't include details for complex + types such as STRUCT, + + MAP or ARRAY. + ComparisonOperator: + type: string + enum: + - EQUAL + - GREATER_THAN + - GREATER_THAN_OR_EQUAL + - IS_NOT_NULL + - IS_NULL + - LESS_THAN + - LESS_THAN_OR_EQUAL + - NOT_EQUAL + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + CreateWarehouseRequestWarehouseType: + type: string + enum: + - CLASSIC + - PRO + - TYPE_UNSPECIFIED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DatePrecision: + type: string + enum: + - DAY_PRECISION + - MINUTE_PRECISION + - SECOND_PRECISION + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DateRangeValueDynamicDateRange: + type: string + enum: + - LAST_12_MONTHS + - LAST_14_DAYS + - LAST_24_HOURS + - LAST_30_DAYS + - LAST_60_DAYS + - LAST_7_DAYS + - LAST_8_HOURS + - LAST_90_DAYS + - LAST_HOUR + - LAST_MONTH + - LAST_WEEK + - LAST_YEAR + - THIS_MONTH + - THIS_WEEK + - THIS_YEAR + - TODAY + - YESTERDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DateValueDynamicDate: + type: string + enum: + - NOW + - YESTERDAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + DefaultWarehouseOverrideType: + type: string + enum: + - CUSTOM + - LAST_SELECTED + description: Type of default warehouse override behavior. + Disposition: + type: string + enum: + - EXTERNAL_LINKS + - INLINE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EditWarehouseRequestWarehouseType: + type: string + enum: + - CLASSIC + - PRO + - TYPE_UNSPECIFIED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EndpointInfoWarehouseType: + type: string + enum: + - CLASSIC + - PRO + - TYPE_UNSPECIFIED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ExecuteStatementRequestOnWaitTimeout: + type: string + enum: + - CANCEL + - CONTINUE + description: >- + When `wait_timeout > 0s`, the call will block up to the specified time. + If the statement + + execution doesn't finish within this time, `on_wait_timeout` determines + whether the execution + + should continue or be canceled. When set to `CONTINUE`, the statement + execution continues + + asynchronously and the call returns a statement ID which can be used for + polling with + + :method:statementexecution/getStatement. When set to `CANCEL`, the + statement execution is + + canceled and the call returns with a `CANCELED` state. + Format: + type: string + enum: + - ARROW_STREAM + - CSV + - JSON_ARRAY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + GetWarehouseResponseWarehouseType: + type: string + enum: + - CLASSIC + - PRO + - TYPE_UNSPECIFIED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + GetWorkspaceWarehouseConfigResponseSecurityPolicy: + type: string + enum: + - DATA_ACCESS_CONTROL + - NONE + - PASSTHROUGH + description: Security policy to be used for warehouses + LegacyAlertState: + type: string + enum: + - ok + - triggered + - unknown + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + LifecycleState: + type: string + enum: + - ACTIVE + - TRASHED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ListOrder: + type: string + enum: + - created_at + - name + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ObjectType: + type: string + enum: + - alert + - dashboard + - data_source + - query + description: A singular noun object type. + ObjectTypePlural: + type: string + enum: + - alerts + - dashboards + - data_sources + - queries + description: Always a plural of the object type. + OwnableObjectType: + type: string + enum: + - alert + - dashboard + - query + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ParameterType: + type: string + enum: + - datetime + - enum + - number + - query + - text + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + PermissionLevel: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_RUN + - CAN_VIEW + description: >- + * `CAN_VIEW`: Can view the query * `CAN_RUN`: Can run the query * + `CAN_EDIT`: Can edit the query + + * `CAN_MANAGE`: Can manage the query + PlansState: + type: string + enum: + - EMPTY + - EXISTS + - IGNORED_LARGE_PLANS_SIZE + - IGNORED_SMALL_DURATION + - IGNORED_SPARK_PLAN_TYPE + - UNKNOWN + description: Possible Reasons for which we have not saved plans in the database + QueryStatementType: + type: string + enum: + - ALTER + - ANALYZE + - COPY + - CREATE + - DELETE + - DESCRIBE + - DROP + - EXPLAIN + - GRANT + - INSERT + - MERGE + - OPTIMIZE + - OTHER + - REFRESH + - REPLACE + - REVOKE + - SELECT + - SET + - SHOW + - TRUNCATE + - UPDATE + - USE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + QueryStatus: + type: string + enum: + - CANCELED + - COMPILED + - COMPILING + - FAILED + - FINISHED + - QUEUED + - RUNNING + - STARTED + description: Statuses which are also used by OperationStatus in runtime + RunAsMode: + type: string + enum: + - OWNER + - VIEWER + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + RunAsRole: + type: string + enum: + - owner + - viewer + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SchedulePauseStatus: + type: string + enum: + - PAUSED + - UNPAUSED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + ServiceErrorCode: + type: string + enum: + - ABORTED + - ALREADY_EXISTS + - BAD_REQUEST + - CANCELLED + - DEADLINE_EXCEEDED + - INTERNAL_ERROR + - IO_ERROR + - NOT_FOUND + - RESOURCE_EXHAUSTED + - SERVICE_UNDER_MAINTENANCE + - TEMPORARILY_UNAVAILABLE + - UNAUTHENTICATED + - UNKNOWN + - WORKSPACE_TEMPORARILY_UNAVAILABLE + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SetWorkspaceWarehouseConfigRequestSecurityPolicy: + type: string + enum: + - DATA_ACCESS_CONTROL + - NONE + - PASSTHROUGH + description: Security policy to be used for warehouses + SpotInstancePolicy: + type: string + enum: + - COST_OPTIMIZED + - POLICY_UNSPECIFIED + - RELIABILITY_OPTIMIZED + description: >- + EndpointSpotInstancePolicy configures whether the endpoint should use + spot instances. + + + The breakdown of how the EndpointSpotInstancePolicy converts to per + cloud configurations is: + + + +-------+--------------------------------------+--------------------------------+ + | Cloud | + + COST_OPTIMIZED | RELIABILITY_OPTIMIZED | + + +-------+--------------------------------------+--------------------------------+ + | AWS | On + + Demand Driver with Spot Executors | On Demand Driver and Executors | | + AZURE | On Demand Driver + + and Executors | On Demand Driver and Executors | + + +-------+--------------------------------------+--------------------------------+ + + + While including "spot" in the enum name may limit the the future + extensibility of this field + + because it limits this enum to denoting "spot or not", this is the field + that PM recommends + + after discussion with customers per SC-48783. + State: + type: string + enum: + - DELETED + - DELETING + - RUNNING + - STARTING + - STOPPED + - STOPPING + description: '* State of a warehouse.' + StatementState: + type: string + enum: + - CANCELED + - CLOSED + - FAILED + - PENDING + - RUNNING + - SUCCEEDED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + Status: + type: string + enum: + - DEGRADED + - FAILED + - HEALTHY + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + SuccessMessage: + type: string + enum: + - Success + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + TerminationReasonCode: + type: string + enum: + - ABUSE_DETECTED + - ACCESS_TOKEN_FAILURE + - ALLOCATION_TIMEOUT + - ALLOCATION_TIMEOUT_NODE_DAEMON_NOT_READY + - ALLOCATION_TIMEOUT_NO_HEALTHY_AND_WARMED_UP_CLUSTERS + - ALLOCATION_TIMEOUT_NO_HEALTHY_CLUSTERS + - ALLOCATION_TIMEOUT_NO_MATCHED_CLUSTERS + - ALLOCATION_TIMEOUT_NO_READY_CLUSTERS + - ALLOCATION_TIMEOUT_NO_UNALLOCATED_CLUSTERS + - ALLOCATION_TIMEOUT_NO_WARMED_UP_CLUSTERS + - ATTACH_PROJECT_FAILURE + - AWS_AUTHORIZATION_FAILURE + - AWS_INACCESSIBLE_KMS_KEY_FAILURE + - AWS_INSTANCE_PROFILE_UPDATE_FAILURE + - AWS_INSUFFICIENT_FREE_ADDRESSES_IN_SUBNET_FAILURE + - AWS_INSUFFICIENT_INSTANCE_CAPACITY_FAILURE + - AWS_INVALID_KEY_PAIR + - AWS_INVALID_KMS_KEY_STATE + - AWS_MAX_SPOT_INSTANCE_COUNT_EXCEEDED_FAILURE + - AWS_REQUEST_LIMIT_EXCEEDED + - AWS_RESOURCE_QUOTA_EXCEEDED + - AWS_UNSUPPORTED_FAILURE + - AZURE_BYOK_KEY_PERMISSION_FAILURE + - AZURE_EPHEMERAL_DISK_FAILURE + - AZURE_INVALID_DEPLOYMENT_TEMPLATE + - AZURE_OPERATION_NOT_ALLOWED_EXCEPTION + - AZURE_PACKED_DEPLOYMENT_PARTIAL_FAILURE + - AZURE_QUOTA_EXCEEDED_EXCEPTION + - AZURE_RESOURCE_MANAGER_THROTTLING + - AZURE_RESOURCE_PROVIDER_THROTTLING + - AZURE_UNEXPECTED_DEPLOYMENT_TEMPLATE_FAILURE + - AZURE_VM_EXTENSION_FAILURE + - AZURE_VNET_CONFIGURATION_FAILURE + - BOOTSTRAP_TIMEOUT + - BOOTSTRAP_TIMEOUT_CLOUD_PROVIDER_EXCEPTION + - BOOTSTRAP_TIMEOUT_DUE_TO_MISCONFIG + - BUDGET_POLICY_LIMIT_ENFORCEMENT_ACTIVATED + - BUDGET_POLICY_RESOLUTION_FAILURE + - CLOUD_ACCOUNT_POD_QUOTA_EXCEEDED + - CLOUD_ACCOUNT_SETUP_FAILURE + - CLOUD_OPERATION_CANCELLED + - CLOUD_PROVIDER_DISK_SETUP_FAILURE + - CLOUD_PROVIDER_INSTANCE_NOT_LAUNCHED + - CLOUD_PROVIDER_LAUNCH_FAILURE + - CLOUD_PROVIDER_LAUNCH_FAILURE_DUE_TO_MISCONFIG + - CLOUD_PROVIDER_RESOURCE_STOCKOUT + - CLOUD_PROVIDER_RESOURCE_STOCKOUT_DUE_TO_MISCONFIG + - CLOUD_PROVIDER_SHUTDOWN + - CLUSTER_OPERATION_THROTTLED + - CLUSTER_OPERATION_TIMEOUT + - COMMUNICATION_LOST + - CONTAINER_LAUNCH_FAILURE + - CONTROL_PLANE_CONNECTION_FAILURE + - CONTROL_PLANE_CONNECTION_FAILURE_DUE_TO_MISCONFIG + - CONTROL_PLANE_REQUEST_FAILURE + - CONTROL_PLANE_REQUEST_FAILURE_DUE_TO_MISCONFIG + - DATABASE_CONNECTION_FAILURE + - DATA_ACCESS_CONFIG_CHANGED + - DBFS_COMPONENT_UNHEALTHY + - DBR_IMAGE_RESOLUTION_FAILURE + - DISASTER_RECOVERY_REPLICATION + - DNS_RESOLUTION_ERROR + - DOCKER_CONTAINER_CREATION_EXCEPTION + - DOCKER_IMAGE_PULL_FAILURE + - DOCKER_IMAGE_TOO_LARGE_FOR_INSTANCE_EXCEPTION + - DOCKER_INVALID_OS_EXCEPTION + - DRIVER_EVICTION + - DRIVER_LAUNCH_TIMEOUT + - DRIVER_NODE_UNREACHABLE + - DRIVER_OUT_OF_DISK + - DRIVER_OUT_OF_MEMORY + - DRIVER_POD_CREATION_FAILURE + - DRIVER_UNEXPECTED_FAILURE + - DRIVER_UNHEALTHY + - DRIVER_UNREACHABLE + - DRIVER_UNRESPONSIVE + - DYNAMIC_SPARK_CONF_SIZE_EXCEEDED + - EOS_SPARK_IMAGE + - EXECUTION_COMPONENT_UNHEALTHY + - EXECUTOR_POD_UNSCHEDULED + - GCP_API_RATE_QUOTA_EXCEEDED + - GCP_DENIED_BY_ORG_POLICY + - GCP_FORBIDDEN + - GCP_IAM_TIMEOUT + - GCP_INACCESSIBLE_KMS_KEY_FAILURE + - GCP_INSUFFICIENT_CAPACITY + - GCP_IP_SPACE_EXHAUSTED + - GCP_KMS_KEY_PERMISSION_DENIED + - GCP_NOT_FOUND + - GCP_QUOTA_EXCEEDED + - GCP_RESOURCE_QUOTA_EXCEEDED + - GCP_SERVICE_ACCOUNT_ACCESS_DENIED + - GCP_SERVICE_ACCOUNT_DELETED + - GCP_SERVICE_ACCOUNT_NOT_FOUND + - GCP_SUBNET_NOT_READY + - GCP_TRUSTED_IMAGE_PROJECTS_VIOLATED + - GKE_BASED_CLUSTER_TERMINATION + - GLOBAL_INIT_SCRIPT_FAILURE + - HIVE_METASTORE_PROVISIONING_FAILURE + - IMAGE_PULL_PERMISSION_DENIED + - INACTIVITY + - INIT_CONTAINER_NOT_FINISHED + - INIT_SCRIPT_FAILURE + - INSTANCE_POOL_CLUSTER_FAILURE + - INSTANCE_POOL_MAX_CAPACITY_REACHED + - INSTANCE_POOL_NOT_FOUND + - INSTANCE_UNREACHABLE + - INSTANCE_UNREACHABLE_DUE_TO_MISCONFIG + - INTERNAL_CAPACITY_FAILURE + - INTERNAL_ERROR + - INVALID_ARGUMENT + - INVALID_AWS_PARAMETER + - INVALID_INSTANCE_PLACEMENT_PROTOCOL + - INVALID_SPARK_IMAGE + - INVALID_WORKER_IMAGE_FAILURE + - IN_PENALTY_BOX + - IP_EXHAUSTION_FAILURE + - JOB_FINISHED + - K8S_ACTIVE_POD_QUOTA_EXCEEDED + - K8S_AUTOSCALING_FAILURE + - K8S_DBR_CLUSTER_LAUNCH_TIMEOUT + - LAZY_ALLOCATION_TIMEOUT + - MAINTENANCE_MODE + - METASTORE_COMPONENT_UNHEALTHY + - MTLS_PORT_CONNECTIVITY_FAILURE + - NEPHOS_RESOURCE_MANAGEMENT + - NETVISOR_SETUP_TIMEOUT + - NETWORK_CHECK_CONTROL_PLANE_FAILURE + - NETWORK_CHECK_CONTROL_PLANE_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_DNS_SERVER_FAILURE + - NETWORK_CHECK_DNS_SERVER_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_METADATA_ENDPOINT_FAILURE + - NETWORK_CHECK_METADATA_ENDPOINT_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_MULTIPLE_COMPONENTS_FAILURE + - NETWORK_CHECK_MULTIPLE_COMPONENTS_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_NIC_FAILURE + - NETWORK_CHECK_NIC_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CHECK_STORAGE_FAILURE + - NETWORK_CHECK_STORAGE_FAILURE_DUE_TO_MISCONFIG + - NETWORK_CONFIGURATION_FAILURE + - NFS_MOUNT_FAILURE + - NO_MATCHED_K8S + - NO_MATCHED_K8S_TESTING_TAG + - NPIP_TUNNEL_SETUP_FAILURE + - NPIP_TUNNEL_TOKEN_FAILURE + - POD_ASSIGNMENT_FAILURE + - POD_SCHEDULING_FAILURE + - RATE_LIMITED + - REQUEST_REJECTED + - REQUEST_THROTTLED + - RESOURCE_USAGE_BLOCKED + - SECRET_CREATION_FAILURE + - SECRET_PERMISSION_DENIED + - SECRET_RESOLUTION_ERROR + - SECURITY_DAEMON_REGISTRATION_EXCEPTION + - SELF_BOOTSTRAP_FAILURE + - SERVERLESS_LONG_RUNNING_TERMINATED + - SKIPPED_SLOW_NODES + - SLOW_IMAGE_DOWNLOAD + - SPARK_ERROR + - SPARK_IMAGE_DOWNLOAD_FAILURE + - SPARK_IMAGE_DOWNLOAD_THROTTLED + - SPARK_IMAGE_NOT_FOUND + - SPARK_STARTUP_FAILURE + - SPOT_INSTANCE_TERMINATION + - SSH_BOOTSTRAP_FAILURE + - STORAGE_DOWNLOAD_FAILURE + - STORAGE_DOWNLOAD_FAILURE_DUE_TO_MISCONFIG + - STORAGE_DOWNLOAD_FAILURE_SLOW + - STORAGE_DOWNLOAD_FAILURE_THROTTLED + - STS_CLIENT_SETUP_FAILURE + - SUBNET_EXHAUSTED_FAILURE + - TEMPORARILY_UNAVAILABLE + - TRIAL_EXPIRED + - UNEXPECTED_LAUNCH_FAILURE + - UNEXPECTED_POD_RECREATION + - UNKNOWN + - UNSUPPORTED_INSTANCE_TYPE + - UPDATE_INSTANCE_PROFILE_FAILURE + - USAGE_POLICY_ENTITLEMENT_DENIED + - USER_INITIATED_VM_TERMINATION + - USER_REQUEST + - WORKER_SETUP_FAILURE + - WORKSPACE_CANCELLED_ERROR + - WORKSPACE_CONFIGURATION_ERROR + - WORKSPACE_UPDATE + description: The status code indicating why the cluster was terminated + TerminationReasonType: + type: string + enum: + - CLIENT_ERROR + - CLOUD_FAILURE + - SERVICE_FAULT + - SUCCESS + description: type of the termination + WarehousePermissionLevel: + type: string + enum: + - CAN_MANAGE + - CAN_MONITOR + - CAN_USE + - CAN_VIEW + - IS_OWNER + description: Permission level + WarehouseTypePairWarehouseType: + type: string + enum: + - CLASSIC + - PRO + - TYPE_UNSPECIFIED + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + x-stackQL-resources: + alerts: + id: databricks_workspace.sql.alerts + name: alerts + title: Alerts + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1sql~1alerts/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1sql~1alerts/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.results + delete: + operation: + $ref: '#/paths/~1api~12.0~1sql~1alerts~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1sql~1alerts~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1sql~1alerts~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/alerts/methods/get' + - $ref: '#/components/x-stackQL-resources/alerts/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/alerts/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/alerts/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/alerts/methods/delete' + replace: [] + alerts_legacy: + id: databricks_workspace.sql.alerts_legacy + name: alerts_legacy + title: Alerts Legacy + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts/get' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts~1{alert_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts~1{alert_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1alerts~1{alert_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/get' + - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/alerts_legacy/methods/update' + alerts_v2: + id: databricks_workspace.sql.alerts_v2 + name: alerts_v2 + title: Alerts V2 + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1alerts/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1alerts/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.alerts + get: + operation: + $ref: '#/paths/~1api~12.0~1alerts~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1alerts~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1alerts~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/alerts_v2/methods/get' + - $ref: '#/components/x-stackQL-resources/alerts_v2/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/alerts_v2/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/alerts_v2/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/alerts_v2/methods/delete' + replace: [] + dashboard_widgets: + id: databricks_workspace.sql.dashboard_widgets + name: dashboard_widgets + title: Dashboard Widgets + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1widgets/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1widgets~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1widgets~1{id}/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: '#/components/x-stackQL-resources/dashboard_widgets/methods/update' + - $ref: '#/components/x-stackQL-resources/dashboard_widgets/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/dashboard_widgets/methods/delete' + replace: [] + dashboards: + id: databricks_workspace.sql.dashboards + name: dashboards + title: Dashboards + methods: + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1sql~1dashboards~1{dashboard_id}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards~1{dashboard_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards~1{dashboard_id}/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1dashboards/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.results + restore: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1sql~1dashboards~1trash~1{dashboard_id}/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/dashboards/methods/get' + - $ref: '#/components/x-stackQL-resources/dashboards/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/dashboards/methods/update' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/dashboards/methods/delete' + replace: [] + data_sources: + id: databricks_workspace.sql.data_sources + name: data_sources + title: Data Sources + methods: + list: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1data_sources/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/data_sources/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + dbsql_permissions: + id: databricks_workspace.sql.dbsql_permissions + name: dbsql_permissions + title: Dbsql Permissions + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1sql~1permissions~1{object_type.value}~1{object_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1sql~1permissions~1{object_type.value}~1{object_id}/post + response: + mediaType: application/json + openAPIDocKey: '200' + transfer_ownership: + operation: + $ref: >- + #/paths/~1api~12.0~1preview~1sql~1permissions~1{object_type.value}~1{object_id}~1transfer/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/dbsql_permissions/methods/get' + insert: + - $ref: '#/components/x-stackQL-resources/dbsql_permissions/methods/set' + update: [] + delete: [] + replace: [] + queries: + id: databricks_workspace.sql.queries + name: queries + title: Queries + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1sql~1queries/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1sql~1queries/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.results + delete: + operation: + $ref: '#/paths/~1api~12.0~1sql~1queries~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1sql~1queries~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1sql~1queries~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/queries/methods/get' + - $ref: '#/components/x-stackQL-resources/queries/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/queries/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/queries/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/queries/methods/delete' + replace: [] + query_visualizations: + id: databricks_workspace.sql.query_visualizations + name: query_visualizations + title: Query Visualizations + methods: + list: + operation: + $ref: '#/paths/~1api~12.0~1sql~1queries~1{id}~1visualizations/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.results + create: + operation: + $ref: '#/paths/~1api~12.0~1sql~1visualizations/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1sql~1visualizations~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1sql~1visualizations~1{id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/query_visualizations/methods/list' + insert: + - $ref: >- + #/components/x-stackQL-resources/query_visualizations/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/query_visualizations/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/query_visualizations/methods/delete + replace: [] + queries_legacy: + id: databricks_workspace.sql.queries_legacy + name: queries_legacy + title: Queries Legacy + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1queries/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1queries/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.results + delete: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1{query_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1{query_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + replace: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1{query_id}/post' + response: + mediaType: application/json + openAPIDocKey: '200' + restore: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1queries~1trash~1{query_id}/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/get' + - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/queries_legacy/methods/replace' + query_history: + id: databricks_workspace.sql.query_history + name: query_history + title: Query History + methods: + list: + operation: + $ref: '#/paths/~1api~12.0~1sql~1history~1queries/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/query_history/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + query_visualizations_legacy: + id: databricks_workspace.sql.query_visualizations_legacy + name: query_visualizations_legacy + title: Query Visualizations Legacy + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1visualizations/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1visualizations~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1preview~1sql~1visualizations~1{id}/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: [] + insert: + - $ref: >- + #/components/x-stackQL-resources/query_visualizations_legacy/methods/update + - $ref: >- + #/components/x-stackQL-resources/query_visualizations_legacy/methods/create + update: [] + delete: + - $ref: >- + #/components/x-stackQL-resources/query_visualizations_legacy/methods/delete + replace: [] + redash_config: + id: databricks_workspace.sql.redash_config + name: redash_config + title: Redash Config + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1redash-v2~1config/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/redash_config/methods/get' + insert: [] + update: [] + delete: [] + replace: [] + statement_execution: + id: databricks_workspace.sql.statement_execution + name: statement_execution + title: Statement Execution + methods: + cancel: + operation: + $ref: '#/paths/~1api~12.0~1sql~1statements~1{statement_id}~1cancel/post' + response: + mediaType: application/json + openAPIDocKey: '200' + execute: + operation: + $ref: '#/paths/~1api~12.0~1sql~1statements/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1sql~1statements~1{statement_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + get_result_chunk: + operation: + $ref: >- + #/paths/~1api~12.0~1sql~1statements~1{statement_id}~1result~1chunks~1{chunk_index}/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/statement_execution/methods/get_result_chunk + - $ref: '#/components/x-stackQL-resources/statement_execution/methods/get' + insert: + - $ref: >- + #/components/x-stackQL-resources/statement_execution/methods/cancel + - $ref: >- + #/components/x-stackQL-resources/statement_execution/methods/execute + update: [] + delete: [] + replace: [] + warehouses: + id: databricks_workspace.sql.warehouses + name: warehouses + title: Warehouses + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1sql~1warehouses/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1sql~1warehouses/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.warehouses + delete: + operation: + $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + edit: + operation: + $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}~1edit/post' + response: + mediaType: application/json + openAPIDocKey: '200' + start: + operation: + $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}~1start/post' + response: + mediaType: application/json + openAPIDocKey: '200' + stop: + operation: + $ref: '#/paths/~1api~12.0~1sql~1warehouses~1{id}~1stop/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/warehouses/methods/get' + - $ref: '#/components/x-stackQL-resources/warehouses/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/warehouses/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/warehouses/methods/delete' + replace: + - $ref: '#/components/x-stackQL-resources/warehouses/methods/edit' + default_warehouse_overrides: + id: databricks_workspace.sql.default_warehouse_overrides + name: default_warehouse_overrides + title: Default Warehouse Overrides + methods: + create: + operation: + $ref: '#/paths/~1api~1warehouses~1v1~1default-warehouse-overrides/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~1warehouses~1v1~1default-warehouse-overrides/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.default_warehouse_overrides + delete: + operation: + $ref: '#/paths/~1api~1warehouses~1v1~1{name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~1warehouses~1v1~1{name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~1warehouses~1v1~1{name}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/default_warehouse_overrides/methods/get + - $ref: >- + #/components/x-stackQL-resources/default_warehouse_overrides/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/default_warehouse_overrides/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/default_warehouse_overrides/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/default_warehouse_overrides/methods/delete + replace: [] + warehouse_permission_levels: + id: databricks_workspace.sql.warehouse_permission_levels + name: warehouse_permission_levels + title: Warehouse Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/warehouse_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + warehouse_permissions: + id: databricks_workspace.sql.warehouse_permissions + name: warehouse_permissions + title: Warehouse Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1warehouses~1{warehouse_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/warehouse_permissions/methods/get' + insert: [] + update: + - $ref: >- + #/components/x-stackQL-resources/warehouse_permissions/methods/update + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/warehouse_permissions/methods/set' + workspace_warehouse_config: + id: databricks_workspace.sql.workspace_warehouse_config + name: workspace_warehouse_config + title: Workspace Warehouse Config + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1sql~1config~1warehouses/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1sql~1config~1warehouses/put' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/workspace_warehouse_config/methods/get + insert: [] + update: [] + delete: [] + replace: + - $ref: >- + #/components/x-stackQL-resources/workspace_warehouse_config/methods/set diff --git a/providers/src/databricks_workspace/v00.00.00000/services/tags.yaml b/providers/src/databricks_workspace/v00.00.00000/services/tags.yaml new file mode 100644 index 00000000..d2b652d2 --- /dev/null +++ b/providers/src/databricks_workspace/v00.00.00000/services/tags.yaml @@ -0,0 +1,835 @@ +openapi: 3.0.0 +info: + title: Databricks Tags API (workspace) + description: >- + OpenAPI specification for the Databricks tags service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.tags +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.1/tag-policies: + post: + operationId: tag_policies_create_tag_policy + summary: >- + Creates a new tag policy, making the associated tag key governed. For + Terraform usage, see the [Tag + tags: + - tags + - tag_policies + description: >- + Creates a new tag policy, making the associated tag key governed. For + Terraform usage, see the [Tag + + Policy Terraform documentation]. To manage permissions for tag policies, + use the [Account Access + + Control Proxy API]. + + + [Account Access Control Proxy API]: + https://docs.databricks.com/api/workspace/accountaccesscontrolproxy + + [Tag Policy Terraform documentation]: + https://registry.terraform.io/providers/databricks/databricks/latest/docs/resources/tag_policy + + + :param tag_policy: :class:`TagPolicy` + + + :returns: :class:`TagPolicy` + requestBody: + content: + application/json: + schema: + type: object + properties: + tag_policy: + type: string + description: ':returns: :class:`TagPolicy`' + required: + - tag_policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TagPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TagPoliciesAPI + get: + operationId: tag_policies_list_tag_policies + summary: >- + Lists the tag policies for all governed tags in the account. For + Terraform usage, see the [Tag Policy + tags: + - tags + - tag_policies + description: >- + Lists the tag policies for all governed tags in the account. For + Terraform usage, see the [Tag Policy + + Terraform documentation]. To list granted permissions for tag policies, + use the [Account Access + + Control Proxy API]. + + + [Account Access Control Proxy API]: + https://docs.databricks.com/api/workspace/accountaccesscontrolproxy + + [Tag Policy Terraform documentation]: + https://registry.terraform.io/providers/databricks/databricks/latest/docs/data-sources/tag_policies + + + :param page_size: int (optional) + The maximum number of results to return in this request. Fewer results may be returned than + requested. If unspecified or set to 0, this defaults to 1000. The maximum value is 1000; values + above 1000 will be coerced down to 1000. + :param page_token: str (optional) + An optional page token received from a previous list tag policies call. + + :returns: Iterator over :class:`TagPolicy` + parameters: + - name: page_size + in: query + required: false + schema: + type: string + description: >- + The maximum number of results to return in this request. Fewer + results may be returned than requested. If unspecified or set to 0, + this defaults to 1000. The maximum value is 1000; values above 1000 + will be coerced down to 1000. + - name: page_token + in: query + required: false + schema: + type: string + description: >- + An optional page token received from a previous list tag policies + call. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListTagPoliciesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TagPoliciesAPI + /api/2.1/tag-policies/{tag_key}: + delete: + operationId: tag_policies_delete_tag_policy + summary: >- + Deletes a tag policy by its associated governed tag's key, leaving that + tag key ungoverned. For + tags: + - tags + - tag_policies + description: >- + Deletes a tag policy by its associated governed tag's key, leaving that + tag key ungoverned. For + + Terraform usage, see the [Tag Policy Terraform documentation]. + + + [Tag Policy Terraform documentation]: + https://registry.terraform.io/providers/databricks/databricks/latest/docs/resources/tag_policy + + + :param tag_key: str + parameters: + - name: tag_key + in: path + required: true + schema: + type: string + description: str + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TagPoliciesAPI + get: + operationId: tag_policies_get_tag_policy + summary: >- + Gets a single tag policy by its associated governed tag's key. For + Terraform usage, see the [Tag + tags: + - tags + - tag_policies + description: >- + Gets a single tag policy by its associated governed tag's key. For + Terraform usage, see the [Tag + + Policy Terraform documentation]. To list granted permissions for tag + policies, use the [Account Access + + Control Proxy API]. + + + [Account Access Control Proxy API]: + https://docs.databricks.com/api/workspace/accountaccesscontrolproxy + + [Tag Policy Terraform documentation]: + https://registry.terraform.io/providers/databricks/databricks/latest/docs/data-sources/tag_policy + + + :param tag_key: str + + + :returns: :class:`TagPolicy` + parameters: + - name: tag_key + in: path + required: true + schema: + type: string + description: ':returns: :class:`TagPolicy`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TagPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TagPoliciesAPI + patch: + operationId: tag_policies_update_tag_policy + summary: >- + Updates an existing tag policy for a single governed tag. For Terraform + usage, see the [Tag Policy + tags: + - tags + - tag_policies + description: >- + Updates an existing tag policy for a single governed tag. For Terraform + usage, see the [Tag Policy + + Terraform documentation]. To manage permissions for tag policies, use + the [Account Access Control + + Proxy API]. + + + [Account Access Control Proxy API]: + https://docs.databricks.com/api/workspace/accountaccesscontrolproxy + + [Tag Policy Terraform documentation]: + https://registry.terraform.io/providers/databricks/databricks/latest/docs/resources/tag_policy + + + :param tag_key: str + + :param tag_policy: :class:`TagPolicy` + + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`TagPolicy` + parameters: + - name: tag_key + in: path + required: true + schema: + type: string + description: ':param tag_policy: :class:`TagPolicy`' + - name: update_mask + in: query + required: true + schema: + type: string + description: >- + The field mask must be a single string, with multiple fields + separated by commas (no spaces). The field path is relative to the + resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map + fields is not allowed, as only the entire collection field can be + specified. Field names must exactly match the resource field names. + A field mask of `*` indicates full replacement. It’s recommended to + always explicitly list the fields being updated and avoid using `*` + wildcards, as it can lead to unintended results if the API changes + in the future. + requestBody: + content: + application/json: + schema: + type: object + properties: + tag_policy: + type: string + required: + - tag_policy + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TagPolicy' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: TagPoliciesAPI + /api/2.0/entity-tag-assignments: + post: + operationId: workspace_entity_tag_assignments_create_tag_assignment + summary: Create a tag assignment + tags: + - tags + - workspace_entity_tag_assignments + description: |- + Create a tag assignment + + :param tag_assignment: :class:`TagAssignment` + + :returns: :class:`TagAssignment` + requestBody: + content: + application/json: + schema: + type: object + properties: + tag_assignment: + type: string + description: ':returns: :class:`TagAssignment`' + required: + - tag_assignment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TagAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceEntityTagAssignmentsAPI + /api/2.0/entity-tag-assignments/{entity_type}/{entity_id}/tags/{tag_key}: + delete: + operationId: workspace_entity_tag_assignments_delete_tag_assignment + summary: Delete a tag assignment + tags: + - tags + - workspace_entity_tag_assignments + description: |- + Delete a tag assignment + + :param entity_type: str + The type of entity to which the tag is assigned. Allowed values are apps, dashboards, geniespaces + :param entity_id: str + The identifier of the entity to which the tag is assigned + :param tag_key: str + The key of the tag. The characters , . : / - = and leading/trailing spaces are not allowed + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of entity to which the tag is assigned. Allowed values are + apps, dashboards, geniespaces + - name: entity_id + in: path + required: true + schema: + type: string + description: The identifier of the entity to which the tag is assigned + - name: tag_key + in: path + required: true + schema: + type: string + description: >- + The key of the tag. The characters , . : / - = and leading/trailing + spaces are not allowed + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceEntityTagAssignmentsAPI + get: + operationId: workspace_entity_tag_assignments_get_tag_assignment + summary: Get a tag assignment + tags: + - tags + - workspace_entity_tag_assignments + description: |- + Get a tag assignment + + :param entity_type: str + The type of entity to which the tag is assigned. Allowed values are apps, dashboards, geniespaces + :param entity_id: str + The identifier of the entity to which the tag is assigned + :param tag_key: str + The key of the tag. The characters , . : / - = and leading/trailing spaces are not allowed + + :returns: :class:`TagAssignment` + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of entity to which the tag is assigned. Allowed values are + apps, dashboards, geniespaces + - name: entity_id + in: path + required: true + schema: + type: string + description: The identifier of the entity to which the tag is assigned + - name: tag_key + in: path + required: true + schema: + type: string + description: >- + The key of the tag. The characters , . : / - = and leading/trailing + spaces are not allowed + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TagAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceEntityTagAssignmentsAPI + patch: + operationId: workspace_entity_tag_assignments_update_tag_assignment + summary: Update a tag assignment + tags: + - tags + - workspace_entity_tag_assignments + description: |- + Update a tag assignment + + :param entity_type: str + The type of entity to which the tag is assigned. Allowed values are apps, dashboards, geniespaces + :param entity_id: str + The identifier of the entity to which the tag is assigned + :param tag_key: str + The key of the tag. The characters , . : / - = and leading/trailing spaces are not allowed + :param tag_assignment: :class:`TagAssignment` + :param update_mask: str + The field mask must be a single string, with multiple fields separated by commas (no spaces). The + field path is relative to the resource object, using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence or map fields is not allowed, as only + the entire collection field can be specified. Field names must exactly match the resource field + names. + + A field mask of `*` indicates full replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it can lead to unintended results if the API + changes in the future. + + :returns: :class:`TagAssignment` + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of entity to which the tag is assigned. Allowed values are + apps, dashboards, geniespaces + - name: entity_id + in: path + required: true + schema: + type: string + description: The identifier of the entity to which the tag is assigned + - name: tag_key + in: path + required: true + schema: + type: string + description: >- + The key of the tag. The characters , . : / - = and leading/trailing + spaces are not allowed + - name: update_mask + in: query + required: true + schema: + type: string + requestBody: + content: + application/json: + schema: + type: object + properties: + tag_assignment: + type: string + description: >- + :param update_mask: str The field mask must be a single + string, with multiple fields separated by commas (no + spaces). The field path is relative to the resource object, + using a dot (`.`) to navigate sub-fields (e.g., + `author.given_name`). Specification of elements in sequence + or map fields is not allowed, as only the entire collection + field can be specified. Field names must exactly match the + resource field names. A field mask of `*` indicates full + replacement. It’s recommended to always explicitly list the + fields being updated and avoid using `*` wildcards, as it + can lead to unintended results if the API changes in the + future. + required: + - tag_assignment + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/TagAssignment' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceEntityTagAssignmentsAPI + /api/2.0/entity-tag-assignments/{entity_type}/{entity_id}/tags: + get: + operationId: workspace_entity_tag_assignments_list_tag_assignments + summary: List the tag assignments for an entity + tags: + - tags + - workspace_entity_tag_assignments + description: |- + List the tag assignments for an entity + + :param entity_type: str + The type of entity to which the tag is assigned. Allowed values are apps, dashboards, geniespaces + :param entity_id: str + The identifier of the entity to which the tag is assigned + :param page_size: int (optional) + Optional. Maximum number of tag assignments to return in a single page + :param page_token: str (optional) + Pagination token to go to the next page of tag assignments. Requests first page if absent. + + :returns: Iterator over :class:`TagAssignment` + parameters: + - name: entity_type + in: path + required: true + schema: + type: string + description: >- + The type of entity to which the tag is assigned. Allowed values are + apps, dashboards, geniespaces + - name: entity_id + in: path + required: true + schema: + type: string + description: The identifier of the entity to which the tag is assigned + - name: page_size + in: query + required: false + schema: + type: string + description: >- + Optional. Maximum number of tag assignments to return in a single + page + - name: page_token + in: query + required: false + schema: + type: string + description: >- + Pagination token to go to the next page of tag assignments. Requests + first page if absent. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListTagAssignmentsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceEntityTagAssignmentsAPI +components: + schemas: + ListTagAssignmentsResponse: + type: object + properties: + next_page_token: + type: string + tag_assignments: + type: array + items: + $ref: '#/components/schemas/TagAssignment' + ListTagPoliciesResponse: + type: object + properties: + next_page_token: + type: string + tag_policies: + type: array + items: + $ref: '#/components/schemas/TagPolicy' + TagAssignment: + type: object + properties: + entity_type: + type: string + entity_id: + type: string + description: The identifier of the entity to which the tag is assigned + tag_key: + type: string + description: >- + The key of the tag. The characters , . : / - = and leading/trailing + spaces are not allowed + tag_value: + type: string + description: The value of the tag + required: + - entity_type + - entity_id + - tag_key + TagPolicy: + type: object + properties: + tag_key: + type: string + create_time: + type: string + description: Timestamp when the tag policy was created + description: + type: string + id: + type: string + update_time: + type: string + description: Timestamp when the tag policy was last updated + values: + type: array + items: + $ref: '#/components/schemas/Value' + required: + - tag_key + Value: + type: object + properties: + name: + type: string + required: + - name + x-stackQL-resources: + tag_policies: + id: databricks_workspace.tags.tag_policies + name: tag_policies + title: Tag Policies + methods: + create: + operation: + $ref: '#/paths/~1api~12.1~1tag-policies/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.1~1tag-policies/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.tag_policies + delete: + operation: + $ref: '#/paths/~1api~12.1~1tag-policies~1{tag_key}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.1~1tag-policies~1{tag_key}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.1~1tag-policies~1{tag_key}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/tag_policies/methods/get' + - $ref: '#/components/x-stackQL-resources/tag_policies/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/tag_policies/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/tag_policies/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/tag_policies/methods/delete' + replace: [] + workspace_entity_tag_assignments: + id: databricks_workspace.tags.workspace_entity_tag_assignments + name: workspace_entity_tag_assignments + title: Workspace Entity Tag Assignments + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1entity-tag-assignments/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1entity-tag-assignments~1{entity_type}~1{entity_id}~1tags~1{tag_key}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: >- + #/paths/~1api~12.0~1entity-tag-assignments~1{entity_type}~1{entity_id}~1tags~1{tag_key}/get + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1entity-tag-assignments~1{entity_type}~1{entity_id}~1tags~1{tag_key}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: >- + #/paths/~1api~12.0~1entity-tag-assignments~1{entity_type}~1{entity_id}~1tags/get + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.tag_assignments + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/workspace_entity_tag_assignments/methods/get + - $ref: >- + #/components/x-stackQL-resources/workspace_entity_tag_assignments/methods/list + insert: + - $ref: >- + #/components/x-stackQL-resources/workspace_entity_tag_assignments/methods/create + update: + - $ref: >- + #/components/x-stackQL-resources/workspace_entity_tag_assignments/methods/update + delete: + - $ref: >- + #/components/x-stackQL-resources/workspace_entity_tag_assignments/methods/delete + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/unitycatalog.yaml b/providers/src/databricks_workspace/v00.00.00000/services/unitycatalog.yaml deleted file mode 100644 index 4d3d6ef2..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/unitycatalog.yaml +++ /dev/null @@ -1,7945 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Unitycatalog API - description: Unitycatalog -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.1/unity-catalog/artifact-allowlists/{artifact_type}: - get: - operationId: artifact-allow-lists-get - externalDocs: - url: https://docs.databricks.com/api/workspace/artifactallowlists/get - x-stackQL-resource: artifact_allow_lists - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: artifact_type - required: true - description: string - in: path - responses: - '200': - description: The artifact allowlist was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - artifact_matchers: - type: array - items: - type: object - properties: - artifact: - type: string - match_type: - type: string - metastore_id: - type: string - created_by: - type: string - created_at: - type: integer - description: Get the artifact allowlist of a certain artifact type. The caller - must be a metastore admin or have the - put: - operationId: artifact-allow-lists-update - externalDocs: - url: https://docs.databricks.com/api/workspace/artifactallowlists/update - x-stackQL-resource: artifact_allow_lists - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: artifact_type - required: true - description: string - in: path - responses: - '200': - description: The artifact allowlist of a certain artifact type was updated - content: - application/json: - schema: - type: object - properties: - artifact_matchers: - type: array - items: - type: object - properties: - artifact: - type: string - match_type: - type: string - metastore_id: - type: string - created_by: - type: string - created_at: - type: integer - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - artifact_matchers: - type: required - example: - artifact_matchers: - - artifact: string - match_type: PREFIX_MATCH - description: Set the artifact allowlist of a certain artifact type. The whole - artifact allowlist is replaced with the new allowlist. The caller must be - a metastore admin or have the - /api/2.1/unity-catalog/catalogs: - post: - operationId: catalogs-create - externalDocs: - url: https://docs.databricks.com/api/workspace/catalogs/create - x-stackQL-resource: catalogs - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new catalog was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - provider_name: - type: string - share_name: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - storage_location: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - isolation_mode: - type: string - connection_name: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - browse_only: - type: boolean - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - comment: - type: string - storage_root: - type: string - provider_name: - type: string - share_name: - type: string - connection_name: - type: string - properties: - type: object - options: - type: object - example: - name: string - comment: string - properties: - property1: string - property2: string - storage_root: string - provider_name: string - share_name: string - connection_name: string - options: - property1: string - property2: string - description: Creates a new catalog instance in the parent metastore if the caller - is a metastore admin or has the - get: - operationId: catalogs-list - externalDocs: - url: https://docs.databricks.com/api/workspace/catalogs/list - x-stackQL-resource: catalogs - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: include_browse - in: query - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The catalog list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - catalogs: - type: array - items: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - provider_name: - type: string - share_name: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - storage_location: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - isolation_mode: - type: string - connection_name: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - browse_only: - type: boolean - next_page_token: - type: string - x-stackQL-objectKey: $.catalogs - description: Gets an array of catalogs in the metastore. If the caller is the - metastore admin, all catalogs will be retrieved. Otherwise, only catalogs - owned by the caller (or for which the caller has the - /api/2.1/unity-catalog/catalogs/{name}: - delete: - operationId: catalogs-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/catalogs/delete - x-stackQL-resource: catalogs - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: The catalog was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes the catalog that matches the supplied name. The caller - must be a metastore admin or the owner of the catalog. - get: - operationId: catalogs-get - externalDocs: - url: https://docs.databricks.com/api/workspace/catalogs/get - x-stackQL-resource: catalogs - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: include_browse - in: query - responses: - '200': - description: The catalog was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - provider_name: - type: string - share_name: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - storage_location: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - isolation_mode: - type: string - connection_name: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - browse_only: - type: boolean - description: Gets the specified catalog in a metastore. The caller must be a - metastore admin, the owner of the catalog, or a user that has the - patch: - operationId: catalogs-update - externalDocs: - url: https://docs.databricks.com/api/workspace/catalogs/update - x-stackQL-resource: catalogs - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The catalog was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - provider_name: - type: string - share_name: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - storage_location: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - isolation_mode: - type: string - connection_name: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - browse_only: - type: boolean - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - comment: - type: string - isolation_mode: - type: string - enable_predictive_optimization: - type: string - new_name: - type: string - properties: - type: object - example: - owner: string - comment: string - properties: - property1: string - property2: string - isolation_mode: OPEN - enable_predictive_optimization: DISABLE - new_name: string - description: Updates the catalog that matches the supplied name. The caller - must be either the owner of the catalog, or a metastore admin (when changing - the owner field of the catalog). - /api/2.1/unity-catalog/connections: - post: - operationId: connections-create - externalDocs: - url: https://docs.databricks.com/api/workspace/connections/create - x-stackQL-resource: connections - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The connection was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - connection_type: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - owner: - type: string - read_only: - type: boolean - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - url: - type: string - credential_type: - type: string - connection_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - connection_type: - type: string - read_only: - type: required - comment: - type: string - options: - type: required - properties: - type: object - example: - name: string - connection_type: MYSQL - options: - property1: string - property2: string - read_only: true - comment: string - properties: - property1: string - property2: string - description: Creates a new connection - get: - operationId: connections-list - externalDocs: - url: https://docs.databricks.com/api/workspace/connections/list - x-stackQL-resource: connections - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The connection list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - connections: - type: array - items: - type: object - properties: - name: - type: string - connection_type: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - owner: - type: string - read_only: - type: boolean - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - url: - type: string - credential_type: - type: string - connection_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.connections - description: List all connections. - /api/2.1/unity-catalog/connections/{name}: - delete: - operationId: connections-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/connections/delete - x-stackQL-resource: connections - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The connection was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes the connection that matches the supplied name. - get: - operationId: connections-get - externalDocs: - url: https://docs.databricks.com/api/workspace/connections/get - x-stackQL-resource: connections - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The connection was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - connection_type: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - owner: - type: string - read_only: - type: boolean - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - url: - type: string - credential_type: - type: string - connection_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - description: Gets a connection from it's name. - patch: - operationId: connections-update - externalDocs: - url: https://docs.databricks.com/api/workspace/connections/update - x-stackQL-resource: connections - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The connection was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - connection_type: - type: string - options: - type: object - properties: - property1: - type: string - property2: - type: string - owner: - type: string - read_only: - type: boolean - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - full_name: - type: string - url: - type: string - credential_type: - type: string - connection_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - securable_kind: - type: string - securable_type: - type: string - provisioning_info: - type: object - properties: - state: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - new_name: - type: string - options: - type: required - example: - options: - property1: string - property2: string - owner: string - new_name: string - description: Updates the connection that matches the supplied name. - /api/2.1/unity-catalog/credentials: - post: - operationId: credentials-createcredential - externalDocs: - url: https://docs.databricks.com/api/workspace/credentials/createcredential - x-stackQL-resource: credentials - x-stackQL-method: createcredential - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - purpose: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - full_name: - type: string - isolation_mode: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - comment: - type: string - read_only: - type: string - purpose: - type: boolean - skip_validation: - type: string - aws_iam_role: - type: object - example: - name: string - aws_iam_role: - role_arn: string - unity_catalog_iam_arn: string - external_id: string - comment: string - read_only: true - purpose: STORAGE - skip_validation: false - description: Creates a new credential. The type of credential to be created - is determined by the - get: - operationId: credentials-listcredentials - externalDocs: - url: https://docs.databricks.com/api/workspace/credentials/listcredentials - x-stackQL-resource: credentials - x-stackQL-method: listcredentials - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - - name: purpose - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - credentials: - type: array - items: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - purpose: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - full_name: - type: string - isolation_mode: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.credentials - description: Gets an array of credentials (as - /api/2.1/unity-catalog/credentials/{name_arg}: - delete: - operationId: credentials-deletecredential - externalDocs: - url: https://docs.databricks.com/api/workspace/credentials/deletecredential - x-stackQL-resource: credentials - x-stackQL-method: deletecredential - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name_arg - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a service or storage credential from the metastore. The - caller must be an owner of the credential. - get: - operationId: credentials-getcredential - externalDocs: - url: https://docs.databricks.com/api/workspace/credentials/getcredential - x-stackQL-resource: credentials - x-stackQL-method: getcredential - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name_arg - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - purpose: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - full_name: - type: string - isolation_mode: - type: string - description: Gets a service or storage credential from the metastore. The caller - must be a metastore admin, the owner of the credential, or have any permission - on the credential. - patch: - operationId: credentials-updatecredential - externalDocs: - url: https://docs.databricks.com/api/workspace/credentials/updatecredential - x-stackQL-resource: credentials - x-stackQL-method: updatecredential - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name_arg - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - purpose: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - full_name: - type: string - isolation_mode: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - new_name: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - isolation_mode: - type: string - skip_validation: - type: boolean - force: - type: boolean - aws_iam_role: - type: object - example: - new_name: string - aws_iam_role: - role_arn: string - unity_catalog_iam_arn: string - external_id: string - comment: string - read_only: true - owner: string - isolation_mode: ISOLATION_MODE_OPEN - skip_validation: true - force: true - description: Updates a service or storage credential on the metastore. - /api/2.1/unity-catalog/temporary-service-credentials: - post: - operationId: credentials-generatetemporaryservicecredential - externalDocs: - url: https://docs.databricks.com/api/workspace/credentials/generatetemporaryservicecredential - x-stackQL-resource: credentials - x-stackQL-method: generatetemporaryservicecredential - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - aws_temp_credentials: - type: object - properties: - access_key_id: - type: string - secret_access_key: - type: string - session_token: - type: string - access_point: - type: string - expiration_time: - type: integer - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - credential_name: - type: required - example: - credential_name: string - description: Returns a set of temporary credentials generated using the specified - service credential. The caller must be a metastore admin or have the metastore - privilege - /api/2.1/unity-catalog/validate-credentials: - post: - operationId: credential-validation-validatecredential - externalDocs: - url: https://docs.databricks.com/api/workspace/credentials/validatecredential - x-stackQL-resource: credential_validation - x-stackQL-method: validatecredential - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - results: - type: array - items: - type: object - properties: - result: - type: string - message: - type: string - isDir: - type: boolean - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - credential_name: - type: string - purpose: - type: string - external_location_name: - type: string - url: - type: string - read_only: - type: boolean - aws_iam_role: - type: object - example: - credential_name: string - aws_iam_role: - role_arn: string - unity_catalog_iam_arn: string - external_id: string - purpose: STORAGE - external_location_name: string - url: string - read_only: true - description: Validates a credential. - /api/2.1/unity-catalog/effective-permissions/{securable_type}/{full_name}: - get: - operationId: effective-grants-geteffective - externalDocs: - url: https://docs.databricks.com/api/workspace/grants/geteffective - x-stackQL-resource: effective_grants - x-stackQL-method: geteffective - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: securable_type - required: true - description: string - in: path - - name: full_name - required: true - description: string - in: path - - name: principal - in: query - responses: - '200': - description: The effective permissions list for securable was successfully - retrieved. - content: - application/json: - schema: - type: object - properties: - privilege_assignments: - type: array - items: - type: object - properties: - principal: - type: string - privileges: - type: array - items: - type: object - properties: - privilege: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - x-stackQL-objectKey: $.privilege_assignments - description: Gets the effective permissions for a securable. - /api/2.1/unity-catalog/external-locations: - post: - operationId: external-locations-create - externalDocs: - url: https://docs.databricks.com/api/workspace/externallocations/create - x-stackQL-resource: external_locations - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new external location was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - url: - type: string - credential_name: - type: string - read_only: - type: boolean - comment: - type: string - owner: - type: string - access_point: - type: string - metastore_id: - type: string - credential_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - browse_only: - type: boolean - isolation_mode: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - url: - type: string - credential_name: - type: required - read_only: - type: string - comment: - type: required - access_point: - type: string - skip_validation: - type: boolean - example: - name: string - url: string - credential_name: string - read_only: true - comment: string - access_point: string - skip_validation: true - description: Creates a new external location entry in the metastore. The caller - must be a metastore admin or have the - get: - operationId: external-locations-list - externalDocs: - url: https://docs.databricks.com/api/workspace/externallocations/list - x-stackQL-resource: external_locations - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - - name: include_browse - in: query - responses: - '200': - description: The external location list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - external_locations: - type: array - items: - type: object - properties: - name: - type: string - url: - type: string - credential_name: - type: string - read_only: - type: boolean - comment: - type: string - owner: - type: string - access_point: - type: string - metastore_id: - type: string - credential_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - browse_only: - type: boolean - isolation_mode: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.external_locations - description: Gets an array of external locations ( - /api/2.1/unity-catalog/external-locations/{name}: - delete: - operationId: external-locations-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/externallocations/delete - x-stackQL-resource: external_locations - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: The external location was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes the specified external location from the metastore. The - caller must be the owner of the external location. - get: - operationId: external-locations-get - externalDocs: - url: https://docs.databricks.com/api/workspace/externallocations/get - x-stackQL-resource: external_locations - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: include_browse - in: query - responses: - '200': - description: The external location was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - url: - type: string - credential_name: - type: string - read_only: - type: boolean - comment: - type: string - owner: - type: string - access_point: - type: string - metastore_id: - type: string - credential_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - browse_only: - type: boolean - isolation_mode: - type: string - description: Gets an external location from the metastore. The caller must be - either a metastore admin, the owner of the external location, or a user that - has some privilege on the external location. - patch: - operationId: external-locations-update - externalDocs: - url: https://docs.databricks.com/api/workspace/externallocations/update - x-stackQL-resource: external_locations - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The external location was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - url: - type: string - credential_name: - type: string - read_only: - type: boolean - comment: - type: string - owner: - type: string - access_point: - type: string - metastore_id: - type: string - credential_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - browse_only: - type: boolean - isolation_mode: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - url: - type: string - credential_name: - type: string - read_only: - type: boolean - comment: - type: string - owner: - type: string - access_point: - type: string - isolation_mode: - type: string - new_name: - type: string - force: - type: boolean - skip_validation: - type: boolean - example: - url: string - credential_name: string - read_only: true - comment: string - owner: string - access_point: string - isolation_mode: ISOLATION_MODE_OPEN - new_name: string - force: true - skip_validation: true - description: Updates an external location in the metastore. The caller must - be the owner of the external location, or be a metastore admin. In the second - case, the admin can only update the name of the external location. - /api/2.1/unity-catalog/functions: - post: - operationId: functions-create - externalDocs: - url: https://docs.databricks.com/api/workspace/functions/create - x-stackQL-resource: functions - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new function was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - input_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - data_type: - type: string - full_data_type: - type: string - return_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - routine_body: - type: string - routine_definition: - type: string - routine_dependencies: - type: object - properties: - dependencies: - type: array - items: - type: object - properties: - table: - type: object - properties: - table_full_name: - type: string - function: - type: object - properties: - function_full_name: - type: string - parameter_style: - type: string - is_deterministic: - type: boolean - sql_data_access: - type: string - is_null_call: - type: boolean - security_type: - type: string - specific_name: - type: string - external_name: - type: string - external_language: - type: string - sql_path: - type: string - owner: - type: string - comment: - type: string - properties: - type: string - metastore_id: - type: string - full_name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - function_id: - type: string - browse_only: - type: boolean - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - function_info: - type: object - example: - function_info: - name: string - catalog_name: string - schema_name: string - input_params: - parameters: - - name: string - type_text: string - type_json: string - type_name: BOOLEAN - type_precision: 0 - type_scale: 0 - type_interval_type: string - position: 0 - parameter_mode: IN - parameter_type: PARAM - parameter_default: string - comment: string - data_type: BOOLEAN - full_data_type: string - return_params: - parameters: - - name: string - type_text: string - type_json: string - type_name: BOOLEAN - type_precision: 0 - type_scale: 0 - type_interval_type: string - position: 0 - parameter_mode: IN - parameter_type: PARAM - parameter_default: string - comment: string - routine_definition: string - routine_dependencies: - dependencies: - - table: - table_full_name: string - function: - function_full_name: string - is_deterministic: true - is_null_call: true - specific_name: string - external_name: string - external_language: string - sql_path: string - comment: string - properties: string - routine_body: SQL - security_type: DEFINER - sql_data_access: CONTAINS_SQL - parameter_style: S - get: - operationId: functions-list - externalDocs: - url: https://docs.databricks.com/api/workspace/functions/list - x-stackQL-resource: functions - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: catalog_name - required: true - description: string - in: query - - name: schema_name - required: true - description: string - in: query - - name: max_results - in: query - - name: page_token - in: query - - name: include_browse - in: query - responses: - '200': - description: The function list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - functions: - type: array - items: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - input_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - data_type: - type: string - full_data_type: - type: string - return_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - routine_body: - type: string - routine_definition: - type: string - routine_dependencies: - type: object - properties: - dependencies: - type: array - items: - type: object - properties: - table: - type: object - properties: - table_full_name: - type: string - function: - type: object - properties: - function_full_name: - type: string - parameter_style: - type: string - is_deterministic: - type: boolean - sql_data_access: - type: string - is_null_call: - type: boolean - security_type: - type: string - specific_name: - type: string - external_name: - type: string - external_language: - type: string - sql_path: - type: string - owner: - type: string - comment: - type: string - properties: - type: string - metastore_id: - type: string - full_name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - function_id: - type: string - browse_only: - type: boolean - next_page_token: - type: string - x-stackQL-objectKey: $.functions - description: List functions within the specified parent catalog and schema. - If the user is a metastore admin, all functions are returned in the output - list. Otherwise, the user must have the - /api/2.1/unity-catalog/functions/{name}: - delete: - operationId: functions-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/functions/delete - x-stackQL-resource: functions - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: The function was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: 'Deletes the function that matches the supplied name. For the deletion - to succeed, the user must satisfy one of the following conditions:' - get: - operationId: functions-get - externalDocs: - url: https://docs.databricks.com/api/workspace/functions/get - x-stackQL-resource: functions - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: include_browse - in: query - responses: - '200': - description: The function was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - input_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - data_type: - type: string - full_data_type: - type: string - return_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - routine_body: - type: string - routine_definition: - type: string - routine_dependencies: - type: object - properties: - dependencies: - type: array - items: - type: object - properties: - table: - type: object - properties: - table_full_name: - type: string - function: - type: object - properties: - function_full_name: - type: string - parameter_style: - type: string - is_deterministic: - type: boolean - sql_data_access: - type: string - is_null_call: - type: boolean - security_type: - type: string - specific_name: - type: string - external_name: - type: string - external_language: - type: string - sql_path: - type: string - owner: - type: string - comment: - type: string - properties: - type: string - metastore_id: - type: string - full_name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - function_id: - type: string - browse_only: - type: boolean - description: 'Gets a function from within a parent catalog and schema. For the - fetch to succeed, the user must satisfy one of the following requirements:' - patch: - operationId: functions-update - externalDocs: - url: https://docs.databricks.com/api/workspace/functions/update - x-stackQL-resource: functions - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The function was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - input_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - data_type: - type: string - full_data_type: - type: string - return_params: - type: object - properties: - parameters: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_json: - type: string - type_name: - type: string - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - position: - type: integer - parameter_mode: - type: string - parameter_type: - type: string - parameter_default: - type: string - comment: - type: string - routine_body: - type: string - routine_definition: - type: string - routine_dependencies: - type: object - properties: - dependencies: - type: array - items: - type: object - properties: - table: - type: object - properties: - table_full_name: - type: string - function: - type: object - properties: - function_full_name: - type: string - parameter_style: - type: string - is_deterministic: - type: boolean - sql_data_access: - type: string - is_null_call: - type: boolean - security_type: - type: string - specific_name: - type: string - external_name: - type: string - external_language: - type: string - sql_path: - type: string - owner: - type: string - comment: - type: string - properties: - type: string - metastore_id: - type: string - full_name: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - function_id: - type: string - browse_only: - type: boolean - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - example: - owner: string - description: Updates the function that matches the supplied name. Only the owner - of the function can be updated. If the user is not a metastore admin, the - user must be a member of the group that is the new function owner. - /api/2.1/unity-catalog/permissions/{securable_type}/{full_name}: - get: - operationId: grants-get - externalDocs: - url: https://docs.databricks.com/api/workspace/grants/get - x-stackQL-resource: grants - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: securable_type - required: true - description: string - in: path - - name: full_name - required: true - description: string - in: path - - name: principal - in: query - responses: - '200': - description: The permissions list for securable was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - privilege_assignments: - type: array - items: - type: object - properties: - principal: - type: string - privileges: - type: array - items: - type: string - x-stackQL-objectKey: $.privilege_assignments - description: Gets the permissions for a securable. - patch: - operationId: grants-update - externalDocs: - url: https://docs.databricks.com/api/workspace/grants/update - x-stackQL-resource: grants - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: securable_type - required: true - description: string - in: path - - name: full_name - required: true - description: string - in: path - responses: - '200': - description: The permissions list for securable was successfully updated. - content: - application/json: - schema: - type: object - properties: - privilege_assignments: - type: array - items: - type: object - properties: - principal: - type: string - privileges: - type: array - items: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - changes: - type: Array of object - example: - changes: - - principal: string - add: - - SELECT - remove: - - SELECT - description: Updates the permissions for a securable. - /api/2.1/unity-catalog/metastores: - post: - operationId: metastores-create - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/create - x-stackQL-resource: metastores - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new metastore was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - storage_root: - type: string - region: - type: string - example: - name: string - storage_root: string - region: string - description: Creates a new metastore based on a provided name and optional storage - root path. By default (if the - get: - operationId: metastores-list - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/list - x-stackQL-resource: metastores - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The metastore list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - metastores: - type: array - items: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - x-stackQL-objectKey: $.metastores - description: Gets an array of the available metastores (as - /api/2.1/unity-catalog/metastores/{id}: - delete: - operationId: metastores-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/delete - x-stackQL-resource: metastores - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: The metastore was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a metastore. The caller must be a metastore admin. - get: - operationId: metastores-get - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/get - x-stackQL-resource: metastores - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: The metastore was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - description: Gets a metastore that matches the supplied ID. The caller must - be a metastore admin to retrieve this info. - patch: - operationId: metastores-update - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/update - x-stackQL-resource: metastores - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: The metastore was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - storage_root: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - delta_sharing_scope: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - region: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_root_credential_name: - type: string - cloud: - type: string - global_metastore_id: - type: string - external_access_enabled: - type: boolean - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - storage_root_credential_id: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - format: int64 - delta_sharing_organization_name: - type: string - owner: - type: string - privilege_model_version: - type: string - new_name: - type: string - delta_sharing_scope: - type: string - example: - storage_root_credential_id: string - delta_sharing_recipient_token_lifetime_in_seconds: 0 - delta_sharing_organization_name: string - owner: string - privilege_model_version: string - new_name: string - delta_sharing_scope: INTERNAL - description: Updates information for a specific metastore. The caller must be - a metastore admin. If the - /api/2.1/unity-catalog/workspaces/{workspace_id}/metastore: - put: - operationId: metastore-assignments-assign - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/assign - x-stackQL-resource: metastore_assignments - x-stackQL-method: assign - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: The metastore was successfully assigned. - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - metastore_id: - type: required - default_catalog_name: - type: string - example: - metastore_id: string - default_catalog_name: string - description: Creates a new metastore assignment. If an assignment for the same - delete: - operationId: metastore-assignments-unassign - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/unassign - x-stackQL-resource: metastore_assignments - x-stackQL-method: unassign - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: workspace_id - required: true - description: int64 - in: path - - name: metastore_id - required: true - description: string - in: query - responses: - '200': - description: The metastore assignment was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a metastore assignment. The caller must be an account administrator. - patch: - operationId: metastore-assignments-updateassignment - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/updateassignment - x-stackQL-resource: metastore_assignments - x-stackQL-method: updateassignment - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: workspace_id - required: true - description: int64 - in: path - responses: - '200': - description: The metastore assignment was successfully updated. - content: - application/json: - schema: - type: object - properties: {} - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - metastore_id: - type: string - default_catalog_name: - type: string - example: - metastore_id: string - default_catalog_name: string - description: Updates a metastore assignment. This operation can be used to update - /api/2.1/unity-catalog/current-metastore-assignment: - get: - operationId: metastore-assignments-current - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/current - x-stackQL-resource: metastore_assignments - x-stackQL-method: current - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The metastore assignment was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - metastore_id: - type: string - workspace_id: - type: integer - default_catalog_name: - type: string - description: Gets the metastore assignment for the workspace being accessed. - /api/2.1/unity-catalog/metastore_summary: - get: - operationId: metastore-summary-summary - externalDocs: - url: https://docs.databricks.com/api/workspace/metastores/summary - x-stackQL-resource: metastore_summary - x-stackQL-method: summary - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The metastore summary was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - metastore_id: - type: string - name: - type: string - default_data_access_config_id: - type: string - storage_root_credential_id: - type: string - cloud: - type: string - region: - type: string - global_metastore_id: - type: string - storage_root_credential_name: - type: string - privilege_model_version: - type: string - delta_sharing_recipient_token_lifetime_in_seconds: - type: integer - delta_sharing_organization_name: - type: string - storage_root: - type: string - owner: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - external_access_enabled: - type: boolean - delta_sharing_scope: - type: string - description: Gets information about a metastore. This summary includes the storage - credential, the cloud vendor, the cloud region, and the global metastore ID. - /api/2.1/unity-catalog/models/{full_name}/versions/{version}: - delete: - operationId: model-versions-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/modelversions/delete - x-stackQL-resource: model_versions - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: full_name - required: true - description: string - in: path - - name: version - required: true - description: integer - in: path - responses: - '200': - description: Successfully deleted the model version - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a model version from the specified registered model. Any - aliases assigned to the model version will also be deleted. - get: - operationId: model-versions-get - externalDocs: - url: https://docs.databricks.com/api/workspace/modelversions/get - x-stackQL-resource: model_versions - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: full_name - required: true - description: string - in: path - - name: version - required: true - description: integer - in: path - - name: include_browse - in: query - - name: include_aliases - in: query - responses: - '200': - description: Successfully retrieved the properties of the model version - content: - application/json: - schema: - type: object - properties: - model_name: - type: string - catalog_name: - type: string - schema_name: - type: string - comment: - type: string - source: - type: string - run_id: - type: string - run_workspace_id: - type: string - version: - type: integer - status: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get a model version. - patch: - operationId: model-versions-update - externalDocs: - url: https://docs.databricks.com/api/workspace/modelversions/update - x-stackQL-resource: model_versions - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: full_name - required: true - description: string - in: path - - name: version - required: true - description: integer - in: path - responses: - '200': - description: Successfully updated the properties of the model version - content: - application/json: - schema: - type: object - properties: - model_name: - type: string - catalog_name: - type: string - schema_name: - type: string - comment: - type: string - source: - type: string - run_id: - type: string - run_workspace_id: - type: string - version: - type: integer - status: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - comment: - type: string - example: - comment: This is my new model version comment - description: Updates the specified model version. - /api/2.1/unity-catalog/models/{full_name}/aliases/{alias}: - get: - operationId: model-versions-getbyalias - externalDocs: - url: https://docs.databricks.com/api/workspace/modelversions/getbyalias - x-stackQL-resource: model_versions - x-stackQL-method: getbyalias - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: full_name - required: true - description: string - in: path - - name: alias - required: true - description: string - in: path - - name: include_aliases - in: query - responses: - '200': - description: Successfully retrieved the properties of the model version - content: - application/json: - schema: - type: object - properties: - model_name: - type: string - catalog_name: - type: string - schema_name: - type: string - comment: - type: string - source: - type: string - run_id: - type: string - run_workspace_id: - type: string - version: - type: integer - status: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - aliases: - type: array - items: - type: object - properties: - alias_name: - type: string - version_num: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get a model version by alias. - delete: - operationId: registered-model-aliases-deletealias - externalDocs: - url: https://docs.databricks.com/api/workspace/registeredmodels/deletealias - x-stackQL-resource: registered_model_aliases - x-stackQL-method: deletealias - x-stackQL-verb: delete - x-numReqParams: 2 - parameters: - - name: full_name - required: true - description: string - in: path - - name: alias - required: true - description: string - in: path - responses: - '200': - description: Successfully deleted the registered model alias - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a registered model alias. - put: - operationId: registered-model-aliases-setalias - externalDocs: - url: https://docs.databricks.com/api/workspace/registeredmodels/setalias - x-stackQL-resource: registered_model_aliases - x-stackQL-method: setalias - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: full_name - required: true - description: string - in: path - - name: alias - required: true - description: string - in: path - responses: - '200': - description: Successfully set the specified alias on the registered model - content: - application/json: - schema: - type: object - properties: - alias_name: - type: string - version_num: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - full_name: - type: required - alias: - type: string - version_num: - type: required - example: - version_num: 2 - description: Set an alias on the specified registered model. - /api/2.1/unity-catalog/models/{full_name}/versions: - get: - operationId: model-versions-list - externalDocs: - url: https://docs.databricks.com/api/workspace/modelversions/list - x-stackQL-resource: model_versions - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - - name: max_results - in: query - - name: page_token - in: query - - name: include_browse - in: query - responses: - '200': - description: The model version list was successfully retrieved - content: - application/json: - schema: - type: object - properties: - model_versions: - type: array - items: - type: object - properties: - model_name: - type: string - catalog_name: - type: string - schema_name: - type: string - comment: - type: string - source: - type: string - run_id: - type: string - run_workspace_id: - type: string - version: - type: integer - status: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.model_versions - description: List model versions. You can list model versions under a particular - schema, or list all model versions in the current metastore. - /api/2.0/online-tables: - post: - operationId: online-tables-create - externalDocs: - url: https://docs.databricks.com/api/workspace/onlinetables/create - x-stackQL-resource: online_tables - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - spec: - type: object - properties: - run_continuously: - type: object - properties: {} - run_triggered: - type: object - properties: {} - source_table_full_name: - type: string - primary_key_columns: - type: array - items: - type: string - timeseries_key: - type: string - perform_full_copy: - type: boolean - pipeline_id: - type: string - status: - type: object - properties: - detailed_state: - type: string - message: - type: string - provisioning_status: - type: object - properties: - initial_pipeline_sync_progress: - type: object - properties: - latest_version_currently_processing: - type: integer - synced_row_count: - type: integer - total_row_count: - type: integer - sync_progress_completion: - type: number - estimated_completion_time_seconds: - type: number - continuous_update_status: - type: object - properties: - last_processed_commit_version: - type: integer - timestamp: - type: string - initial_pipeline_sync_progress: - type: object - properties: - latest_version_currently_processing: - type: integer - synced_row_count: - type: integer - total_row_count: - type: integer - sync_progress_completion: - type: number - estimated_completion_time_seconds: - type: number - triggered_update_status: - type: object - properties: - last_processed_commit_version: - type: integer - timestamp: - type: string - triggered_update_progress: - type: object - properties: - latest_version_currently_processing: - type: integer - synced_row_count: - type: integer - total_row_count: - type: integer - sync_progress_completion: - type: number - estimated_completion_time_seconds: - type: number - failed_status: - type: object - properties: - last_processed_commit_version: - type: integer - timestamp: - type: string - table_serving_url: - type: string - unity_catalog_provisioning_state: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - spec: - type: object - example: - name: string - spec: - run_continuously: {} - run_triggered: {} - source_table_full_name: string - primary_key_columns: - - string - timeseries_key: string - perform_full_copy: false - pipeline_id: string - description: Online Table information. - description: Create a new Online Table. - /api/2.0/online-tables/{name}: - delete: - operationId: online-tables-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/onlinetables/delete - x-stackQL-resource: online_tables - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - description: 'Delete an online table. Warning: This will delete all the data - in the online table. If the source Delta table was deleted or modified since - this Online Table was created, this will lose the data forever!' - get: - operationId: online-tables-get - externalDocs: - url: https://docs.databricks.com/api/workspace/onlinetables/get - x-stackQL-resource: online_tables - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - name: - type: string - spec: - type: object - properties: - run_continuously: - type: object - properties: {} - run_triggered: - type: object - properties: {} - source_table_full_name: - type: string - primary_key_columns: - type: array - items: - type: string - timeseries_key: - type: string - perform_full_copy: - type: boolean - pipeline_id: - type: string - status: - type: object - properties: - detailed_state: - type: string - message: - type: string - provisioning_status: - type: object - properties: - initial_pipeline_sync_progress: - type: object - properties: - latest_version_currently_processing: - type: integer - synced_row_count: - type: integer - total_row_count: - type: integer - sync_progress_completion: - type: number - estimated_completion_time_seconds: - type: number - continuous_update_status: - type: object - properties: - last_processed_commit_version: - type: integer - timestamp: - type: string - initial_pipeline_sync_progress: - type: object - properties: - latest_version_currently_processing: - type: integer - synced_row_count: - type: integer - total_row_count: - type: integer - sync_progress_completion: - type: number - estimated_completion_time_seconds: - type: number - triggered_update_status: - type: object - properties: - last_processed_commit_version: - type: integer - timestamp: - type: string - triggered_update_progress: - type: object - properties: - latest_version_currently_processing: - type: integer - synced_row_count: - type: integer - total_row_count: - type: integer - sync_progress_completion: - type: number - estimated_completion_time_seconds: - type: number - failed_status: - type: object - properties: - last_processed_commit_version: - type: integer - timestamp: - type: string - table_serving_url: - type: string - unity_catalog_provisioning_state: - type: string - description: Get information about an existing online table and its status. - /api/2.1/unity-catalog/tables/{table_name}/monitor: - post: - operationId: quality-monitors-create - externalDocs: - url: https://docs.databricks.com/api/workspace/qualitymonitors/create - x-stackQL-resource: quality_monitors - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 1 - parameters: - - name: table_name - required: true - description: string - in: path - responses: - '200': - description: The new monitor was successfully created. - content: - application/json: - schema: - type: object - properties: - assets_dir: - type: string - output_schema_name: - type: string - inference_log: - type: object - properties: - problem_type: - type: string - timestamp_col: - type: string - granularities: - type: array - items: - type: string - prediction_col: - type: string - label_col: - type: string - model_id_col: - type: string - prediction_proba_col: - type: string - time_series: - type: object - properties: - timestamp_col: - type: string - granularities: - type: array - items: - type: string - snapshot: - type: object - properties: {} - slicing_exprs: - type: array - items: - type: string - custom_metrics: - type: array - items: - type: object - properties: - name: - type: string - definition: - type: string - input_columns: - type: array - items: - type: string - output_data_type: - type: string - type: - type: string - baseline_table_name: - type: string - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - notifications: - type: object - properties: - on_failure: - type: object - properties: - email_addresses: - type: array - items: - type: string - table_name: - type: string - status: - type: string - latest_monitor_failure_msg: - type: string - profile_metrics_table_name: - type: string - drift_metrics_table_name: - type: string - dashboard_id: - type: string - monitor_version: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - skip_builtin_dashboard: - type: boolean - warehouse_id: - type: string - assets_dir: - type: required - output_schema_name: - type: string - snapshot: - type: required - slicing_exprs: - type: string - baseline_table_name: - type: object - inference_log: - type: object - time_series: - type: object - custom_metrics: - type: Array of object - schedule: - type: object - notifications: - type: object - example: - skip_builtin_dashboard: true - warehouse_id: string - assets_dir: string - output_schema_name: string - inference_log: - problem_type: PROBLEM_TYPE_CLASSIFICATION - timestamp_col: string - granularities: - - string - prediction_col: string - label_col: string - model_id_col: string - prediction_proba_col: string - time_series: - timestamp_col: string - granularities: - - string - snapshot: {} - slicing_exprs: - - string - custom_metrics: - - name: string - definition: string - input_columns: - - string - output_data_type: string - type: CUSTOM_METRIC_TYPE_AGGREGATE - baseline_table_name: string - schedule: - quartz_cron_expression: string - timezone_id: string - pause_status: UNPAUSED - notifications: - on_failure: - email_addresses: - - string - description: Creates a new monitor for the specified table. - delete: - operationId: quality-monitors-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/qualitymonitors/delete - x-stackQL-resource: quality_monitors - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: table_name - required: true - description: string - in: path - responses: - '200': - description: The monitor was successfully deleted. - description: Deletes a monitor for the specified table. - get: - operationId: quality-monitors-get - externalDocs: - url: https://docs.databricks.com/api/workspace/qualitymonitors/get - x-stackQL-resource: quality_monitors - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: table_name - required: true - description: string - in: path - responses: - '200': - description: The monitor was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - assets_dir: - type: string - output_schema_name: - type: string - inference_log: - type: object - properties: - problem_type: - type: string - timestamp_col: - type: string - granularities: - type: array - items: - type: string - prediction_col: - type: string - label_col: - type: string - model_id_col: - type: string - prediction_proba_col: - type: string - time_series: - type: object - properties: - timestamp_col: - type: string - granularities: - type: array - items: - type: string - snapshot: - type: object - properties: {} - slicing_exprs: - type: array - items: - type: string - custom_metrics: - type: array - items: - type: object - properties: - name: - type: string - definition: - type: string - input_columns: - type: array - items: - type: string - output_data_type: - type: string - type: - type: string - baseline_table_name: - type: string - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - notifications: - type: object - properties: - on_failure: - type: object - properties: - email_addresses: - type: array - items: - type: string - table_name: - type: string - status: - type: string - latest_monitor_failure_msg: - type: string - profile_metrics_table_name: - type: string - drift_metrics_table_name: - type: string - dashboard_id: - type: string - monitor_version: - type: string - description: Gets a monitor for the specified table. - put: - operationId: quality-monitors-update - externalDocs: - url: https://docs.databricks.com/api/workspace/qualitymonitors/update - x-stackQL-resource: quality_monitors - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: table_name - required: true - description: string - in: path - responses: - '200': - description: The monitor was successfully updated. - content: - application/json: - schema: - type: object - properties: - assets_dir: - type: string - output_schema_name: - type: string - inference_log: - type: object - properties: - problem_type: - type: string - timestamp_col: - type: string - granularities: - type: array - items: - type: string - prediction_col: - type: string - label_col: - type: string - model_id_col: - type: string - prediction_proba_col: - type: string - time_series: - type: object - properties: - timestamp_col: - type: string - granularities: - type: array - items: - type: string - snapshot: - type: object - properties: {} - slicing_exprs: - type: array - items: - type: string - custom_metrics: - type: array - items: - type: object - properties: - name: - type: string - definition: - type: string - input_columns: - type: array - items: - type: string - output_data_type: - type: string - type: - type: string - baseline_table_name: - type: string - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - notifications: - type: object - properties: - on_failure: - type: object - properties: - email_addresses: - type: array - items: - type: string - table_name: - type: string - status: - type: string - latest_monitor_failure_msg: - type: string - profile_metrics_table_name: - type: string - drift_metrics_table_name: - type: string - dashboard_id: - type: string - monitor_version: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - output_schema_name: - type: required - snapshot: - type: string - slicing_exprs: - type: object - baseline_table_name: - type: Array of string - dashboard_id: - type: string - inference_log: - type: object - time_series: - type: object - custom_metrics: - type: Array of object - schedule: - type: object - notifications: - type: object - example: - output_schema_name: string - inference_log: - problem_type: PROBLEM_TYPE_CLASSIFICATION - timestamp_col: string - granularities: - - string - prediction_col: string - label_col: string - model_id_col: string - prediction_proba_col: string - time_series: - timestamp_col: string - granularities: - - string - snapshot: {} - slicing_exprs: - - string - custom_metrics: - - name: string - definition: string - input_columns: - - string - output_data_type: string - type: CUSTOM_METRIC_TYPE_AGGREGATE - baseline_table_name: string - schedule: - quartz_cron_expression: string - timezone_id: string - pause_status: UNPAUSED - notifications: - on_failure: - email_addresses: - - string - dashboard_id: string - description: Updates a monitor for the specified table. - /api/2.1/unity-catalog/tables/{table_name}/monitor/refreshes/{refresh_id}: - get: - operationId: quality-monitor-refreshes-getrefresh - externalDocs: - url: https://docs.databricks.com/api/workspace/qualitymonitors/getrefresh - x-stackQL-resource: quality_monitor_refreshes - x-stackQL-method: getrefresh - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: table_name - required: true - description: string - in: path - - name: refresh_id - required: true - description: string - in: path - responses: - '200': - description: The refresh info was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - refresh_id: - type: integer - state: - type: string - message: - type: string - start_time_ms: - type: integer - end_time_ms: - type: integer - trigger: - type: string - description: Gets info about a specific monitor refresh using the given refresh - ID. - /api/2.1/unity-catalog/tables/{table_name}/monitor/refreshes: - get: - operationId: quality-monitor-refreshes-listrefreshes - externalDocs: - url: https://docs.databricks.com/api/workspace/qualitymonitors/listrefreshes - x-stackQL-resource: quality_monitor_refreshes - x-stackQL-method: listrefreshes - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: table_name - required: true - description: string - in: path - responses: - '200': - description: The list of refresh info was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - refreshes: - type: array - items: - type: object - properties: - refresh_id: - type: integer - state: - type: string - message: - type: string - start_time_ms: - type: integer - end_time_ms: - type: integer - trigger: - type: string - x-stackQL-objectKey: $.refreshes - description: Gets an array containing the history of the most recent refreshes - (up to 25) for this table. - post: - operationId: quality-monitor-refreshes-runrefresh - externalDocs: - url: https://docs.databricks.com/api/workspace/qualitymonitors/runrefresh - x-stackQL-resource: quality_monitor_refreshes - x-stackQL-method: runrefresh - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: table_name - required: true - description: string - in: path - responses: - '200': - description: The metric refresh was successfully queued. - content: - application/json: - schema: - type: object - properties: - refresh_id: - type: integer - state: - type: string - message: - type: string - start_time_ms: - type: integer - end_time_ms: - type: integer - trigger: - type: string - description: Queues a metric refresh on the monitor for the specified table. - The refresh will execute in the background. - /api/2.1/unity-catalog/models: - post: - operationId: registered-models-create - externalDocs: - url: https://docs.databricks.com/api/workspace/registeredmodels/create - x-stackQL-resource: registered_models - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Successfully created the registered model - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - full_name: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - securable_type: - type: string - securable_kind: - type: string - comment: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - catalog_name: - type: required - schema_name: - type: string - name: - type: required - storage_location: - type: string - comment: - type: required - example: - name: revenue_forecasting_model - catalog_name: main - schema_name: default - comment: This model contains model versions that forecast future revenue, - given historical data - description: Creates a new registered model in Unity Catalog. - get: - operationId: registered-models-list - externalDocs: - url: https://docs.databricks.com/api/workspace/registeredmodels/list - x-stackQL-resource: registered_models - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: catalog_name - in: query - - name: schema_name - in: query - - name: max_results - in: query - - name: page_token - in: query - - name: include_browse - in: query - responses: - '200': - description: The registered model list was successfully retrieved - content: - application/json: - schema: - type: object - properties: - registered_models: - type: array - items: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - full_name: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - securable_type: - type: string - securable_kind: - type: string - comment: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.registered_models - description: List registered models. You can list registered models under a - particular schema, or list all registered models in the current metastore. - /api/2.1/unity-catalog/models/{full_name}: - delete: - operationId: registered-models-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/registeredmodels/delete - x-stackQL-resource: registered_models - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - responses: - '200': - description: Successfully deleted the registered model - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a registered model and all its model versions from the - specified parent catalog and schema. - get: - operationId: registered-models-get - externalDocs: - url: https://docs.databricks.com/api/workspace/registeredmodels/get - x-stackQL-resource: registered_models - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - - name: include_browse - in: query - - name: include_aliases - in: query - responses: - '200': - description: Successfully retrieved the properties of the registered model - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - full_name: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - securable_type: - type: string - securable_kind: - type: string - comment: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get a registered model. - patch: - operationId: registered-models-update - externalDocs: - url: https://docs.databricks.com/api/workspace/registeredmodels/update - x-stackQL-resource: registered_models - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - responses: - '200': - description: Successfully updated the properties of the registered model - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - full_name: - type: string - owner: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - securable_type: - type: string - securable_kind: - type: string - comment: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - comment: - type: string - new_name: - type: string - example: - new_name: my_new_model - description: Updates the specified registered model. - /api/2.1/unity-catalog/resource-quotas/{parent_securable_type}/{parent_full_name}/{quota_name}: - get: - operationId: resource-quotas-getquota - externalDocs: - url: https://docs.databricks.com/api/workspace/resourcequotas/getquota - x-stackQL-resource: resource_quotas - x-stackQL-method: getquota - x-stackQL-verb: select - x-numReqParams: 3 - parameters: - - name: parent_securable_type - required: true - description: string - in: path - - name: parent_full_name - required: true - description: string - in: path - - name: quota_name - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - quota_info: - type: object - properties: - parent_securable_type: - type: string - parent_full_name: - type: string - quota_name: - type: string - quota_count: - type: integer - quota_limit: - type: integer - last_refreshed_at: - type: integer - x-stackQL-objectKey: $.quota_info - description: The GetQuota API returns usage information for a single resource - quota, defined as a child-parent pair. This API also refreshes the quota count - if it is out of date. Refreshes are triggered asynchronously. The updated - count might not be returned in the first call. - /api/2.1/unity-catalog/resource-quotas/all-resource-quotas: - get: - operationId: resource-quotas-listquotas - externalDocs: - url: https://docs.databricks.com/api/workspace/resourcequotas/listquotas - x-stackQL-resource: resource_quotas - x-stackQL-method: listquotas - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - quotas: - type: array - items: - type: object - properties: - parent_securable_type: - type: string - parent_full_name: - type: string - quota_name: - type: string - quota_count: - type: integer - quota_limit: - type: integer - last_refreshed_at: - type: integer - next_page_token: - type: string - x-stackQL-objectKey: $.quotas - description: ListQuotas returns all quota values under the metastore. There - are no SLAs on the freshness of the counts returned. This API does not trigger - a refresh of quota counts. - /api/2.1/unity-catalog/schemas: - post: - operationId: schemas-create - externalDocs: - url: https://docs.databricks.com/api/workspace/schemas/create - x-stackQL-resource: schemas - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new schema was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - full_name: - type: string - storage_location: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - browse_only: - type: boolean - schema_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - catalog_name: - type: string - comment: - type: required - storage_root: - type: string - properties: - type: object - example: - name: string - catalog_name: string - comment: string - properties: - property1: string - property2: string - storage_root: string - description: Creates a new schema for catalog in the Metatastore. The caller - must be a metastore admin, or have the - get: - operationId: schemas-list - externalDocs: - url: https://docs.databricks.com/api/workspace/schemas/list - x-stackQL-resource: schemas - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: catalog_name - required: true - description: string - in: query - - name: max_results - in: query - - name: page_token - in: query - - name: include_browse - in: query - responses: - '200': - description: The schemas list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: object - properties: - name: - type: string - catalog_name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - full_name: - type: string - storage_location: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - browse_only: - type: boolean - schema_id: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.schemas - description: Gets an array of schemas for a catalog in the metastore. If the - caller is the metastore admin or the owner of the parent catalog, all schemas - for the catalog will be retrieved. Otherwise, only schemas owned by the caller - (or for which the caller has the - /api/2.1/unity-catalog/schemas/{full_name}: - delete: - operationId: schemas-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/schemas/delete - x-stackQL-resource: schemas - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: The schema was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes the specified schema from the parent catalog. The caller - must be the owner of the schema or an owner of the parent catalog. - get: - operationId: schemas-get - externalDocs: - url: https://docs.databricks.com/api/workspace/schemas/get - x-stackQL-resource: schemas - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - - name: include_browse - in: query - responses: - '200': - description: The schema was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - full_name: - type: string - storage_location: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - browse_only: - type: boolean - schema_id: - type: string - description: Gets the specified schema within the metastore. The caller must - be a metastore admin, the owner of the schema, or a user that has the - patch: - operationId: schemas-update - externalDocs: - url: https://docs.databricks.com/api/workspace/schemas/update - x-stackQL-resource: schemas - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - responses: - '200': - description: The schema was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_root: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - full_name: - type: string - storage_location: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - catalog_type: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - browse_only: - type: boolean - schema_id: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - comment: - type: string - enable_predictive_optimization: - type: string - new_name: - type: string - properties: - type: object - example: - owner: string - comment: string - properties: - property1: string - property2: string - enable_predictive_optimization: DISABLE - new_name: string - description: Updates a schema for a catalog. The caller must be the owner of - the schema or a metastore admin. If the caller is a metastore admin, only - the - /api/2.1/unity-catalog/storage-credentials: - post: - operationId: storage-credentials-create - externalDocs: - url: https://docs.databricks.com/api/workspace/storagecredentials/create - x-stackQL-resource: storage_credentials - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new storage credential was successfully created. - content: - application/json: - schema: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - comment: - type: string - read_only: - type: string - skip_validation: - type: boolean - aws_iam_role: - type: object - example: - name: string - comment: string - read_only: true - aws_iam_role: - role_arn: string - skip_validation: false - description: Creates a new storage credential. - get: - operationId: storage-credentials-list - externalDocs: - url: https://docs.databricks.com/api/workspace/storagecredentials/list - x-stackQL-resource: storage_credentials - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The storage credentials list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - storage_credentials: - type: array - items: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.storage_credentials - description: Gets an array of storage credentials (as - /api/2.1/unity-catalog/storage-credentials/{name}: - delete: - operationId: storage-credentials-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/storagecredentials/delete - x-stackQL-resource: storage_credentials - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: force - in: query - responses: - '200': - description: The storage credential was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a storage credential from the metastore. The caller must - be an owner of the storage credential. - get: - operationId: storage-credentials-get - externalDocs: - url: https://docs.databricks.com/api/workspace/storagecredentials/get - x-stackQL-resource: storage_credentials - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The storage credential was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - description: Gets a storage credential from the metastore. The caller must be - a metastore admin, the owner of the storage credential, or have some permission - on the storage credential. - patch: - operationId: storage-credentials-update - externalDocs: - url: https://docs.databricks.com/api/workspace/storagecredentials/update - x-stackQL-resource: storage_credentials - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: The storage credential was successfully updated. - content: - application/json: - schema: - type: object - properties: - name: - type: string - aws_iam_role: - type: object - properties: - role_arn: - type: string - unity_catalog_iam_arn: - type: string - external_id: - type: string - comment: - type: string - read_only: - type: boolean - owner: - type: string - full_name: - type: string - id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - used_for_managed_storage: - type: boolean - isolation_mode: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - comment: - type: string - read_only: - type: boolean - owner: - type: string - isolation_mode: - type: string - new_name: - type: string - skip_validation: - type: boolean - force: - type: boolean - aws_iam_role: - type: object - example: - comment: string - read_only: true - owner: string - aws_iam_role: - role_arn: string - isolation_mode: ISOLATION_MODE_OPEN - new_name: string - skip_validation: false - force: true - description: Updates a storage credential on the metastore. - /api/2.1/unity-catalog/validate-storage-credentials: - post: - operationId: storage-credential-validation-validate - externalDocs: - url: https://docs.databricks.com/api/workspace/storagecredentials/validate - x-stackQL-resource: storage_credential_validation - x-stackQL-method: validate - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The storage credential validation operation completed successfully. - content: - application/json: - schema: - type: object - properties: - isDir: - type: boolean - results: - type: array - items: - type: object - properties: - operation: - type: string - result: - type: string - message: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - storage_credential_name: - type: string - external_location_name: - type: string - url: - type: string - read_only: - type: boolean - aws_iam_role: - type: object - example: - storage_credential_name: string - aws_iam_role: - role_arn: string - external_location_name: string - url: string - read_only: true - description: Validates a storage credential. At least one of - /api/2.1/unity-catalog/metastores/{metastore_id}/systemschemas/{schema_name}: - delete: - operationId: system-schemas-disable - externalDocs: - url: https://docs.databricks.com/api/workspace/systemschemas/disable - x-stackQL-resource: system_schemas - x-stackQL-method: disable - x-stackQL-verb: exec - x-numReqParams: 2 - parameters: - - name: metastore_id - required: true - description: string - in: path - - name: schema_name - required: true - description: string - in: path - responses: - '200': - description: The schema was successfully disabled. - description: Disables the system schema and removes it from the system catalog. - The caller must be an account admin or a metastore admin. - put: - operationId: system-schemas-enable - externalDocs: - url: https://docs.databricks.com/api/workspace/systemschemas/enable - x-stackQL-resource: system_schemas - x-stackQL-method: enable - x-stackQL-verb: exec - x-numReqParams: 2 - parameters: - - name: metastore_id - required: true - description: string - in: path - - name: schema_name - required: true - description: string - in: path - responses: - '200': - description: The system schema was successfully enabled. - description: Enables the system schema and adds it to the system catalog. The - caller must be an account admin or a metastore admin. - /api/2.1/unity-catalog/metastores/{metastore_id}/systemschemas: - get: - operationId: system-schemas-list - externalDocs: - url: https://docs.databricks.com/api/workspace/systemschemas/list - x-stackQL-resource: system_schemas - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: metastore_id - required: true - description: string - in: path - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: The schemas list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - schemas: - type: array - items: - type: object - properties: - schema: - type: string - state: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.schemas - description: Gets an array of system schemas for a metastore. The caller must - be an account admin or a metastore admin. - /api/2.1/unity-catalog/tables/{full_name}: - delete: - operationId: tables-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/tables/delete - x-stackQL-resource: tables - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - responses: - '200': - description: The table was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a table from the specified parent catalog and schema. The - caller must be the owner of the parent catalog, have the - get: - operationId: tables-get - externalDocs: - url: https://docs.databricks.com/api/workspace/tables/get - x-stackQL-resource: tables - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - - name: include_delta_metadata - in: query - - name: include_browse - in: query - - name: include_manifest_capabilities - in: query - responses: - '200': - description: The table was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - table_type: - type: string - data_source_format: - type: string - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - storage_location: - type: string - view_definition: - type: string - view_dependencies: - type: object - properties: - dependencies: - type: array - items: - type: object - properties: - table: - type: object - properties: - table_full_name: - type: string - function: - type: object - properties: - function_full_name: - type: string - sql_path: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_credential_name: - type: string - table_constraints: - type: array - items: - type: object - properties: - primary_key_constraint: - type: object - properties: - name: - type: string - child_columns: - type: array - items: - type: string - foreign_key_constraint: - type: object - properties: - name: - type: string - child_columns: - type: array - items: - type: string - parent_table: - type: string - parent_columns: - type: array - items: - type: string - named_table_constraint: - type: object - properties: - name: - type: string - row_filter: - type: object - properties: - function_name: - type: string - input_column_names: - type: array - items: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - full_name: - type: string - data_access_configuration_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - deleted_at: - type: integer - table_id: - type: string - delta_runtime_properties_kvpairs: - type: object - properties: - delta_runtime_properties: - type: object - properties: - property1: - type: string - property2: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - access_point: - type: string - pipeline_id: - type: string - browse_only: - type: boolean - description: 'Gets a table from the metastore for a specific catalog and schema. - The caller must satisfy one of the following requirements:' - /api/2.1/unity-catalog/tables: - get: - operationId: tables-list - externalDocs: - url: https://docs.databricks.com/api/workspace/tables/list - x-stackQL-resource: tables - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: catalog_name - required: true - description: string - in: query - - name: schema_name - required: true - description: string - in: query - - name: max_results - in: query - - name: page_token - in: query - - name: include_delta_metadata - in: query - - name: omit_columns - in: query - - name: omit_properties - in: query - - name: omit_username - in: query - - name: include_browse - in: query - - name: include_manifest_capabilities - in: query - responses: - '200': - description: The tables list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - tables: - type: array - items: - type: object - properties: - name: - type: string - catalog_name: - type: string - schema_name: - type: string - table_type: - type: string - data_source_format: - type: string - columns: - type: array - items: - type: object - properties: - name: - type: string - type_text: - type: string - type_name: - type: string - position: - type: integer - type_precision: - type: integer - type_scale: - type: integer - type_interval_type: - type: string - type_json: - type: string - comment: - type: string - nullable: - type: boolean - partition_index: - type: integer - mask: - type: object - properties: - function_name: - type: string - using_column_names: - type: array - items: - type: string - storage_location: - type: string - view_definition: - type: string - view_dependencies: - type: object - properties: - dependencies: - type: array - items: - type: object - properties: - table: - type: object - properties: - table_full_name: - type: string - function: - type: object - properties: - function_full_name: - type: string - sql_path: - type: string - owner: - type: string - comment: - type: string - properties: - type: object - properties: - property1: - type: string - property2: - type: string - storage_credential_name: - type: string - table_constraints: - type: array - items: - type: object - properties: - primary_key_constraint: - type: object - properties: - name: - type: string - child_columns: - type: array - items: - type: string - foreign_key_constraint: - type: object - properties: - name: - type: string - child_columns: - type: array - items: - type: string - parent_table: - type: string - parent_columns: - type: array - items: - type: string - named_table_constraint: - type: object - properties: - name: - type: string - row_filter: - type: object - properties: - function_name: - type: string - input_column_names: - type: array - items: - type: string - enable_predictive_optimization: - type: string - metastore_id: - type: string - full_name: - type: string - data_access_configuration_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - deleted_at: - type: integer - table_id: - type: string - delta_runtime_properties_kvpairs: - type: object - properties: - delta_runtime_properties: - type: object - properties: - property1: - type: string - property2: - type: string - effective_predictive_optimization_flag: - type: object - properties: - value: - type: string - inherited_from_type: - type: string - inherited_from_name: - type: string - access_point: - type: string - pipeline_id: - type: string - browse_only: - type: boolean - next_page_token: - type: string - x-stackQL-objectKey: $.tables - description: Gets an array of all tables for the current metastore under the - parent catalog and schema. The caller must be a metastore admin or an owner - of (or have the - /api/2.1/unity-catalog/constraints: - post: - operationId: table-constraints-create - externalDocs: - url: https://docs.databricks.com/api/workspace/tableconstraints/create - x-stackQL-resource: table_constraints - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: The new table constraint was successfully created. - content: - application/json: - schema: - type: object - properties: - primary_key_constraint: - type: object - properties: - name: - type: string - child_columns: - type: array - items: - type: string - foreign_key_constraint: - type: object - properties: - name: - type: string - child_columns: - type: array - items: - type: string - parent_table: - type: string - parent_columns: - type: array - items: - type: string - named_table_constraint: - type: object - properties: - name: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - full_name_arg: - type: required - constraint: - type: required - example: - full_name_arg: string - constraint: - primary_key_constraint: - name: string - child_columns: - - string - foreign_key_constraint: - name: string - child_columns: - - string - parent_table: string - parent_columns: - - string - named_table_constraint: - name: string - description: Creates a new table constraint. - /api/2.1/unity-catalog/constraints/{full_name}: - delete: - operationId: table-constraints-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/tableconstraints/delete - x-stackQL-resource: table_constraints - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 3 - parameters: - - name: full_name - required: true - description: string - in: path - - name: constraint_name - required: true - description: string - in: query - - name: cascade - required: true - description: boolean - in: query - responses: - '200': - description: The table constraint was successfully deleted. - content: - application/json: - schema: - type: object - properties: {} - description: Deletes a table constraint. - /api/2.1/unity-catalog/tables/{full_name}/exists: - get: - operationId: table-exists-exists - externalDocs: - url: https://docs.databricks.com/api/workspace/tables/exists - x-stackQL-resource: table_exists - x-stackQL-method: exists - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: full_name - required: true - description: string - in: path - responses: - '200': - description: The boolean reflecting if the table exists was successfully - returned. - content: - application/json: - schema: - type: object - properties: - table_exists: - type: boolean - description: 'Gets if a table exists in the metastore for a specific catalog - and schema. The caller must satisfy one of the following requirements:' - /api/2.1/unity-catalog/table-summaries: - get: - operationId: table-summaries-listsummaries - externalDocs: - url: https://docs.databricks.com/api/workspace/tables/listsummaries - x-stackQL-resource: table_summaries - x-stackQL-method: listsummaries - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: catalog_name - required: true - description: string - in: query - - name: schema_name_pattern - in: query - - name: table_name_pattern - in: query - - name: max_results - in: query - - name: page_token - in: query - - name: include_manifest_capabilities - in: query - responses: - '200': - description: The table summaries list was successfully retrieved. - content: - application/json: - schema: - type: object - properties: - tables: - type: array - items: - type: object - properties: - full_name: - type: string - table_type: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.tables - description: 'Gets an array of summaries for tables for a schema and catalog - within the metastore. The table summaries returned are either:' - /api/2.0/unity-catalog/temporary-table-credentials: - post: - operationId: temporary-table-credentials-generatetemporarytablecredentials - externalDocs: - url: https://docs.databricks.com/api/workspace/temporarytablecredentials/generatetemporarytablecredentials - x-stackQL-resource: temporary_table_credentials - x-stackQL-method: generatetemporarytablecredentials - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - aws_temp_credentials: - type: object - properties: - access_key_id: - type: string - secret_access_key: - type: string - session_token: - type: string - access_point: - type: string - azure_user_delegation_sas: - type: object - properties: - sas_token: - type: string - gcp_oauth_token: - type: object - properties: - oauth_token: - type: string - r2_temp_credentials: - type: object - properties: - access_key_id: - type: string - secret_access_key: - type: string - session_token: - type: string - expiration_time: - type: integer - url: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - table_id: - type: string - operation: - type: string - example: - table_id: string - operation: READ - description: Get a short-lived credential for directly accessing the table data - on cloud storage. The metastore must have external_access_enabled flag set - to true (default false). The caller must have EXTERNAL_USE_SCHEMA privilege - on the parent schema and this privilege can only be granted by catalog owners. - /api/2.1/unity-catalog/volumes: - post: - operationId: volumes-create - externalDocs: - url: https://docs.databricks.com/api/workspace/volumes/create - x-stackQL-resource: volumes - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Successfully created the volume - content: - application/json: - schema: - type: object - properties: - catalog_name: - type: string - schema_name: - type: string - name: - type: string - full_name: - type: string - volume_type: - type: string - owner: - type: string - volume_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - comment: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - catalog_name: - type: required - schema_name: - type: string - name: - type: required - volume_type: - type: string - storage_location: - type: required - comment: - type: string - example: - catalog_name: main - schema_name: default - name: my_volume - volume_type: EXTERNAL - storage_location: s3://my-bucket/hello/world/my-volume - comment: This is my first volume - description: Creates a new volume. - get: - operationId: volumes-list - externalDocs: - url: https://docs.databricks.com/api/workspace/volumes/list - x-stackQL-resource: volumes - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: catalog_name - required: true - description: string - in: query - - name: schema_name - required: true - description: string - in: query - - name: max_results - in: query - - name: page_token - in: query - - name: include_browse - in: query - responses: - '200': - description: The volume list was successfully retrieved - content: - application/json: - schema: - type: object - properties: - volumes: - type: array - items: - type: object - properties: - catalog_name: - type: string - schema_name: - type: string - name: - type: string - full_name: - type: string - volume_type: - type: string - owner: - type: string - volume_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - comment: - type: string - x-stackQL-objectKey: $.volumes - description: Gets an array of volumes for the current metastore under the parent - catalog and schema. - /api/2.1/unity-catalog/volumes/{name}: - delete: - operationId: volumes-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/volumes/delete - x-stackQL-resource: volumes - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Successfully deleted the volume - description: Deletes a volume from the specified parent catalog and schema. - get: - operationId: volumes-read - externalDocs: - url: https://docs.databricks.com/api/workspace/volumes/read - x-stackQL-resource: volumes - x-stackQL-method: read - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - - name: include_browse - in: query - responses: - '200': - description: Successfully retrieved the properties of the volume - content: - application/json: - schema: - type: object - properties: - catalog_name: - type: string - schema_name: - type: string - name: - type: string - full_name: - type: string - volume_type: - type: string - owner: - type: string - volume_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - comment: - type: string - description: Gets a volume from the metastore for a specific catalog and schema. - patch: - operationId: volumes-update - externalDocs: - url: https://docs.databricks.com/api/workspace/volumes/update - x-stackQL-resource: volumes - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: Successfully updated the properties of the volume - content: - application/json: - schema: - type: object - properties: - catalog_name: - type: string - schema_name: - type: string - name: - type: string - full_name: - type: string - volume_type: - type: string - owner: - type: string - volume_id: - type: string - metastore_id: - type: string - created_at: - type: integer - created_by: - type: string - updated_at: - type: integer - updated_by: - type: string - storage_location: - type: string - comment: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - owner: - type: string - comment: - type: string - new_name: - type: string - example: - new_name: my_new_volume - description: Updates the specified volume under the specified parent catalog - and schema. - /api/2.1/unity-catalog/bindings/{securable_type}/{securable_name}: - get: - operationId: workspace-bindings-getbindings - externalDocs: - url: https://docs.databricks.com/api/workspace/workspacebindings/getbindings - x-stackQL-resource: workspace_bindings - x-stackQL-method: getbindings - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: securable_type - required: true - description: string - in: path - - name: securable_name - required: true - description: string - in: path - - name: max_results - in: query - - name: page_token - in: query - responses: - '200': - description: A list of workspace IDs that are bound to the securable - content: - application/json: - schema: - type: object - properties: - bindings: - type: array - items: - type: object - properties: - workspace_id: - type: integer - binding_type: - type: string - next_page_token: - type: string - x-stackQL-objectKey: $.bindings - description: Gets workspace bindings of the securable. The caller must be a - metastore admin or an owner of the securable. - patch: - operationId: workspace-bindings-updatebindings - externalDocs: - url: https://docs.databricks.com/api/workspace/workspacebindings/updatebindings - x-stackQL-resource: workspace_bindings - x-stackQL-method: updatebindings - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: securable_type - required: true - description: string - in: path - - name: securable_name - required: true - description: string - in: path - responses: - '200': - description: A list of workspace bindings that are bound to the securable. - content: - application/json: - schema: - type: object - properties: - bindings: - type: array - items: - type: object - properties: - workspace_id: - type: integer - binding_type: - type: string - next_page_token: - type: string - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - add: - type: Array of object - remove: - type: Array of object - example: - add: - - workspace_id: 0 - binding_type: BINDING_TYPE_READ_WRITE - remove: - - workspace_id: 0 - binding_type: BINDING_TYPE_READ_WRITE - description: Updates workspace bindings of the securable. The caller must be - a metastore admin or an owner of the securable. - /api/2.1/unity-catalog/workspace-bindings/catalogs/{name}: - patch: - operationId: workspace-bindings-update - externalDocs: - url: https://docs.databricks.com/api/workspace/workspacebindings/update - x-stackQL-resource: workspace_bindings - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: name - required: true - description: string - in: path - responses: - '200': - description: A list of workspace IDs that are bound to the catalog. - content: - application/json: - schema: - type: object - properties: - workspaces: - type: array - items: - type: integer - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - assign_workspaces: - type: Array of int64 - unassign_workspaces: - type: Array of int64 - example: - assign_workspaces: - - 0 - unassign_workspaces: - - 0 - description: Updates workspace bindings of the catalog. The caller must be a - metastore admin or an owner of the catalog. -components: - x-stackQL-resources: - artifact_allow_lists: - id: databricks_workspace.unitycatalog.artifact_allow_lists - name: artifact_allow_lists - title: Artifact_allow_lists - methods: - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1artifact-allowlists~1{artifact_type}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1artifact-allowlists~1{artifact_type}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/artifact_allow_lists/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/artifact_allow_lists/methods/update' - replace: [] - delete: [] - exec: [] - catalogs: - id: databricks_workspace.unitycatalog.catalogs - name: catalogs - title: Catalogs - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.catalogs - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1catalogs~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/catalogs/methods/get' - - $ref: '#/components/x-stackQL-resources/catalogs/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/catalogs/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/catalogs/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/catalogs/methods/delete' - exec: [] - connections: - id: databricks_workspace.unitycatalog.connections - name: connections - title: Connections - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1connections/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1connections/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.connections - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1connections~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1connections~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1connections~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/connections/methods/get' - - $ref: '#/components/x-stackQL-resources/connections/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/connections/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/connections/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/connections/methods/delete' - exec: [] - credentials: - id: databricks_workspace.unitycatalog.credentials - name: credentials - title: Credentials - methods: - createcredential: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listcredentials: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.credentials - deletecredential: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials~1{name_arg}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getcredential: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials~1{name_arg}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatecredential: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1credentials~1{name_arg}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - generatetemporaryservicecredential: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1temporary-service-credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/credentials/methods/generatetemporaryservicecredential' - - $ref: '#/components/x-stackQL-resources/credentials/methods/getcredential' - - $ref: '#/components/x-stackQL-resources/credentials/methods/listcredentials' - insert: - - $ref: '#/components/x-stackQL-resources/credentials/methods/createcredential' - update: - - $ref: '#/components/x-stackQL-resources/credentials/methods/updatecredential' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/credentials/methods/deletecredential' - exec: [] - credential_validation: - id: databricks_workspace.unitycatalog.credential_validation - name: credential_validation - title: Credential_validation - methods: - validatecredential: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1validate-credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/credential_validation/methods/validatecredential' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - effective_grants: - id: databricks_workspace.unitycatalog.effective_grants - name: effective_grants - title: Effective_grants - methods: - geteffective: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1effective-permissions~1{securable_type}~1{full_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.privilege_assignments - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/effective_grants/methods/geteffective' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - external_locations: - id: databricks_workspace.unitycatalog.external_locations - name: external_locations - title: External_locations - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.external_locations - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1external-locations~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/external_locations/methods/get' - - $ref: '#/components/x-stackQL-resources/external_locations/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/external_locations/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/external_locations/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/external_locations/methods/delete' - exec: [] - functions: - id: databricks_workspace.unitycatalog.functions - name: functions - title: Functions - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1functions/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1functions/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.functions - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1functions~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1functions~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1functions~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/functions/methods/get' - - $ref: '#/components/x-stackQL-resources/functions/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/functions/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/functions/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/functions/methods/delete' - exec: [] - grants: - id: databricks_workspace.unitycatalog.grants - name: grants - title: Grants - methods: - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1permissions~1{securable_type}~1{full_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.privilege_assignments - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1permissions~1{securable_type}~1{full_name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/grants/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/grants/methods/update' - replace: [] - delete: [] - exec: [] - metastores: - id: databricks_workspace.unitycatalog.metastores - name: metastores - title: Metastores - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.metastores - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/metastores/methods/get' - - $ref: '#/components/x-stackQL-resources/metastores/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/metastores/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/metastores/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/metastores/methods/delete' - exec: [] - metastore_assignments: - id: databricks_workspace.unitycatalog.metastore_assignments - name: metastore_assignments - title: Metastore_assignments - methods: - assign: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1workspaces~1{workspace_id}~1metastore/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - unassign: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1workspaces~1{workspace_id}~1metastore/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updateassignment: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1workspaces~1{workspace_id}~1metastore/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - current: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1current-metastore-assignment/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/current' - insert: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/assign' - update: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/updateassignment' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/metastore_assignments/methods/unassign' - exec: [] - metastore_summary: - id: databricks_workspace.unitycatalog.metastore_summary - name: metastore_summary - title: Metastore_summary - methods: - summary: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastore_summary/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/metastore_summary/methods/summary' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - model_versions: - id: databricks_workspace.unitycatalog.model_versions - name: model_versions - title: Model_versions - methods: - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions~1{version}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions~1{version}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions~1{version}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getbyalias: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1aliases~1{alias}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1versions/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.model_versions - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/model_versions/methods/list' - - $ref: '#/components/x-stackQL-resources/model_versions/methods/getbyalias' - - $ref: '#/components/x-stackQL-resources/model_versions/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/model_versions/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/model_versions/methods/delete' - exec: [] - registered_model_aliases: - id: databricks_workspace.unitycatalog.registered_model_aliases - name: registered_model_aliases - title: Registered_model_aliases - methods: - deletealias: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1aliases~1{alias}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setalias: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}~1aliases~1{alias}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/registered_model_aliases/methods/setalias' - delete: - - $ref: '#/components/x-stackQL-resources/registered_model_aliases/methods/deletealias' - exec: [] - online_tables: - id: databricks_workspace.unitycatalog.online_tables - name: online_tables - title: Online_tables - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1online-tables/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1online-tables~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1online-tables~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/online_tables/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/online_tables/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/online_tables/methods/delete' - exec: [] - quality_monitors: - id: databricks_workspace.unitycatalog.quality_monitors - name: quality_monitors - title: Quality_monitors - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/quality_monitors/methods/delete' - exec: [] - quality_monitor_refreshes: - id: databricks_workspace.unitycatalog.quality_monitor_refreshes - name: quality_monitor_refreshes - title: Quality_monitor_refreshes - methods: - getrefresh: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor~1refreshes~1{refresh_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listrefreshes: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor~1refreshes/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.refreshes - runrefresh: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{table_name}~1monitor~1refreshes/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/quality_monitor_refreshes/methods/listrefreshes' - - $ref: '#/components/x-stackQL-resources/quality_monitor_refreshes/methods/getrefresh' - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/quality_monitor_refreshes/methods/runrefresh' - registered_models: - id: databricks_workspace.unitycatalog.registered_models - name: registered_models - title: Registered_models - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.registered_models - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1models~1{full_name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/registered_models/methods/get' - - $ref: '#/components/x-stackQL-resources/registered_models/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/registered_models/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/registered_models/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/registered_models/methods/delete' - exec: [] - resource_quotas: - id: databricks_workspace.unitycatalog.resource_quotas - name: resource_quotas - title: Resource_quotas - methods: - getquota: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1resource-quotas~1{parent_securable_type}~1{parent_full_name}~1{quota_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.quota_info - listquotas: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1resource-quotas~1all-resource-quotas/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.quotas - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/resource_quotas/methods/listquotas' - - $ref: '#/components/x-stackQL-resources/resource_quotas/methods/getquota' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - schemas: - id: databricks_workspace.unitycatalog.schemas - name: schemas - title: Schemas - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.schemas - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas~1{full_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas~1{full_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1schemas~1{full_name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/schemas/methods/get' - - $ref: '#/components/x-stackQL-resources/schemas/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/schemas/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/schemas/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/schemas/methods/delete' - exec: [] - storage_credentials: - id: databricks_workspace.unitycatalog.storage_credentials - name: storage_credentials - title: Storage_credentials - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.storage_credentials - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1storage-credentials~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/get' - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/storage_credentials/methods/delete' - exec: [] - storage_credential_validation: - id: databricks_workspace.unitycatalog.storage_credential_validation - name: storage_credential_validation - title: Storage_credential_validation - methods: - validate: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1validate-storage-credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/storage_credential_validation/methods/validate' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - system_schemas: - id: databricks_workspace.unitycatalog.system_schemas - name: system_schemas - title: System_schemas - methods: - disable: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{metastore_id}~1systemschemas~1{schema_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - enable: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{metastore_id}~1systemschemas~1{schema_name}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1metastores~1{metastore_id}~1systemschemas/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.schemas - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/system_schemas/methods/list' - insert: [] - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/system_schemas/methods/enable' - - $ref: '#/components/x-stackQL-resources/system_schemas/methods/disable' - tables: - id: databricks_workspace.unitycatalog.tables - name: tables - title: Tables - methods: - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{full_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{full_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.tables - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/tables/methods/list' - - $ref: '#/components/x-stackQL-resources/tables/methods/get' - insert: [] - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/tables/methods/delete' - exec: [] - table_constraints: - id: databricks_workspace.unitycatalog.table_constraints - name: table_constraints - title: Table_constraints - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1constraints/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1constraints~1{full_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: - - $ref: '#/components/x-stackQL-resources/table_constraints/methods/create' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/table_constraints/methods/delete' - exec: [] - table_exists: - id: databricks_workspace.unitycatalog.table_exists - name: table_exists - title: Table_exists - methods: - exists: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1tables~1{full_name}~1exists/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/table_exists/methods/exists' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - table_summaries: - id: databricks_workspace.unitycatalog.table_summaries - name: table_summaries - title: Table_summaries - methods: - listsummaries: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1table-summaries/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.tables - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/table_summaries/methods/listsummaries' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - temporary_table_credentials: - id: databricks_workspace.unitycatalog.temporary_table_credentials - name: temporary_table_credentials - title: Temporary_table_credentials - methods: - generatetemporarytablecredentials: - operation: - $ref: '#/paths/~1api~12.0~1unity-catalog~1temporary-table-credentials/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/temporary_table_credentials/methods/generatetemporarytablecredentials' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - volumes: - id: databricks_workspace.unitycatalog.volumes - name: volumes - title: Volumes - methods: - create: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.volumes - delete: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes~1{name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - read: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes~1{name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1volumes~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/volumes/methods/read' - - $ref: '#/components/x-stackQL-resources/volumes/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/volumes/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/volumes/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/volumes/methods/delete' - exec: [] - workspace_bindings: - id: databricks_workspace.unitycatalog.workspace_bindings - name: workspace_bindings - title: Workspace_bindings - methods: - getbindings: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1bindings~1{securable_type}~1{securable_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.bindings - updatebindings: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1bindings~1{securable_type}~1{securable_name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.1~1unity-catalog~1workspace-bindings~1catalogs~1{name}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/workspace_bindings/methods/getbindings' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/workspace_bindings/methods/update' - - $ref: '#/components/x-stackQL-resources/workspace_bindings/methods/updatebindings' - replace: [] - delete: [] - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/vectorsearch.yaml b/providers/src/databricks_workspace/v00.00.00000/services/vectorsearch.yaml index 0aa67f26..780ebcc2 100644 --- a/providers/src/databricks_workspace/v00.00.00000/services/vectorsearch.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/services/vectorsearch.yaml @@ -1,982 +1,1730 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Vectorsearch API - description: Vectorsearch -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/vector-search/indexes/{index_name}/delete-data: - post: - operationId: data-vector-indexes-deletedatavectorindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/deletedatavectorindex - x-stackQL-resource: data_vector_indexes - x-stackQL-method: deletedatavectorindex - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Response indicating the result of the delete operation. - content: - application/json: - schema: - type: object - properties: - status: - type: string - result: - type: object - properties: - success_row_count: - type: integer - failed_primary_keys: - type: array - items: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - primary_keys: - type: required - example: - primary_keys: - - '1' - description: Data to be deleted from the vector index. - description: Handles the deletion of data from a specified vector index. - /api/2.0/vector-search/indexes/{index_name}/upsert-data: - post: - operationId: data-vector-indexes-upsertdatavectorindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/upsertdatavectorindex - x-stackQL-resource: data_vector_indexes - x-stackQL-method: upsertdatavectorindex - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Response indicating the result of the upsert operation. - content: - application/json: - schema: - type: object - properties: - status: - type: string - result: - type: object - properties: - success_row_count: - type: integer - failed_primary_keys: - type: array - items: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - inputs_json: - type: required - example: - inputs_json: '[{"id": "1", "text": "hello world", "text_vector": [1.0, - 2.0, 3.0]}]' - description: Data to be upserted into the vector index. - description: Handles the upserting of data into a specified vector index. - /api/2.0/vector-search/endpoints: - post: - operationId: endpoints-createendpoint - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchendpoints/createendpoint - x-stackQL-resource: endpoints - x-stackQL-method: createendpoint - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Successful response with details of the created endpoint - content: - application/json: - schema: - type: object - properties: - name: - type: string - creator: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - endpoint_type: - type: string - last_updated_user: - type: string - id: - type: string - endpoint_status: - type: object - properties: - state: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - endpoint_type: - type: string - example: - name: docs-endpoint - endpoint_type: STANDARD - description: Create a new endpoint. - get: - operationId: endpoints-listendpoints - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchendpoints/listendpoints - x-stackQL-resource: endpoints - x-stackQL-method: listendpoints - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - responses: - '200': - description: List of all endpoints - content: - application/json: - schema: - type: object - properties: - endpoints: - type: array - items: - type: object - properties: - name: - type: string - creator: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - endpoint_type: - type: string - last_updated_user: - type: string - id: - type: string - endpoint_status: - type: object - properties: - state: - type: string - message: - type: string - num_indexes: - type: integer - next_page_token: - type: string - x-stackQL-objectKey: $.endpoints - /api/2.0/vector-search/endpoints/{endpoint_name}: - delete: - operationId: endpoints-deleteendpoint - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchendpoints/deleteendpoint - x-stackQL-resource: endpoints - x-stackQL-method: deleteendpoint - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: endpoint_name - required: true - description: string - in: path - responses: - '200': - description: Endpoint deletion initiated. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - get: - operationId: endpoints-getendpoint - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchendpoints/getendpoint - x-stackQL-resource: endpoints - x-stackQL-method: getendpoint - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: endpoint_name - required: true - description: string - in: path - responses: - '200': - description: Details of the endpoint. - content: - application/json: - schema: - type: object - properties: - name: - type: string - creator: - type: string - creation_timestamp: - type: integer - last_updated_timestamp: - type: integer - endpoint_type: - type: string - last_updated_user: - type: string - id: - type: string - endpoint_status: - type: object - properties: - state: - type: string - message: - type: string - num_indexes: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - /api/2.0/vector-search/indexes: - post: - operationId: indexes-createindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/createindex - x-stackQL-resource: indexes - x-stackQL-method: createindex - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Successful response with details of the created index - content: - application/json: - schema: - type: object - properties: - name: - type: string - endpoint_name: - type: string - primary_key: - type: string - index_type: - type: string - delta_sync_index_spec: - type: object - properties: - source_table: - type: string - embedding_source_columns: - type: array - items: - type: object - properties: - name: - type: string - embedding_model_endpoint_name: - type: string - columns_to_sync: - type: array - items: - type: string - pipeline_type: - type: string - pipeline_id: - type: string - status: - type: object - properties: - message: - type: string - indexed_row_count: - type: integer - ready: - type: boolean - index_url: - type: string - creator: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: required - endpoint_name: - type: string - primary_key: - type: required - index_type: - type: string - delta_sync_index_spec: - type: object - direct_access_index_spec: - type: object - example: - name: main_catalog.docs.en_wiki_index - primary_key: id - index_type: DELTA_SYNC - delta_sync_index_spec: - source_table: main_catalog.docs.en_wiki - pipeline_type: TRIGGERED - embedding_source_columns: - - name: text - embedding_model_endpoint_name: e5-small-v2 - columns_to_sync: - - id - - text - description: Create a new index. - get: - operationId: indexes-listindexes - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/listindexes - x-stackQL-resource: indexes - x-stackQL-method: listindexes - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: endpoint_name - required: true - description: string - in: query - - name: page_token - in: query - responses: - '200': - description: Successful response with list of endpoints. - content: - application/json: - schema: - type: object - properties: - vector_indexes: - type: array - items: - type: object - properties: - name: - type: string - endpoint_name: - type: string - primary_key: - type: string - index_type: - type: string - creator: - type: string - next_page_token: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.vector_indexes - description: List all indexes in the given endpoint. - /api/2.0/vector-search/indexes/{index_name}: - delete: - operationId: indexes-deleteindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/deleteindex - x-stackQL-resource: indexes - x-stackQL-method: deleteindex - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Index deletion has been initiated. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Delete an index. - get: - operationId: indexes-getindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/getindex - x-stackQL-resource: indexes - x-stackQL-method: getindex - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Successful response with details of the index - content: - application/json: - schema: - type: object - properties: - name: - type: string - endpoint_name: - type: string - primary_key: - type: string - index_type: - type: string - delta_sync_index_spec: - type: object - properties: - source_table: - type: string - embedding_source_columns: - type: array - items: - type: object - properties: - name: - type: string - embedding_model_endpoint_name: - type: string - pipeline_type: - type: string - pipeline_id: - type: string - status: - type: object - properties: - message: - type: string - indexed_row_count: - type: integer - ready: - type: boolean - index_url: - type: string - creator: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Get an index. - /api/2.0/vector-search/indexes/{index_name}/query-next-page: - post: - operationId: indexes-querynextpage - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/querynextpage - x-stackQL-resource: indexes - x-stackQL-method: querynextpage - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Successful response with query results. - content: - application/json: - schema: - type: object - properties: - manifest: - type: object - properties: - column_count: - type: integer - columns: - type: array - items: - type: object - properties: - name: - type: string - result: - type: object - properties: - row_count: - type: integer - data_array: - type: array - items: - type: array - items: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - endpoint_name: - type: string - page_token: - type: string - example: - endpoint_name: demo-endpoint - page_token: dummy-page-token - description: Use - /api/2.0/vector-search/indexes/{index_name}/sync: - post: - operationId: indexes-syncindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/syncindex - x-stackQL-resource: indexes - x-stackQL-method: syncindex - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Successful response indicating the sync process has been initiated. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Triggers a synchronization process for a specified vector index. - /api/2.0/vector-search/indexes/{index_name}/query: - post: - operationId: index-query-queryindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/queryindex - x-stackQL-resource: index_query - x-stackQL-method: queryindex - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Successful response with query results. - content: - application/json: - schema: - type: object - properties: - manifest: - type: object - properties: - column_count: - type: integer - columns: - type: array - items: - type: object - properties: - name: - type: string - result: - type: object - properties: - row_count: - type: integer - data_array: - type: array - items: - type: array - items: - type: string - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - num_results: - type: integer - columns: - type: required - filters_json: - type: Array of string - query_vector: - type: string - query_text: - type: Array of double - score_threshold: - type: string - query_type: - type: double - example: - query_vector: - - 1 - - 2 - - 3 - columns: - - id - - text - num_results: 10 - description: Query the specified vector index. - /api/2.0/vector-search/indexes/{index_name}/scan: - post: - operationId: index-scan-scanindex - externalDocs: - url: https://docs.databricks.com/api/workspace/vectorsearchindexes/scanindex - x-stackQL-resource: index_scan - x-stackQL-method: scanindex - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: index_name - required: true - description: string - in: path - responses: - '200': - description: Successful response with scan results. - content: - application/json: - schema: - type: object - properties: - last_primary_key: - type: string - data: - type: array - items: - type: object - properties: - fields: - type: array - items: - type: object - properties: - key: - type: string - value: - type: object - properties: - number_value: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - num_results: - type: integer - last_primary_key: - type: string - example: - num_results: 2 - description: Scan the specified vector index and return the first -components: - x-stackQL-resources: - data_vector_indexes: - id: databricks_workspace.vectorsearch.data_vector_indexes - name: data_vector_indexes - title: Data_vector_indexes - methods: - deletedatavectorindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1delete-data/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - upsertdatavectorindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1upsert-data/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/data_vector_indexes/methods/upsertdatavectorindex' - delete: - - $ref: '#/components/x-stackQL-resources/data_vector_indexes/methods/deletedatavectorindex' - exec: [] - endpoints: - id: databricks_workspace.vectorsearch.endpoints - name: endpoints - title: Endpoints - methods: - createendpoint: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1endpoints/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listendpoints: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1endpoints/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.endpoints - deleteendpoint: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1endpoints~1{endpoint_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getendpoint: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1endpoints~1{endpoint_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/endpoints/methods/getendpoint' - - $ref: '#/components/x-stackQL-resources/endpoints/methods/listendpoints' - insert: - - $ref: '#/components/x-stackQL-resources/endpoints/methods/createendpoint' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/endpoints/methods/deleteendpoint' - exec: [] - indexes: - id: databricks_workspace.vectorsearch.indexes - name: indexes - title: Indexes - methods: - createindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listindexes: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.vector_indexes - deleteindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - querynextpage: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1query-next-page/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - syncindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1sync/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/indexes/methods/getindex' - - $ref: '#/components/x-stackQL-resources/indexes/methods/listindexes' - insert: - - $ref: '#/components/x-stackQL-resources/indexes/methods/createindex' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/indexes/methods/deleteindex' - exec: - - $ref: '#/components/x-stackQL-resources/indexes/methods/syncindex' - - $ref: '#/components/x-stackQL-resources/indexes/methods/querynextpage' - index_query: - id: databricks_workspace.vectorsearch.index_query - name: index_query - title: Index_query - methods: - queryindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1query/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/index_query/methods/queryindex' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - index_scan: - id: databricks_workspace.vectorsearch.index_scan - name: index_scan - title: Index_scan - methods: - scanindex: - operation: - $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1scan/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/index_scan/methods/scanindex' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] +openapi: 3.0.0 +info: + title: Databricks Vectorsearch API (workspace) + description: >- + OpenAPI specification for the Databricks vectorsearch service + (workspace-level APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.vectorsearch +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/vector-search/endpoints: + post: + operationId: vector_search_endpoints_create_endpoint + summary: Create a new endpoint. + tags: + - vectorsearch + - vector_search_endpoints + description: |- + Create a new endpoint. + + :param name: str + Name of the vector search endpoint + :param endpoint_type: :class:`EndpointType` + Type of endpoint + :param budget_policy_id: str (optional) + The budget policy id to be applied + + :returns: + Long-running operation waiter for :class:`EndpointInfo`. + See :method:wait_get_endpoint_vector_search_endpoint_online for more details. + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the vector search endpoint + endpoint_type: + type: string + description: Type of endpoint + budget_policy_id: + type: string + description: The budget policy id to be applied + required: + - name + - endpoint_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EndpointInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchEndpointsAPI + get: + operationId: vector_search_endpoints_list_endpoints + summary: List all vector search endpoints in the workspace. + tags: + - vectorsearch + - vector_search_endpoints + description: |- + List all vector search endpoints in the workspace. + + :param page_token: str (optional) + Token for pagination + + :returns: Iterator over :class:`EndpointInfo` + parameters: + - name: page_token + in: query + required: false + schema: + type: string + description: Token for pagination + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListEndpointResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchEndpointsAPI + /api/2.0/vector-search/endpoints/{endpoint_name}: + delete: + operationId: vector_search_endpoints_delete_endpoint + summary: Delete a vector search endpoint. + tags: + - vectorsearch + - vector_search_endpoints + description: |- + Delete a vector search endpoint. + + :param endpoint_name: str + Name of the vector search endpoint + parameters: + - name: endpoint_name + in: path + required: true + schema: + type: string + description: Name of the vector search endpoint + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchEndpointsAPI + get: + operationId: vector_search_endpoints_get_endpoint + summary: Get details for a single vector search endpoint. + tags: + - vectorsearch + - vector_search_endpoints + description: |- + Get details for a single vector search endpoint. + + :param endpoint_name: str + Name of the endpoint + + :returns: :class:`EndpointInfo` + parameters: + - name: endpoint_name + in: path + required: true + schema: + type: string + description: Name of the endpoint + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/EndpointInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchEndpointsAPI + /api/2.0/vector-search/endpoints/{name}/metrics: + post: + operationId: vector_search_endpoints_retrieve_user_visible_metrics + summary: Retrieve user-visible metrics for an endpoint + tags: + - vectorsearch + - vector_search_endpoints + description: |- + Retrieve user-visible metrics for an endpoint + + :param name: str + Vector search endpoint name + :param end_time: str (optional) + End time for metrics query + :param granularity_in_seconds: int (optional) + Granularity in seconds + :param metrics: List[:class:`Metric`] (optional) + List of metrics to retrieve + :param page_token: str (optional) + Token for pagination + :param start_time: str (optional) + Start time for metrics query + + :returns: :class:`RetrieveUserVisibleMetricsResponse` + parameters: + - name: name + in: path + required: true + schema: + type: string + description: Vector search endpoint name + requestBody: + content: + application/json: + schema: + type: object + properties: + end_time: + type: string + description: End time for metrics query + granularity_in_seconds: + type: string + description: Granularity in seconds + metrics: + type: string + description: List of metrics to retrieve + page_token: + type: string + description: Token for pagination + start_time: + type: string + description: Start time for metrics query + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RetrieveUserVisibleMetricsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchEndpointsAPI + /api/2.0/vector-search/endpoints/{endpoint_name}/budget-policy: + patch: + operationId: vector_search_endpoints_update_endpoint_budget_policy + summary: Update the budget policy of an endpoint + tags: + - vectorsearch + - vector_search_endpoints + description: |- + Update the budget policy of an endpoint + + :param endpoint_name: str + Name of the vector search endpoint + :param budget_policy_id: str + The budget policy id to be applied + + :returns: :class:`PatchEndpointBudgetPolicyResponse` + parameters: + - name: endpoint_name + in: path + required: true + schema: + type: string + description: Name of the vector search endpoint + requestBody: + content: + application/json: + schema: + type: object + properties: + budget_policy_id: + type: string + description: The budget policy id to be applied + required: + - budget_policy_id + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/PatchEndpointBudgetPolicyResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchEndpointsAPI + /api/2.0/vector-search/endpoints/{endpoint_name}/tags: + patch: + operationId: vector_search_endpoints_update_endpoint_custom_tags + summary: Update the custom tags of an endpoint. + tags: + - vectorsearch + - vector_search_endpoints + description: |- + Update the custom tags of an endpoint. + + :param endpoint_name: str + Name of the vector search endpoint + :param custom_tags: List[:class:`CustomTag`] + The new custom tags for the vector search endpoint + + :returns: :class:`UpdateEndpointCustomTagsResponse` + parameters: + - name: endpoint_name + in: path + required: true + schema: + type: string + description: Name of the vector search endpoint + requestBody: + content: + application/json: + schema: + type: object + properties: + custom_tags: + type: string + description: The new custom tags for the vector search endpoint + required: + - custom_tags + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpdateEndpointCustomTagsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchEndpointsAPI + /api/2.0/vector-search/indexes: + post: + operationId: vector_search_indexes_create_index + summary: Create a new index. + tags: + - vectorsearch + - vector_search_indexes + description: >- + Create a new index. + + + :param name: str + Name of the index + :param endpoint_name: str + Name of the endpoint to be used for serving the index + :param primary_key: str + Primary key of the index + :param index_type: :class:`VectorIndexType` + + :param delta_sync_index_spec: :class:`DeltaSyncVectorIndexSpecRequest` + (optional) + Specification for Delta Sync Index. Required if `index_type` is `DELTA_SYNC`. + :param direct_access_index_spec: :class:`DirectAccessVectorIndexSpec` + (optional) + Specification for Direct Vector Access Index. Required if `index_type` is `DIRECT_ACCESS`. + + :returns: :class:`VectorIndex` + requestBody: + content: + application/json: + schema: + type: object + properties: + name: + type: string + description: Name of the index + endpoint_name: + type: string + description: Name of the endpoint to be used for serving the index + primary_key: + type: string + description: Primary key of the index + index_type: + type: string + description: >- + :param delta_sync_index_spec: + :class:`DeltaSyncVectorIndexSpecRequest` (optional) + Specification for Delta Sync Index. Required if `index_type` + is `DELTA_SYNC`. + delta_sync_index_spec: + type: string + direct_access_index_spec: + type: string + description: >- + Specification for Direct Vector Access Index. Required if + `index_type` is `DIRECT_ACCESS`. + required: + - name + - endpoint_name + - primary_key + - index_type + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VectorIndex' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + get: + operationId: vector_search_indexes_list_indexes + summary: List all indexes in the given endpoint. + tags: + - vectorsearch + - vector_search_indexes + description: |- + List all indexes in the given endpoint. + + :param endpoint_name: str + Name of the endpoint + :param page_token: str (optional) + Token for pagination + + :returns: Iterator over :class:`MiniVectorIndex` + parameters: + - name: endpoint_name + in: query + required: true + schema: + type: string + description: Name of the endpoint + - name: page_token + in: query + required: false + schema: + type: string + description: Token for pagination + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListVectorIndexesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + /api/2.0/vector-search/indexes/{index_name}/delete-data: + delete: + operationId: vector_search_indexes_delete_data_vector_index + summary: Handles the deletion of data from a specified vector index. + tags: + - vectorsearch + - vector_search_indexes + description: |- + Handles the deletion of data from a specified vector index. + + :param index_name: str + Name of the vector index where data is to be deleted. Must be a Direct Vector Access Index. + :param primary_keys: List[str] + List of primary keys for the data to be deleted. + + :returns: :class:`DeleteDataVectorIndexResponse` + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: >- + Name of the vector index where data is to be deleted. Must be a + Direct Vector Access Index. + - name: primary_keys + in: query + required: true + schema: + type: string + description: List of primary keys for the data to be deleted. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/DeleteDataVectorIndexResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + /api/2.0/vector-search/indexes/{index_name}: + delete: + operationId: vector_search_indexes_delete_index + summary: Delete an index. + tags: + - vectorsearch + - vector_search_indexes + description: |- + Delete an index. + + :param index_name: str + Name of the index + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: Name of the index + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + get: + operationId: vector_search_indexes_get_index + summary: Get an index. + tags: + - vectorsearch + - vector_search_indexes + description: |- + Get an index. + + :param index_name: str + Name of the index + :param ensure_reranker_compatible: bool (optional) + If true, the URL returned for the index is guaranteed to be compatible with the reranker. Currently + this means we return the CP URL regardless of how the index is being accessed. If not set or set to + false, the URL may still be compatible with the reranker depending on what URL we return. + + :returns: :class:`VectorIndex` + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: Name of the index + - name: ensure_reranker_compatible + in: query + required: false + schema: + type: string + description: >- + If true, the URL returned for the index is guaranteed to be + compatible with the reranker. Currently this means we return the CP + URL regardless of how the index is being accessed. If not set or set + to false, the URL may still be compatible with the reranker + depending on what URL we return. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/VectorIndex' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + /api/2.0/vector-search/indexes/{index_name}/query: + post: + operationId: vector_search_indexes_query_index + summary: Query the specified vector index. + tags: + - vectorsearch + - vector_search_indexes + description: |- + Query the specified vector index. + + :param index_name: str + Name of the vector index to query. + :param columns: List[str] + List of column names to include in the response. + :param columns_to_rerank: List[str] (optional) + Column names used to retrieve data to send to the reranker. + :param filters_json: str (optional) + JSON string representing query filters. + + Example filters: + + - `{"id <": 5}`: Filter for id less than 5. - `{"id >": 5}`: Filter for id greater than 5. - `{"id + <=": 5}`: Filter for id less than equal to 5. - `{"id >=": 5}`: Filter for id greater than equal to + 5. - `{"id": 5}`: Filter for id equal to 5. + :param num_results: int (optional) + Number of results to return. Defaults to 10. + :param query_text: str (optional) + Query text. Required for Delta Sync Index using model endpoint. + :param query_type: str (optional) + The query type to use. Choices are `ANN` and `HYBRID` and `FULL_TEXT`. Defaults to `ANN`. + :param query_vector: List[float] (optional) + Query vector. Required for Direct Vector Access Index and Delta Sync Index using self-managed + vectors. + :param reranker: :class:`RerankerConfig` (optional) + If set, the top 50 results are reranked with the Databricks Reranker model before returning the + `num_results` results to the user. The setting `columns_to_rerank` selects which columns are used + for reranking. For each datapoint, the columns selected are concatenated before being sent to the + reranking model. See https://docs.databricks.com/aws/en/vector-search/query-vector-search#rerank for + more information. + :param score_threshold: float (optional) + Threshold for the approximate nearest neighbor search. Defaults to 0.0. + + :returns: :class:`QueryVectorIndexResponse` + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: Name of the vector index to query. + requestBody: + content: + application/json: + schema: + type: object + properties: + columns: + type: string + description: List of column names to include in the response. + columns_to_rerank: + type: string + description: Column names used to retrieve data to send to the reranker. + filters_json: + type: string + description: >- + JSON string representing query filters. Example filters: - + `{"id <": 5}`: Filter for id less than 5. - `{"id >": 5}`: + Filter for id greater than 5. - `{"id <=": 5}`: Filter for + id less than equal to 5. - `{"id >=": 5}`: Filter for id + greater than equal to 5. - `{"id": 5}`: Filter for id equal + to 5. + num_results: + type: string + description: Number of results to return. Defaults to 10. + query_text: + type: string + description: >- + Query text. Required for Delta Sync Index using model + endpoint. + query_type: + type: string + description: >- + The query type to use. Choices are `ANN` and `HYBRID` and + `FULL_TEXT`. Defaults to `ANN`. + query_vector: + type: string + description: >- + Query vector. Required for Direct Vector Access Index and + Delta Sync Index using self-managed vectors. + reranker: + type: string + description: >- + If set, the top 50 results are reranked with the Databricks + Reranker model before returning the `num_results` results to + the user. The setting `columns_to_rerank` selects which + columns are used for reranking. For each datapoint, the + columns selected are concatenated before being sent to the + reranking model. See + https://docs.databricks.com/aws/en/vector-search/query-vector-search#rerank + for more information. + score_threshold: + type: string + description: >- + Threshold for the approximate nearest neighbor search. + Defaults to 0.0. + required: + - columns + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/QueryVectorIndexResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + /api/2.0/vector-search/indexes/{index_name}/query-next-page: + post: + operationId: vector_search_indexes_query_next_page + summary: >- + Use `next_page_token` returned from previous `QueryVectorIndex` or + `QueryVectorIndexNextPage` request + tags: + - vectorsearch + - vector_search_indexes + description: >- + Use `next_page_token` returned from previous `QueryVectorIndex` or + `QueryVectorIndexNextPage` request + + to fetch next page of results. + + + :param index_name: str + Name of the vector index to query. + :param endpoint_name: str (optional) + Name of the endpoint. + :param page_token: str (optional) + Page token returned from previous `QueryVectorIndex` or `QueryVectorIndexNextPage` API. + + :returns: :class:`QueryVectorIndexResponse` + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: Name of the vector index to query. + requestBody: + content: + application/json: + schema: + type: object + properties: + endpoint_name: + type: string + description: Name of the endpoint. + page_token: + type: string + description: >- + Page token returned from previous `QueryVectorIndex` or + `QueryVectorIndexNextPage` API. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/QueryVectorIndexResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + /api/2.0/vector-search/indexes/{index_name}/scan: + post: + operationId: vector_search_indexes_scan_index + summary: >- + Scan the specified vector index and return the first `num_results` + entries after the exclusive + tags: + - vectorsearch + - vector_search_indexes + description: >- + Scan the specified vector index and return the first `num_results` + entries after the exclusive + + `primary_key`. + + + :param index_name: str + Name of the vector index to scan. + :param last_primary_key: str (optional) + Primary key of the last entry returned in the previous scan. + :param num_results: int (optional) + Number of results to return. Defaults to 10. + + :returns: :class:`ScanVectorIndexResponse` + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: Name of the vector index to scan. + requestBody: + content: + application/json: + schema: + type: object + properties: + last_primary_key: + type: string + description: Primary key of the last entry returned in the previous scan. + num_results: + type: string + description: Number of results to return. Defaults to 10. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ScanVectorIndexResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + /api/2.0/vector-search/indexes/{index_name}/sync: + post: + operationId: vector_search_indexes_sync_index + summary: Triggers a synchronization process for a specified vector index. + tags: + - vectorsearch + - vector_search_indexes + description: |- + Triggers a synchronization process for a specified vector index. + + :param index_name: str + Name of the vector index to synchronize. Must be a Delta Sync Index. + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: Name of the vector index to synchronize. Must be a Delta Sync Index. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI + /api/2.0/vector-search/indexes/{index_name}/upsert-data: + post: + operationId: vector_search_indexes_upsert_data_vector_index + summary: Handles the upserting of data into a specified vector index. + tags: + - vectorsearch + - vector_search_indexes + description: |- + Handles the upserting of data into a specified vector index. + + :param index_name: str + Name of the vector index where data is to be upserted. Must be a Direct Vector Access Index. + :param inputs_json: str + JSON string representing the data to be upserted. + + :returns: :class:`UpsertDataVectorIndexResponse` + parameters: + - name: index_name + in: path + required: true + schema: + type: string + description: >- + Name of the vector index where data is to be upserted. Must be a + Direct Vector Access Index. + requestBody: + content: + application/json: + schema: + type: object + properties: + inputs_json: + type: string + description: JSON string representing the data to be upserted. + required: + - inputs_json + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/UpsertDataVectorIndexResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: VectorSearchIndexesAPI +components: + schemas: + ColumnInfo: + type: object + properties: + name: + type: string + type_text: + type: string + description: Data type of the column (e.g., "string", "int", "array") + CustomTag: + type: object + properties: + key: + type: string + value: + type: string + description: '[Optional] Value field for a vector search endpoint tag.' + required: + - key + DeleteDataResult: + type: object + properties: + failed_primary_keys: + type: array + items: + type: string + success_row_count: + type: integer + description: Count of successfully processed rows. + DeleteDataVectorIndexResponse: + type: object + properties: + result: + $ref: '#/components/schemas/DeleteDataResult' + status: + $ref: '#/components/schemas/DeleteDataStatus' + description: Status of the delete operation. + DeleteEndpointResponse: + type: object + properties: {} + DeleteIndexResponse: + type: object + properties: {} + DeltaSyncVectorIndexSpecRequest: + type: object + properties: + columns_to_sync: + type: array + items: + type: string + embedding_source_columns: + type: array + items: + $ref: '#/components/schemas/EmbeddingSourceColumn' + description: The columns that contain the embedding source. + embedding_vector_columns: + type: array + items: + $ref: '#/components/schemas/EmbeddingVectorColumn' + description: The columns that contain the embedding vectors. + embedding_writeback_table: + type: string + description: >- + [Optional] Name of the Delta table to sync the vector index contents + and computed embeddings to. + pipeline_type: + $ref: '#/components/schemas/PipelineType' + description: >- + Pipeline execution mode. - `TRIGGERED`: If the pipeline uses the + triggered execution mode, the system stops processing after + successfully refreshing the source table in the pipeline once, + ensuring the table is updated based on the data available when the + update started. - `CONTINUOUS`: If the pipeline uses continuous + execution, the pipeline processes new data as it arrives in the + source table to keep vector index fresh. + source_table: + type: string + description: The name of the source table. + DeltaSyncVectorIndexSpecResponse: + type: object + properties: + embedding_source_columns: + type: array + items: + $ref: '#/components/schemas/EmbeddingSourceColumn' + embedding_vector_columns: + type: array + items: + $ref: '#/components/schemas/EmbeddingVectorColumn' + description: The columns that contain the embedding vectors. + embedding_writeback_table: + type: string + description: >- + [Optional] Name of the Delta table to sync the vector index contents + and computed embeddings to. + pipeline_id: + type: string + description: The ID of the pipeline that is used to sync the index. + pipeline_type: + $ref: '#/components/schemas/PipelineType' + description: >- + Pipeline execution mode. - `TRIGGERED`: If the pipeline uses the + triggered execution mode, the system stops processing after + successfully refreshing the source table in the pipeline once, + ensuring the table is updated based on the data available when the + update started. - `CONTINUOUS`: If the pipeline uses continuous + execution, the pipeline processes new data as it arrives in the + source table to keep vector index fresh. + source_table: + type: string + description: The name of the source table. + DirectAccessVectorIndexSpec: + type: object + properties: + embedding_source_columns: + type: array + items: + $ref: '#/components/schemas/EmbeddingSourceColumn' + embedding_vector_columns: + type: array + items: + $ref: '#/components/schemas/EmbeddingVectorColumn' + description: >- + The columns that contain the embedding vectors. The format should be + array[double]. + schema_json: + type: string + description: >- + The schema of the index in JSON format. Supported types are + `integer`, `long`, `float`, `double`, `boolean`, `string`, `date`, + `timestamp`. Supported types for vector column: `array`, + `array`,`. + EmbeddingSourceColumn: + type: object + properties: + embedding_model_endpoint_name: + type: string + model_endpoint_name_for_query: + type: string + description: >- + Name of the embedding model endpoint which, if specified, is used + for querying (not ingestion). + name: + type: string + description: Name of the column + EmbeddingVectorColumn: + type: object + properties: + embedding_dimension: + type: integer + name: + type: string + description: Name of the column + EndpointInfo: + type: object + properties: + creation_timestamp: + type: integer + creator: + type: string + description: Creator of the endpoint + custom_tags: + type: array + items: + $ref: '#/components/schemas/CustomTag' + description: The custom tags assigned to the endpoint + effective_budget_policy_id: + type: string + description: The budget policy id applied to the endpoint + endpoint_status: + $ref: '#/components/schemas/EndpointStatus' + description: Current status of the endpoint + endpoint_type: + $ref: '#/components/schemas/EndpointType' + description: Type of endpoint + id: + type: string + description: Unique identifier of the endpoint + last_updated_timestamp: + type: integer + description: Timestamp of last update to the endpoint + last_updated_user: + type: string + description: User who last updated the endpoint + name: + type: string + description: Name of the vector search endpoint + num_indexes: + type: integer + description: Number of indexes on the endpoint + EndpointStatus: + type: object + properties: + message: + type: string + description: Additional status message + state: + $ref: '#/components/schemas/EndpointStatusState' + description: Current state of the endpoint + description: Status information of an endpoint + ListEndpointResponse: + type: object + properties: + endpoints: + type: array + items: + $ref: '#/components/schemas/EndpointInfo' + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If not + present, there are no more results to show. + ListValue: + type: object + properties: + values: + type: array + items: + $ref: '#/components/schemas/Value' + ListVectorIndexesResponse: + type: object + properties: + next_page_token: + type: string + vector_indexes: + type: array + items: + $ref: '#/components/schemas/MiniVectorIndex' + MapStringValueEntry: + type: object + properties: + key: + type: string + description: Column name. + value: + $ref: '#/components/schemas/Value' + description: Column value, nullable. + description: Key-value pair. + Metric: + type: object + properties: + labels: + type: array + items: + $ref: '#/components/schemas/MetricLabel' + description: Metric labels + name: + type: string + description: Metric name + percentile: + type: number + description: Percentile for the metric + description: Metric specification + MetricLabel: + type: object + properties: + name: + type: string + description: Label name + value: + type: string + description: Label value + description: Label for a metric + MetricValue: + type: object + properties: + timestamp: + type: integer + description: Timestamp of the metric value (milliseconds since epoch) + value: + type: number + description: Metric value + description: Single metric value at a specific timestamp + MetricValues: + type: object + properties: + metric: + $ref: '#/components/schemas/Metric' + description: Metric specification + values: + type: array + items: + $ref: '#/components/schemas/MetricValue' + description: Time series of metric values + description: Collection of metric values for a specific metric + MiniVectorIndex: + type: object + properties: + creator: + type: string + endpoint_name: + type: string + description: Name of the endpoint associated with the index + index_type: + $ref: '#/components/schemas/VectorIndexType' + name: + type: string + description: Name of the index + primary_key: + type: string + description: Primary key of the index + PatchEndpointBudgetPolicyResponse: + type: object + properties: + effective_budget_policy_id: + type: string + QueryVectorIndexResponse: + type: object + properties: + manifest: + $ref: '#/components/schemas/ResultManifest' + next_page_token: + type: string + description: >- + [Optional] Token that can be used in `QueryVectorIndexNextPage` API + to get next page of results. If more than 1000 results satisfy the + query, they are returned in groups of 1000. Empty value means no + more results. The maximum number of results that can be returned is + 10,000. + result: + $ref: '#/components/schemas/ResultData' + description: Data returned in the query result. + RerankerConfig: + type: object + properties: + model: + type: string + parameters: + $ref: '#/components/schemas/RerankerConfigRerankerParameters' + RerankerConfigRerankerParameters: + type: object + properties: + columns_to_rerank: + type: array + items: + type: string + ResultData: + type: object + properties: + data_array: + type: array + items: + type: array + items: + type: string + description: Data rows returned in the query. + row_count: + type: integer + description: Number of rows in the result set. + description: Data returned in the query result. + ResultManifest: + type: object + properties: + column_count: + type: integer + description: Number of columns in the result set. + columns: + type: array + items: + $ref: '#/components/schemas/ColumnInfo' + description: Information about each column in the result set. + description: Metadata about the result set. + RetrieveUserVisibleMetricsResponse: + type: object + properties: + metric_values: + type: array + items: + $ref: '#/components/schemas/MetricValues' + description: Collection of metric values + next_page_token: + type: string + description: >- + A token that can be used to get the next page of results. If not + present, there are no more results to show. + description: Response containing user-visible metrics + ScanVectorIndexResponse: + type: object + properties: + data: + type: array + items: + $ref: '#/components/schemas/Struct' + description: List of data entries + last_primary_key: + type: string + description: Primary key of the last entry. + description: Response to a scan vector index request. + Struct: + type: object + properties: + fields: + type: array + items: + $ref: '#/components/schemas/MapStringValueEntry' + SyncIndexResponse: + type: object + properties: {} + UpdateEndpointCustomTagsResponse: + type: object + properties: + custom_tags: + type: array + items: + $ref: '#/components/schemas/CustomTag' + name: + type: string + description: >- + The name of the vector search endpoint whose custom tags were + updated. + UpsertDataResult: + type: object + properties: + failed_primary_keys: + type: array + items: + type: string + success_row_count: + type: integer + description: Count of successfully processed rows. + UpsertDataVectorIndexResponse: + type: object + properties: + result: + $ref: '#/components/schemas/UpsertDataResult' + status: + $ref: '#/components/schemas/UpsertDataStatus' + description: Status of the upsert operation. + Value: + type: object + properties: + bool_value: + type: boolean + list_value: + $ref: '#/components/schemas/ListValue' + number_value: + type: number + string_value: + type: string + struct_value: + $ref: '#/components/schemas/Struct' + VectorIndex: + type: object + properties: + creator: + type: string + delta_sync_index_spec: + $ref: '#/components/schemas/DeltaSyncVectorIndexSpecResponse' + direct_access_index_spec: + $ref: '#/components/schemas/DirectAccessVectorIndexSpec' + endpoint_name: + type: string + description: Name of the endpoint associated with the index + index_type: + $ref: '#/components/schemas/VectorIndexType' + name: + type: string + description: Name of the index + primary_key: + type: string + description: Primary key of the index + status: + $ref: '#/components/schemas/VectorIndexStatus' + VectorIndexStatus: + type: object + properties: + index_url: + type: string + indexed_row_count: + type: integer + description: Number of rows indexed + message: + type: string + description: Message associated with the index status + ready: + type: boolean + description: Whether the index is ready for search + DeleteDataStatus: + type: string + enum: + - FAILURE + - PARTIAL_SUCCESS + - SUCCESS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + EndpointStatusState: + type: string + enum: + - DELETED + - OFFLINE + - ONLINE + - PROVISIONING + - RED_STATE + - YELLOW_STATE + description: Current state of the endpoint + EndpointType: + type: string + enum: + - STANDARD + description: Type of endpoint. + PipelineType: + type: string + enum: + - CONTINUOUS + - TRIGGERED + description: >- + Pipeline execution mode. - `TRIGGERED`: If the pipeline uses the + triggered execution mode, the + + system stops processing after successfully refreshing the source table + in the pipeline once, + + ensuring the table is updated based on the data available when the + update started. - + + `CONTINUOUS`: If the pipeline uses continuous execution, the pipeline + processes new data as it + + arrives in the source table to keep vector index fresh. + UpsertDataStatus: + type: string + enum: + - FAILURE + - PARTIAL_SUCCESS + - SUCCESS + description: |- + Create a collection of name/value pairs. + + Example enumeration: + + >>> class Color(Enum): + ... RED = 1 + ... BLUE = 2 + ... GREEN = 3 + + Access them by: + + - attribute access: + + >>> Color.RED + + + - value lookup: + + >>> Color(1) + + + - name lookup: + + >>> Color['RED'] + + + Enumerations can be iterated over, and know how many members they have: + + >>> len(Color) + 3 + + >>> list(Color) + [, , ] + + Methods can be added to enumerations, and members can have their own + attributes -- see the documentation for details. + VectorIndexType: + type: string + enum: + - DELTA_SYNC + - DIRECT_ACCESS + description: >- + There are 2 types of Vector Search indexes: - `DELTA_SYNC`: An index + that automatically syncs + + with a source Delta Table, automatically and incrementally updating the + index as the underlying + + data in the Delta Table changes. - `DIRECT_ACCESS`: An index that + supports direct read and write + + of vectors and metadata through our REST and SDK APIs. With this model, + the user manages index + + updates. + x-stackQL-resources: + endpoints: + id: databricks_workspace.vectorsearch.endpoints + name: endpoints + title: Endpoints + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1endpoints/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1endpoints/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.endpoints + delete: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1endpoints~1{endpoint_name}/delete + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1endpoints~1{endpoint_name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + retrieve_user_visible_metrics: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1endpoints~1{name}~1metrics/post' + response: + mediaType: application/json + openAPIDocKey: '200' + update_endpoint_budget_policy: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1endpoints~1{endpoint_name}~1budget-policy/patch + response: + mediaType: application/json + openAPIDocKey: '200' + update_endpoint_custom_tags: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1endpoints~1{endpoint_name}~1tags/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/endpoints/methods/get' + - $ref: '#/components/x-stackQL-resources/endpoints/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/endpoints/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/endpoints/methods/delete' + replace: [] + indexes: + id: databricks_workspace.vectorsearch.indexes + name: indexes + title: Indexes + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1indexes/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1indexes/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.vector_indexes + delete_data_vector_index: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1delete-data/delete + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1vector-search~1indexes~1{index_name}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + query_index: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1query/post + response: + mediaType: application/json + openAPIDocKey: '200' + query_next_page: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1query-next-page/post + response: + mediaType: application/json + openAPIDocKey: '200' + scan_index: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1scan/post + response: + mediaType: application/json + openAPIDocKey: '200' + sync_index: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1sync/post + response: + mediaType: application/json + openAPIDocKey: '200' + upsert_data_vector_index: + operation: + $ref: >- + #/paths/~1api~12.0~1vector-search~1indexes~1{index_name}~1upsert-data/post + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/indexes/methods/get' + - $ref: '#/components/x-stackQL-resources/indexes/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/indexes/methods/create' + update: [] + delete: + - $ref: '#/components/x-stackQL-resources/indexes/methods/delete' + replace: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/workflows.yaml b/providers/src/databricks_workspace/v00.00.00000/services/workflows.yaml deleted file mode 100644 index f9fe6d13..00000000 --- a/providers/src/databricks_workspace/v00.00.00000/services/workflows.yaml +++ /dev/null @@ -1,4507 +0,0 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Workflows API - description: Workflows -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.2/jobs/create: - post: - operationId: jobs-create - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/create - x-stackQL-resource: jobs - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - job_id: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - name: - type: string - description: - type: string - timeout_seconds: - type: int32 - max_concurrent_runs: - type: int32 - format: - type: string - edit_mode: - type: string - email_notifications: - type: object - webhook_notifications: - type: object - notification_settings: - type: object - health: - type: object - schedule: - type: object - trigger: - type: object - continuous: - type: object - tasks: - type: Array of object - job_clusters: - type: Array of object - git_source: - type: object - tags: - type: object - queue: - type: object - parameters: - type: Array of object - run_as: - type: object - deployment: - type: object - environments: - type: Array of object - access_control_list: - type: Array of object - example: - name: A multitask job - description: This job contain multiple tasks that are required to - produce the weekly shark sightings report. - email_notifications: - on_start: - - user.name@databricks.com - on_success: - - user.name@databricks.com - on_failure: - - user.name@databricks.com - on_duration_warning_threshold_exceeded: - - user.name@databricks.com - on_streaming_backlog_exceeded: - - user.name@databricks.com - no_alert_for_skipped_runs: false - webhook_notifications: - on_start: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_success: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_failure: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_duration_warning_threshold_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_streaming_backlog_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - notification_settings: - no_alert_for_skipped_runs: false - no_alert_for_canceled_runs: false - timeout_seconds: 86400 - health: - rules: - - metric: RUN_DURATION_SECONDS - op: GREATER_THAN - value: 10 - schedule: - quartz_cron_expression: 20 30 * * * ? - timezone_id: Europe/London - pause_status: UNPAUSED - trigger: - pause_status: UNPAUSED - file_arrival: - url: string - min_time_between_triggers_seconds: 0 - wait_after_last_change_seconds: 0 - periodic: - interval: 0 - unit: HOURS - continuous: - pause_status: UNPAUSED - max_concurrent_runs: 10 - tasks: - - max_retries: 3 - task_key: Sessionize - description: Extracts session data from events - min_retry_interval_millis: 2000 - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.Sessionize - parameters: - - --data - - dbfs:/path/to/data.json - libraries: - - jar: dbfs:/mnt/databricks/Sessionize.jar - retry_on_timeout: false - existing_cluster_id: 0923-164208-meows279 - - max_retries: 3 - task_key: Orders_Ingest - description: Ingests order data - job_cluster_key: auto_scaling_cluster - min_retry_interval_millis: 2000 - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.OrdersIngest - parameters: - - --data - - dbfs:/path/to/order-data.json - libraries: - - jar: dbfs:/mnt/databricks/OrderIngest.jar - retry_on_timeout: false - - max_retries: 3 - task_key: Match - description: Matches orders with user sessions - notebook_task: - base_parameters: - age: '35' - name: John Doe - notebook_path: /Users/user.name@databricks.com/Match - min_retry_interval_millis: 2000 - depends_on: - - task_key: Orders_Ingest - - task_key: Sessionize - new_cluster: - autoscale: - max_workers: 16 - min_workers: 2 - node_type_id: null - spark_conf: - spark.speculation: true - spark_version: 7.3.x-scala2.12 - timeout_seconds: 86400 - retry_on_timeout: false - run_if: ALL_SUCCESS - job_clusters: - - job_cluster_key: auto_scaling_cluster - new_cluster: - autoscale: - max_workers: 16 - min_workers: 2 - node_type_id: null - spark_conf: - spark.speculation: true - spark_version: 7.3.x-scala2.12 - git_source: - git_branch: main - git_provider: gitHub - git_url: https://github.com/databricks/databricks-cli - tags: - cost-center: engineering - team: jobs - format: SINGLE_TASK - queue: - enabled: true - parameters: - - default: users - name: table - run_as: - user_name: user@databricks.com - service_principal_name: 692bc6d0-ffa3-11ed-be56-0242ac120002 - edit_mode: UI_LOCKED - deployment: - kind: BUNDLE - metadata_file_path: string - environments: - - environment_key: string - spec: - client: '1' - dependencies: - - string - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Create a new job. - /api/2.2/jobs/delete: - post: - operationId: jobs-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/delete - x-stackQL-resource: jobs - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - job_id: - type: required - example: - job_id: 11223344 - description: Deletes a job. - /api/2.2/jobs/get: - get: - operationId: jobs-get - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/get - x-stackQL-resource: jobs - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: job_id - required: true - description: int64 - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - job_id: - type: integer - creator_user_name: - type: string - run_as_user_name: - type: string - settings: - type: object - properties: - name: - type: string - description: - type: string - email_notifications: - type: object - properties: - on_start: - type: array - items: - type: string - on_success: - type: array - items: - type: string - on_failure: - type: array - items: - type: string - on_duration_warning_threshold_exceeded: - type: array - items: - type: string - on_streaming_backlog_exceeded: - type: array - items: - type: string - no_alert_for_skipped_runs: - type: boolean - webhook_notifications: - type: object - properties: - on_start: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_success: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_failure: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_duration_warning_threshold_exceeded: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_streaming_backlog_exceeded: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - notification_settings: - type: object - properties: - no_alert_for_skipped_runs: - type: boolean - no_alert_for_canceled_runs: - type: boolean - timeout_seconds: - type: integer - health: - type: object - properties: - rules: - type: array - items: - type: object - properties: - metric: - type: string - op: - type: string - value: - type: integer - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - trigger: - type: object - properties: - pause_status: - type: string - file_arrival: - type: object - properties: - url: - type: string - min_time_between_triggers_seconds: - type: integer - wait_after_last_change_seconds: - type: integer - periodic: - type: object - properties: - interval: - type: integer - unit: - type: string - continuous: - type: object - properties: - pause_status: - type: string - max_concurrent_runs: - type: integer - tasks: - type: array - items: - type: object - properties: - max_retries: - type: integer - task_key: - type: string - description: - type: string - min_retry_interval_millis: - type: integer - depends_on: - type: array - items: {} - timeout_seconds: - type: integer - spark_jar_task: - type: object - properties: - main_class_name: - type: string - parameters: - type: array - items: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - retry_on_timeout: - type: boolean - existing_cluster_id: - type: string - job_clusters: - type: array - items: - type: object - properties: - job_cluster_key: - type: string - new_cluster: - type: object - properties: - autoscale: - type: object - properties: - max_workers: - type: integer - min_workers: - type: integer - node_type_id: - type: 'null' - spark_conf: - type: object - properties: - spark.speculation: - type: boolean - spark_version: - type: string - git_source: - type: object - properties: - git_branch: - type: string - git_provider: - type: string - git_url: - type: string - tags: - type: object - properties: - cost-center: - type: string - team: - type: string - format: - type: string - queue: - type: object - properties: - enabled: - type: boolean - parameters: - type: array - items: - type: object - properties: - default: - type: string - name: - type: string - run_as: - type: object - properties: - user_name: - type: string - service_principal_name: - type: string - edit_mode: - type: string - deployment: - type: object - properties: - kind: - type: string - metadata_file_path: - type: string - environments: - type: array - items: - type: object - properties: - environment_key: - type: string - spec: - type: object - properties: - client: - type: string - dependencies: - type: array - items: - type: string - created_time: - type: integer - has_more: - type: boolean - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: Retrieves the details for a single job. - /api/2.2/jobs/list: - get: - operationId: jobs-list - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/list - x-stackQL-resource: jobs - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: offset - in: query - - name: limit - in: query - - name: expand_tasks - in: query - - name: name - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - jobs: - type: array - items: - type: object - properties: - job_id: - type: integer - creator_user_name: - type: string - settings: - type: object - properties: - name: - type: string - description: - type: string - email_notifications: - type: object - properties: - on_start: - type: array - items: - type: string - on_success: - type: array - items: - type: string - on_failure: - type: array - items: - type: string - on_duration_warning_threshold_exceeded: - type: array - items: - type: string - on_streaming_backlog_exceeded: - type: array - items: - type: string - no_alert_for_skipped_runs: - type: boolean - webhook_notifications: - type: object - properties: - on_start: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_success: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_failure: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_duration_warning_threshold_exceeded: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_streaming_backlog_exceeded: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - notification_settings: - type: object - properties: - no_alert_for_skipped_runs: - type: boolean - no_alert_for_canceled_runs: - type: boolean - timeout_seconds: - type: integer - health: - type: object - properties: - rules: - type: array - items: - type: object - properties: - metric: - type: string - op: - type: string - value: - type: integer - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - trigger: - type: object - properties: - pause_status: - type: string - file_arrival: - type: object - properties: - url: - type: string - min_time_between_triggers_seconds: - type: integer - wait_after_last_change_seconds: - type: integer - periodic: - type: object - properties: - interval: - type: integer - unit: - type: string - continuous: - type: object - properties: - pause_status: - type: string - max_concurrent_runs: - type: integer - tasks: - type: array - items: - type: object - properties: - max_retries: - type: integer - task_key: - type: string - description: - type: string - min_retry_interval_millis: - type: integer - depends_on: - type: array - items: {} - timeout_seconds: - type: integer - spark_jar_task: - type: object - properties: - main_class_name: - type: string - parameters: - type: array - items: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - retry_on_timeout: - type: boolean - existing_cluster_id: - type: string - job_clusters: - type: array - items: - type: object - properties: - job_cluster_key: - type: string - new_cluster: - type: object - properties: - autoscale: - type: object - properties: - max_workers: - type: integer - min_workers: - type: integer - node_type_id: - type: 'null' - spark_conf: - type: object - properties: - spark.speculation: - type: boolean - spark_version: - type: string - git_source: - type: object - properties: - git_branch: - type: string - git_provider: - type: string - git_url: - type: string - tags: - type: object - properties: - cost-center: - type: string - team: - type: string - format: - type: string - queue: - type: object - properties: - enabled: - type: boolean - parameters: - type: array - items: - type: object - properties: - default: - type: string - name: - type: string - run_as: - type: object - properties: - user_name: - type: string - service_principal_name: - type: string - edit_mode: - type: string - deployment: - type: object - properties: - kind: - type: string - metadata_file_path: - type: string - environments: - type: array - items: - type: object - properties: - environment_key: - type: string - spec: - type: object - properties: - client: - type: string - dependencies: - type: array - items: - type: string - created_time: - type: integer - has_more: - type: boolean - next_page_token: - type: string - prev_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - x-stackQL-objectKey: $.jobs - description: Retrieves a list of jobs. - /api/2.2/jobs/reset: - post: - operationId: jobs-reset - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/reset - x-stackQL-resource: jobs - x-stackQL-method: reset - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - job_id: - type: required - new_settings: - type: required - example: - job_id: 11223344 - new_settings: - name: A multitask job - description: This job contain multiple tasks that are required to - produce the weekly shark sightings report. - email_notifications: - on_start: - - user.name@databricks.com - on_success: - - user.name@databricks.com - on_failure: - - user.name@databricks.com - on_duration_warning_threshold_exceeded: - - user.name@databricks.com - on_streaming_backlog_exceeded: - - user.name@databricks.com - no_alert_for_skipped_runs: false - webhook_notifications: - on_start: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_success: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_failure: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_duration_warning_threshold_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_streaming_backlog_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - notification_settings: - no_alert_for_skipped_runs: false - no_alert_for_canceled_runs: false - timeout_seconds: 86400 - health: - rules: - - metric: RUN_DURATION_SECONDS - op: GREATER_THAN - value: 10 - schedule: - quartz_cron_expression: 20 30 * * * ? - timezone_id: Europe/London - pause_status: UNPAUSED - trigger: - pause_status: UNPAUSED - file_arrival: - url: string - min_time_between_triggers_seconds: 0 - wait_after_last_change_seconds: 0 - periodic: - interval: 0 - unit: HOURS - continuous: - pause_status: UNPAUSED - max_concurrent_runs: 10 - tasks: - - max_retries: 3 - task_key: Sessionize - description: Extracts session data from events - min_retry_interval_millis: 2000 - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.Sessionize - parameters: - - --data - - dbfs:/path/to/data.json - libraries: - - jar: dbfs:/mnt/databricks/Sessionize.jar - retry_on_timeout: false - existing_cluster_id: 0923-164208-meows279 - - max_retries: 3 - task_key: Orders_Ingest - description: Ingests order data - job_cluster_key: auto_scaling_cluster - min_retry_interval_millis: 2000 - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.OrdersIngest - parameters: - - --data - - dbfs:/path/to/order-data.json - libraries: - - jar: dbfs:/mnt/databricks/OrderIngest.jar - retry_on_timeout: false - - max_retries: 3 - task_key: Match - description: Matches orders with user sessions - notebook_task: - base_parameters: - age: '35' - name: John Doe - notebook_path: /Users/user.name@databricks.com/Match - min_retry_interval_millis: 2000 - depends_on: - - task_key: Orders_Ingest - - task_key: Sessionize - new_cluster: - autoscale: - max_workers: 16 - min_workers: 2 - node_type_id: null - spark_conf: - spark.speculation: true - spark_version: 7.3.x-scala2.12 - timeout_seconds: 86400 - retry_on_timeout: false - run_if: ALL_SUCCESS - job_clusters: - - job_cluster_key: auto_scaling_cluster - new_cluster: - autoscale: - max_workers: 16 - min_workers: 2 - node_type_id: null - spark_conf: - spark.speculation: true - spark_version: 7.3.x-scala2.12 - git_source: - git_branch: main - git_provider: gitHub - git_url: https://github.com/databricks/databricks-cli - tags: - cost-center: engineering - team: jobs - format: SINGLE_TASK - queue: - enabled: true - parameters: - - default: users - name: table - run_as: - user_name: user@databricks.com - service_principal_name: 692bc6d0-ffa3-11ed-be56-0242ac120002 - edit_mode: UI_LOCKED - deployment: - kind: BUNDLE - metadata_file_path: string - environments: - - environment_key: string - spec: - client: '1' - dependencies: - - string - description: Overwrite all settings for the given job. Use the - /api/2.2/jobs/run-now: - post: - operationId: jobs-runnow - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/runnow - x-stackQL-resource: jobs - x-stackQL-method: runnow - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - run_id: - type: integer - number_in_job: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - job_id: - type: required - idempotency_token: - type: integer - format: int64 - job_parameters: - type: object - pipeline_params: - type: object - queue: - type: object - example: - job_id: 11223344 - job_parameters: - property1: string - property2: string - pipeline_params: - full_refresh: false - idempotency_token: 8f018174-4792-40d5-bcbc-3e6a527352c8 - queue: - enabled: true - description: Run a job and return the - /api/2.2/jobs/runs/submit: - post: - operationId: jobs-submit - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/submit - x-stackQL-resource: jobs - x-stackQL-method: submit - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - run_id: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_name: - type: string - timeout_seconds: - type: int32 - idempotency_token: - type: string - health: - type: object - tasks: - type: Array of object - git_source: - type: object - webhook_notifications: - type: object - email_notifications: - type: object - notification_settings: - type: object - environments: - type: Array of object - access_control_list: - type: Array of object - queue: - type: object - run_as: - type: object - example: - run_name: A multitask job run - timeout_seconds: 86400 - health: - rules: - - metric: RUN_DURATION_SECONDS - op: GREATER_THAN - value: 10 - idempotency_token: 8f018174-4792-40d5-bcbc-3e6a527352c8 - tasks: - - task_key: Sessionize - description: Extracts session data from events - min_retry_interval_millis: 2000 - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.Sessionize - parameters: - - --data - - dbfs:/path/to/data.json - libraries: - - jar: dbfs:/mnt/databricks/Sessionize.jar - retry_on_timeout: false - existing_cluster_id: 0923-164208-meows279 - - task_key: Orders_Ingest - description: Ingests order data - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.OrdersIngest - parameters: - - --data - - dbfs:/path/to/order-data.json - libraries: - - jar: dbfs:/mnt/databricks/OrderIngest.jar - existing_cluster_id: 0923-164208-meows279 - - task_key: Match - description: Matches orders with user sessions - notebook_task: - base_parameters: - age: '35' - name: John Doe - notebook_path: /Users/user.name@databricks.com/Match - depends_on: - - task_key: Orders_Ingest - - task_key: Sessionize - new_cluster: - autoscale: - max_workers: 16 - min_workers: 2 - node_type_id: null - spark_conf: - spark.speculation: true - spark_version: 7.3.x-scala2.12 - timeout_seconds: 86400 - retry_on_timeout: false - run_if: ALL_SUCCESS - git_source: - git_branch: main - git_provider: gitHub - git_url: https://github.com/databricks/databricks-cli - webhook_notifications: - on_start: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_success: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_failure: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_duration_warning_threshold_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_streaming_backlog_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - email_notifications: - on_start: - - user.name@databricks.com - on_success: - - user.name@databricks.com - on_failure: - - user.name@databricks.com - on_duration_warning_threshold_exceeded: - - user.name@databricks.com - on_streaming_backlog_exceeded: - - user.name@databricks.com - no_alert_for_skipped_runs: false - notification_settings: - no_alert_for_skipped_runs: false - no_alert_for_canceled_runs: false - environments: - - environment_key: string - spec: - client: '1' - dependencies: - - string - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - queue: - enabled: true - run_as: - user_name: user@databricks.com - service_principal_name: 692bc6d0-ffa3-11ed-be56-0242ac120002 - description: "Submit a one-time run. This endpoint allows you to submit a workload\ - \ directly without creating a job. Runs submitted using this endpoint don\u2019\ - t display in the UI. Use the" - /api/2.2/jobs/update: - post: - operationId: jobs-update - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/update - x-stackQL-resource: jobs - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - job_id: - type: required - fields_to_remove: - type: integer - format: int64 - new_settings: - type: object - example: - job_id: 11223344 - new_settings: - name: A multitask job - description: This job contain multiple tasks that are required to - produce the weekly shark sightings report. - email_notifications: - on_start: - - user.name@databricks.com - on_success: - - user.name@databricks.com - on_failure: - - user.name@databricks.com - on_duration_warning_threshold_exceeded: - - user.name@databricks.com - on_streaming_backlog_exceeded: - - user.name@databricks.com - no_alert_for_skipped_runs: false - webhook_notifications: - on_start: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_success: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_failure: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_duration_warning_threshold_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - on_streaming_backlog_exceeded: - - - id: 0481e838-0a59-4eff-9541-a4ca6f149574 - notification_settings: - no_alert_for_skipped_runs: false - no_alert_for_canceled_runs: false - timeout_seconds: 86400 - health: - rules: - - metric: RUN_DURATION_SECONDS - op: GREATER_THAN - value: 10 - schedule: - quartz_cron_expression: 20 30 * * * ? - timezone_id: Europe/London - pause_status: UNPAUSED - trigger: - pause_status: UNPAUSED - file_arrival: - url: string - min_time_between_triggers_seconds: 0 - wait_after_last_change_seconds: 0 - periodic: - interval: 0 - unit: HOURS - continuous: - pause_status: UNPAUSED - max_concurrent_runs: 10 - tasks: - - max_retries: 3 - task_key: Sessionize - description: Extracts session data from events - min_retry_interval_millis: 2000 - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.Sessionize - parameters: - - --data - - dbfs:/path/to/data.json - libraries: - - jar: dbfs:/mnt/databricks/Sessionize.jar - retry_on_timeout: false - existing_cluster_id: 0923-164208-meows279 - - max_retries: 3 - task_key: Orders_Ingest - description: Ingests order data - job_cluster_key: auto_scaling_cluster - min_retry_interval_millis: 2000 - depends_on: [] - timeout_seconds: 86400 - spark_jar_task: - main_class_name: com.databricks.OrdersIngest - parameters: - - --data - - dbfs:/path/to/order-data.json - libraries: - - jar: dbfs:/mnt/databricks/OrderIngest.jar - retry_on_timeout: false - - max_retries: 3 - task_key: Match - description: Matches orders with user sessions - notebook_task: - base_parameters: - age: '35' - name: John Doe - notebook_path: /Users/user.name@databricks.com/Match - min_retry_interval_millis: 2000 - depends_on: - - task_key: Orders_Ingest - - task_key: Sessionize - new_cluster: - autoscale: - max_workers: 16 - min_workers: 2 - node_type_id: null - spark_conf: - spark.speculation: true - spark_version: 7.3.x-scala2.12 - timeout_seconds: 86400 - retry_on_timeout: false - run_if: ALL_SUCCESS - job_clusters: - - job_cluster_key: auto_scaling_cluster - new_cluster: - autoscale: - max_workers: 16 - min_workers: 2 - node_type_id: null - spark_conf: - spark.speculation: true - spark_version: 7.3.x-scala2.12 - git_source: - git_branch: main - git_provider: gitHub - git_url: https://github.com/databricks/databricks-cli - tags: - cost-center: engineering - team: jobs - format: SINGLE_TASK - queue: - enabled: true - parameters: - - default: users - name: table - run_as: - user_name: user@databricks.com - service_principal_name: 692bc6d0-ffa3-11ed-be56-0242ac120002 - edit_mode: UI_LOCKED - deployment: - kind: BUNDLE - metadata_file_path: string - environments: - - environment_key: string - spec: - client: '1' - dependencies: - - string - fields_to_remove: - - libraries - - schedule - - tasks/task_1 - - job_clusters/Default - description: Add, update, or remove specific settings of an existing job. Use - the - /api/2.0/policies/jobs/enforce-compliance: - post: - operationId: job-compliance-enforcecompliance - externalDocs: - url: https://docs.databricks.com/api/workspace/policycomplianceforjobs/enforcecompliance - x-stackQL-resource: job_compliance - x-stackQL-method: enforcecompliance - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - has_changes: - type: boolean - job_cluster_changes: - type: array - items: - type: object - properties: - field: - type: string - previous_value: - type: string - new_value: - type: string - settings: - type: object - properties: - name: - type: string - description: - type: string - email_notifications: - type: object - properties: - on_start: - type: array - items: - type: string - on_success: - type: array - items: - type: string - on_failure: - type: array - items: - type: string - on_duration_warning_threshold_exceeded: - type: array - items: - type: string - on_streaming_backlog_exceeded: - type: array - items: - type: string - no_alert_for_skipped_runs: - type: boolean - webhook_notifications: - type: object - properties: - on_start: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_success: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_failure: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_duration_warning_threshold_exceeded: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - on_streaming_backlog_exceeded: - type: array - items: - type: array - items: - type: object - properties: - id: - type: string - notification_settings: - type: object - properties: - no_alert_for_skipped_runs: - type: boolean - no_alert_for_canceled_runs: - type: boolean - timeout_seconds: - type: integer - health: - type: object - properties: - rules: - type: array - items: - type: object - properties: - metric: - type: string - op: - type: string - value: - type: integer - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - trigger: - type: object - properties: - pause_status: - type: string - file_arrival: - type: object - properties: - url: - type: string - min_time_between_triggers_seconds: - type: integer - wait_after_last_change_seconds: - type: integer - periodic: - type: object - properties: - interval: - type: integer - unit: - type: string - continuous: - type: object - properties: - pause_status: - type: string - max_concurrent_runs: - type: integer - tasks: - type: array - items: - type: object - properties: - max_retries: - type: integer - task_key: - type: string - description: - type: string - min_retry_interval_millis: - type: integer - depends_on: - type: array - items: {} - timeout_seconds: - type: integer - spark_jar_task: - type: object - properties: - main_class_name: - type: string - parameters: - type: array - items: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - retry_on_timeout: - type: boolean - existing_cluster_id: - type: string - job_clusters: - type: array - items: - type: object - properties: - job_cluster_key: - type: string - new_cluster: - type: object - properties: - autoscale: - type: object - properties: - max_workers: - type: integer - min_workers: - type: integer - node_type_id: - type: 'null' - spark_conf: - type: object - properties: - spark.speculation: - type: boolean - spark_version: - type: string - git_source: - type: object - properties: - git_branch: - type: string - git_provider: - type: string - git_url: - type: string - tags: - type: object - properties: - cost-center: - type: string - team: - type: string - format: - type: string - queue: - type: object - properties: - enabled: - type: boolean - parameters: - type: array - items: - type: object - properties: - default: - type: string - name: - type: string - run_as: - type: object - properties: - user_name: - type: string - service_principal_name: - type: string - edit_mode: - type: string - deployment: - type: object - properties: - kind: - type: string - metadata_file_path: - type: string - environments: - type: array - items: - type: object - properties: - environment_key: - type: string - spec: - type: object - properties: - client: - type: string - dependencies: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - job_id: - type: required - validate_only: - type: integer - format: int64 - example: - job_id: 0 - validate_only: false - description: Updates a job so the job clusters that are created when running - the job (specified in - /api/2.0/policies/jobs/get-compliance: - get: - operationId: job-compliance-getcompliance - externalDocs: - url: https://docs.databricks.com/api/workspace/policycomplianceforjobs/getcompliance - x-stackQL-resource: job_compliance - x-stackQL-method: getcompliance - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: job_id - required: true - description: int64 - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - is_compliant: - type: boolean - violations: - type: object - properties: - job_clusters[job_cluster_key].new_cluster.custom_tags.test_tag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: Returns the policy compliance status of a job. Jobs could be out - of compliance if a cluster policy they use was updated after the job was last - edited and some of its job clusters no longer comply with their updated policies. - /api/2.0/policies/jobs/list-compliance: - get: - operationId: job-compliance-listcompliance - externalDocs: - url: https://docs.databricks.com/api/workspace/policycomplianceforjobs/listcompliance - x-stackQL-resource: job_compliance - x-stackQL-method: listcompliance - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: policy_id - required: true - description: string - in: query - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - jobs: - type: array - items: - type: object - properties: - job_id: - type: integer - is_compliant: - type: boolean - violations: - type: object - properties: - job_clusters[job_cluster_key].new_cluster.custom_tags.test_tag: - type: string - next_page_token: - type: string - prev_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - x-stackQL-objectKey: $.jobs - description: Returns the policy compliance status of all jobs that use a given - policy. Jobs could be out of compliance if a cluster policy they use was updated - after the job was last edited and its job clusters no longer comply with the - updated policy. - /api/2.0/permissions/jobs/{job_id}: - get: - operationId: job-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/getpermissions - x-stackQL-resource: job_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: job_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a job. Jobs can inherit permissions from - their root object. - put: - operationId: job-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/setpermissions - x-stackQL-resource: job_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: job_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: job-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/updatepermissions - x-stackQL-resource: job_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: job_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a job. Jobs can inherit permissions - from their root object. - /api/2.0/permissions/jobs/{job_id}/permissionLevels: - get: - operationId: job-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/getpermissionlevels - x-stackQL-resource: job_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: job_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.2/jobs/runs/cancel-all: - post: - operationId: job-runs-cancelallruns - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/cancelallruns - x-stackQL-resource: job_runs - x-stackQL-method: cancelallruns - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - job_id: - type: integer - format: int64 - all_queued_runs: - type: boolean - example: - job_id: 11223344 - all_queued_runs: false - description: Cancels all active runs of a job. The runs are canceled asynchronously, - so it doesn't prevent new runs from being started. - /api/2.2/jobs/runs/cancel: - post: - operationId: job-runs-cancelrun - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/cancelrun - x-stackQL-resource: job_runs - x-stackQL-method: cancelrun - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: required - example: - run_id: 455644833 - description: Cancels a job run or a task run. The run is canceled asynchronously, - so it may still be running when this request completes. - /api/2.2/jobs/runs/delete: - post: - operationId: job-runs-deleterun - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/deleterun - x-stackQL-resource: job_runs - x-stackQL-method: deleterun - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: required - example: - run_id: 455644833 - description: Deletes a non-active run. Returns an error if the run is active. - /api/2.2/jobs/runs/export: - get: - operationId: job-runs-exportrun - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/exportrun - x-stackQL-resource: job_runs - x-stackQL-method: exportrun - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: run_id - required: true - description: int64 - in: query - - name: views_to_export - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - views: - type: array - items: - type: object - properties: - content: - type: string - name: - type: string - type: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: Export and retrieve the job run task. - /api/2.2/jobs/runs/get: - get: - operationId: job-runs-getrun - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/getrun - x-stackQL-resource: job_runs - x-stackQL-method: getrun - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: run_id - required: true - description: int64 - in: query - - name: include_history - in: query - - name: include_resolved_values - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - job_id: - type: integer - run_id: - type: integer - creator_user_name: - type: string - number_in_job: - type: integer - original_attempt_run_id: - type: integer - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - queue_reason: - type: string - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - cluster_spec: - type: object - properties: - existing_cluster_id: - type: string - new_cluster: - type: object - properties: - num_workers: - type: integer - autoscale: - type: object - properties: - min_workers: - type: integer - max_workers: - type: integer - kind: - type: string - cluster_name: - type: string - spark_version: - type: string - use_ml_runtime: - type: boolean - is_single_node: - type: boolean - spark_conf: - type: object - properties: - property1: - type: string - property2: - type: string - aws_attributes: - type: object - properties: - first_on_demand: - type: string - availability: - type: string - zone_id: - type: string - instance_profile_arn: - type: string - spot_bid_price_percent: - type: string - ebs_volume_type: - type: string - ebs_volume_count: - type: string - ebs_volume_size: - type: integer - ebs_volume_iops: - type: integer - ebs_volume_throughput: - type: integer - node_type_id: - type: string - driver_node_type_id: - type: string - ssh_public_keys: - type: array - items: - type: string - custom_tags: - type: object - properties: - property1: - type: string - property2: - type: string - cluster_log_conf: - type: object - properties: - dbfs: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - init_scripts: - type: array - items: - type: object - properties: - workspace: - type: object - properties: - destination: - type: string - volumes: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - file: - type: object - properties: - destination: - type: string - dbfs: - type: object - properties: - destination: - type: string - abfss: - type: object - properties: - destination: - type: string - gcs: - type: object - properties: - destination: - type: string - spark_env_vars: - type: object - properties: - property1: - type: string - property2: - type: string - autotermination_minutes: - type: integer - enable_elastic_disk: - type: boolean - instance_pool_id: - type: string - policy_id: - type: string - enable_local_disk_encryption: - type: boolean - driver_instance_pool_id: - type: string - workload_type: - type: object - properties: - clients: - type: object - properties: - notebooks: - type: string - jobs: - type: string - runtime_engine: - type: string - docker_image: - type: object - properties: - url: - type: string - basic_auth: - type: object - properties: - username: - type: string - password: - type: string - data_security_mode: - type: string - single_user_name: - type: string - apply_policy_default_values: - type: boolean - job_cluster_key: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - egg: - type: string - pypi: - type: object - properties: - package: - type: string - repo: - type: string - maven: - type: object - properties: - coordinates: - type: string - repo: - type: string - exclusions: - type: array - items: - type: string - cran: - type: object - properties: - package: - type: string - repo: - type: string - whl: - type: string - requirements: - type: string - cluster_instance: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: string - job_parameters: - type: array - items: - type: object - properties: - default: - type: string - name: - type: string - value: - type: string - overriding_parameters: - type: object - properties: - pipeline_params: - type: object - properties: - full_refresh: - type: boolean - start_time: - type: integer - setup_duration: - type: integer - execution_duration: - type: integer - cleanup_duration: - type: integer - end_time: - type: integer - run_duration: - type: integer - queue_duration: - type: integer - trigger: - type: string - trigger_info: - type: object - properties: - run_id: - type: integer - run_name: - type: string - run_page_url: - type: string - run_type: - type: string - tasks: - type: array - items: - type: object - properties: - setup_duration: - type: integer - start_time: - type: integer - task_key: - type: string - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - description: - type: string - job_cluster_key: - type: string - end_time: - type: integer - run_page_url: - type: string - run_id: - type: integer - cluster_instance: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: string - spark_jar_task: - type: object - properties: - main_class_name: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - attempt_number: - type: integer - cleanup_duration: - type: integer - execution_duration: - type: integer - run_if: - type: string - description: - type: string - attempt_number: - type: integer - job_clusters: - type: array - items: - type: object - properties: - job_cluster_key: - type: string - new_cluster: - type: object - properties: - autoscale: - type: object - properties: - max_workers: - type: integer - min_workers: - type: integer - node_type_id: - type: 'null' - spark_conf: - type: object - properties: - spark.speculation: - type: boolean - spark_version: - type: string - git_source: - type: object - properties: - git_branch: - type: string - git_provider: - type: string - git_url: - type: string - repair_history: - type: array - items: - type: object - properties: - type: - type: string - start_time: - type: integer - end_time: - type: integer - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - queue_reason: - type: string - id: - type: integer - task_run_ids: - type: array - items: - type: integer - status: - type: object - properties: - state: - type: string - termination_details: - type: object - properties: - code: - type: string - type: - type: string - message: - type: string - queue_details: - type: object - properties: - code: - type: string - message: - type: string - status: - type: object - properties: - state: - type: string - termination_details: - type: object - properties: - code: - type: string - type: - type: string - message: - type: string - queue_details: - type: object - properties: - code: - type: string - message: - type: string - job_run_id: - type: integer - has_more: - type: boolean - next_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: Retrieves the metadata of a run. - /api/2.2/jobs/runs/list: - get: - operationId: job-runs-listruns - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/listruns - x-stackQL-resource: job_runs - x-stackQL-method: listruns - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: job_id - in: query - - name: active_only - in: query - - name: completed_only - in: query - - name: offset - in: query - - name: limit - in: query - - name: run_type - in: query - - name: expand_tasks - in: query - - name: start_time_from - in: query - - name: start_time_to - in: query - - name: page_token - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - runs: - type: array - items: - type: object - properties: - job_id: - type: integer - run_id: - type: integer - creator_user_name: - type: string - number_in_job: - type: integer - original_attempt_run_id: - type: integer - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - queue_reason: - type: string - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - cluster_spec: - type: object - properties: - existing_cluster_id: - type: string - new_cluster: - type: object - properties: - num_workers: - type: integer - autoscale: - type: object - properties: - min_workers: - type: integer - max_workers: - type: integer - kind: - type: string - cluster_name: - type: string - spark_version: - type: string - use_ml_runtime: - type: boolean - is_single_node: - type: boolean - spark_conf: - type: object - properties: - property1: - type: string - property2: - type: string - aws_attributes: - type: object - properties: - first_on_demand: - type: string - availability: - type: string - zone_id: - type: string - instance_profile_arn: - type: string - spot_bid_price_percent: - type: string - ebs_volume_type: - type: string - ebs_volume_count: - type: string - ebs_volume_size: - type: integer - ebs_volume_iops: - type: integer - ebs_volume_throughput: - type: integer - node_type_id: - type: string - driver_node_type_id: - type: string - ssh_public_keys: - type: array - items: - type: string - custom_tags: - type: object - properties: - property1: - type: string - property2: - type: string - cluster_log_conf: - type: object - properties: - dbfs: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - init_scripts: - type: array - items: - type: object - properties: - workspace: - type: object - properties: - destination: - type: string - volumes: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - file: - type: object - properties: - destination: - type: string - dbfs: - type: object - properties: - destination: - type: string - abfss: - type: object - properties: - destination: - type: string - gcs: - type: object - properties: - destination: - type: string - spark_env_vars: - type: object - properties: - property1: - type: string - property2: - type: string - autotermination_minutes: - type: integer - enable_elastic_disk: - type: boolean - instance_pool_id: - type: string - policy_id: - type: string - enable_local_disk_encryption: - type: boolean - driver_instance_pool_id: - type: string - workload_type: - type: object - properties: - clients: - type: object - properties: - notebooks: - type: string - jobs: - type: string - runtime_engine: - type: string - docker_image: - type: object - properties: - url: - type: string - basic_auth: - type: object - properties: - username: - type: string - password: - type: string - data_security_mode: - type: string - single_user_name: - type: string - apply_policy_default_values: - type: boolean - job_cluster_key: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - egg: - type: string - pypi: - type: object - properties: - package: - type: string - repo: - type: string - maven: - type: object - properties: - coordinates: - type: string - repo: - type: string - exclusions: - type: array - items: - type: string - cran: - type: object - properties: - package: - type: string - repo: - type: string - whl: - type: string - requirements: - type: string - cluster_instance: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: string - job_parameters: - type: array - items: - type: object - properties: - default: - type: string - name: - type: string - value: - type: string - overriding_parameters: - type: object - properties: - pipeline_params: - type: object - properties: - full_refresh: - type: boolean - start_time: - type: integer - setup_duration: - type: integer - execution_duration: - type: integer - cleanup_duration: - type: integer - end_time: - type: integer - run_duration: - type: integer - queue_duration: - type: integer - trigger: - type: string - trigger_info: - type: object - properties: - run_id: - type: integer - run_name: - type: string - run_page_url: - type: string - run_type: - type: string - tasks: - type: array - items: - type: object - properties: - setup_duration: - type: integer - start_time: - type: integer - task_key: - type: string - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - description: - type: string - job_cluster_key: - type: string - end_time: - type: integer - run_page_url: - type: string - run_id: - type: integer - cluster_instance: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: string - spark_jar_task: - type: object - properties: - main_class_name: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - attempt_number: - type: integer - cleanup_duration: - type: integer - execution_duration: - type: integer - run_if: - type: string - description: - type: string - attempt_number: - type: integer - job_clusters: - type: array - items: - type: object - properties: - job_cluster_key: - type: string - new_cluster: - type: object - properties: - autoscale: - type: object - properties: - max_workers: - type: integer - min_workers: - type: integer - node_type_id: - type: 'null' - spark_conf: - type: object - properties: - spark.speculation: - type: boolean - spark_version: - type: string - git_source: - type: object - properties: - git_branch: - type: string - git_provider: - type: string - git_url: - type: string - repair_history: - type: array - items: - type: object - properties: - type: - type: string - start_time: - type: integer - end_time: - type: integer - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - queue_reason: - type: string - id: - type: integer - task_run_ids: - type: array - items: - type: integer - status: - type: object - properties: - state: - type: string - termination_details: - type: object - properties: - code: - type: string - type: - type: string - message: - type: string - queue_details: - type: object - properties: - code: - type: string - message: - type: string - status: - type: object - properties: - state: - type: string - termination_details: - type: object - properties: - code: - type: string - type: - type: string - message: - type: string - queue_details: - type: object - properties: - code: - type: string - message: - type: string - job_run_id: - type: integer - has_more: - type: boolean - next_page_token: - type: string - prev_page_token: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - x-stackQL-objectKey: $.runs - description: List runs in descending order by start time. - /api/2.2/jobs/runs/repair: - post: - operationId: job-runs-repairrun - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/repairrun - x-stackQL-resource: job_runs - x-stackQL-method: repairrun - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - repair_id: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - run_id: - type: required - latest_repair_id: - type: integer - format: int64 - rerun_tasks: - type: integer - format: int64 - rerun_all_failed_tasks: - type: Array of string - rerun_dependent_tasks: - type: boolean - job_parameters: - type: object - pipeline_params: - type: object - example: - run_id: 455644833 - latest_repair_id: 734650698524280 - rerun_tasks: - - task0 - - task1 - job_parameters: - property1: string - property2: string - pipeline_params: - full_refresh: false - rerun_all_failed_tasks: false - rerun_dependent_tasks: false - description: Re-run one or more tasks. Tasks are re-run as part of the original - job run. They use the current job and task settings, and can be viewed in - the history for the original job run. - /api/2.2/jobs/runs/get-output: - get: - operationId: job-run-output-getrunoutput - externalDocs: - url: https://docs.databricks.com/api/workspace/jobs/getrunoutput - x-stackQL-resource: job_run_output - x-stackQL-method: getrunoutput - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: run_id - required: true - description: int64 - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - metadata: - type: object - properties: - job_id: - type: integer - run_id: - type: integer - creator_user_name: - type: string - number_in_job: - type: integer - original_attempt_run_id: - type: integer - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - queue_reason: - type: string - schedule: - type: object - properties: - quartz_cron_expression: - type: string - timezone_id: - type: string - pause_status: - type: string - cluster_spec: - type: object - properties: - existing_cluster_id: - type: string - new_cluster: - type: object - properties: - num_workers: - type: integer - autoscale: - type: object - properties: - min_workers: - type: integer - max_workers: - type: integer - kind: - type: string - cluster_name: - type: string - spark_version: - type: string - use_ml_runtime: - type: boolean - is_single_node: - type: boolean - spark_conf: - type: object - properties: - property1: - type: string - property2: - type: string - aws_attributes: - type: object - properties: - first_on_demand: - type: string - availability: - type: string - zone_id: - type: string - instance_profile_arn: - type: string - spot_bid_price_percent: - type: string - ebs_volume_type: - type: string - ebs_volume_count: - type: string - ebs_volume_size: - type: integer - ebs_volume_iops: - type: integer - ebs_volume_throughput: - type: integer - node_type_id: - type: string - driver_node_type_id: - type: string - ssh_public_keys: - type: array - items: - type: string - custom_tags: - type: object - properties: - property1: - type: string - property2: - type: string - cluster_log_conf: - type: object - properties: - dbfs: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - init_scripts: - type: array - items: - type: object - properties: - workspace: - type: object - properties: - destination: - type: string - volumes: - type: object - properties: - destination: - type: string - s3: - type: object - properties: - destination: - type: string - region: - type: string - endpoint: - type: string - enable_encryption: - type: boolean - encryption_type: - type: string - kms_key: - type: string - canned_acl: - type: string - file: - type: object - properties: - destination: - type: string - dbfs: - type: object - properties: - destination: - type: string - abfss: - type: object - properties: - destination: - type: string - gcs: - type: object - properties: - destination: - type: string - spark_env_vars: - type: object - properties: - property1: - type: string - property2: - type: string - autotermination_minutes: - type: integer - enable_elastic_disk: - type: boolean - instance_pool_id: - type: string - policy_id: - type: string - enable_local_disk_encryption: - type: boolean - driver_instance_pool_id: - type: string - workload_type: - type: object - properties: - clients: - type: object - properties: - notebooks: - type: string - jobs: - type: string - runtime_engine: - type: string - docker_image: - type: object - properties: - url: - type: string - basic_auth: - type: object - properties: - username: - type: string - password: - type: string - data_security_mode: - type: string - single_user_name: - type: string - apply_policy_default_values: - type: boolean - job_cluster_key: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - egg: - type: string - pypi: - type: object - properties: - package: - type: string - repo: - type: string - maven: - type: object - properties: - coordinates: - type: string - repo: - type: string - exclusions: - type: array - items: - type: string - cran: - type: object - properties: - package: - type: string - repo: - type: string - whl: - type: string - requirements: - type: string - cluster_instance: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: string - job_parameters: - type: array - items: - type: object - properties: - default: - type: string - name: - type: string - value: - type: string - overriding_parameters: - type: object - properties: - pipeline_params: - type: object - properties: - full_refresh: - type: boolean - start_time: - type: integer - setup_duration: - type: integer - execution_duration: - type: integer - cleanup_duration: - type: integer - end_time: - type: integer - run_duration: - type: integer - queue_duration: - type: integer - trigger: - type: string - trigger_info: - type: object - properties: - run_id: - type: integer - run_name: - type: string - run_page_url: - type: string - run_type: - type: string - tasks: - type: array - items: - type: object - properties: - setup_duration: - type: integer - start_time: - type: integer - task_key: - type: string - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - description: - type: string - job_cluster_key: - type: string - end_time: - type: integer - run_page_url: - type: string - run_id: - type: integer - cluster_instance: - type: object - properties: - cluster_id: - type: string - spark_context_id: - type: string - spark_jar_task: - type: object - properties: - main_class_name: - type: string - libraries: - type: array - items: - type: object - properties: - jar: - type: string - attempt_number: - type: integer - cleanup_duration: - type: integer - execution_duration: - type: integer - run_if: - type: string - description: - type: string - attempt_number: - type: integer - job_clusters: - type: array - items: - type: object - properties: - job_cluster_key: - type: string - new_cluster: - type: object - properties: - autoscale: - type: object - properties: - max_workers: - type: integer - min_workers: - type: integer - node_type_id: - type: 'null' - spark_conf: - type: object - properties: - spark.speculation: - type: boolean - spark_version: - type: string - git_source: - type: object - properties: - git_branch: - type: string - git_provider: - type: string - git_url: - type: string - repair_history: - type: array - items: - type: object - properties: - type: - type: string - start_time: - type: integer - end_time: - type: integer - state: - type: object - properties: - life_cycle_state: - type: string - result_state: - type: string - state_message: - type: string - user_cancelled_or_timedout: - type: boolean - queue_reason: - type: string - id: - type: integer - task_run_ids: - type: array - items: - type: integer - status: - type: object - properties: - state: - type: string - termination_details: - type: object - properties: - code: - type: string - type: - type: string - message: - type: string - queue_details: - type: object - properties: - code: - type: string - message: - type: string - status: - type: object - properties: - state: - type: string - termination_details: - type: object - properties: - code: - type: string - type: - type: string - message: - type: string - queue_details: - type: object - properties: - code: - type: string - message: - type: string - job_run_id: - type: integer - has_more: - type: boolean - next_page_token: - type: string - error: - type: string - logs: - type: string - logs_truncated: - type: boolean - error_trace: - type: string - info: - type: string - notebook_output: - type: object - properties: - result: - type: string - truncated: - type: boolean - sql_output: - type: object - properties: - query_output: - type: object - properties: - query_text: - type: string - endpoint_id: - type: string - sql_statements: - type: array - items: - type: object - properties: - lookup_key: - type: string - output_link: - type: string - warehouse_id: - type: string - dashboard_output: - type: object - properties: - widgets: - type: array - items: - type: object - properties: - widget_id: - type: string - widget_title: - type: string - output_link: - type: string - status: - type: string - error: - type: object - properties: - message: - type: string - start_time: - type: integer - end_time: - type: integer - warehouse_id: - type: string - alert_output: - type: object - properties: - query_text: - type: string - sql_statements: - type: array - items: - type: object - properties: - lookup_key: - type: string - output_link: - type: string - warehouse_id: - type: string - alert_state: - type: string - dbt_output: - type: object - properties: - artifacts_link: - type: string - artifacts_headers: - type: object - properties: - property1: - type: string - property2: - type: string - run_job_output: - type: object - properties: - run_id: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '429': - description: Operation is rejected due to throttling, e.g. some resource - has been exhausted, per-user quota. - '500': - description: Internal error. - description: Retrieve the output and metadata of a single task run. When a notebook - task returns a value through the -components: - x-stackQL-resources: - jobs: - id: databricks_workspace.workflows.jobs - name: jobs - title: Jobs - methods: - create: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.jobs - reset: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1reset/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - runnow: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1run-now/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - submit: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1submit/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1update/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/jobs/methods/list' - - $ref: '#/components/x-stackQL-resources/jobs/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/jobs/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/jobs/methods/update' - replace: - - $ref: '#/components/x-stackQL-resources/jobs/methods/reset' - delete: - - $ref: '#/components/x-stackQL-resources/jobs/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/jobs/methods/submit' - - $ref: '#/components/x-stackQL-resources/jobs/methods/runnow' - job_compliance: - id: databricks_workspace.workflows.job_compliance - name: job_compliance - title: Job_compliance - methods: - enforcecompliance: - operation: - $ref: '#/paths/~1api~12.0~1policies~1jobs~1enforce-compliance/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getcompliance: - operation: - $ref: '#/paths/~1api~12.0~1policies~1jobs~1get-compliance/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listcompliance: - operation: - $ref: '#/paths/~1api~12.0~1policies~1jobs~1list-compliance/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.jobs - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/job_compliance/methods/listcompliance' - - $ref: '#/components/x-stackQL-resources/job_compliance/methods/getcompliance' - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/job_compliance/methods/enforcecompliance' - delete: [] - exec: [] - job_permissions: - id: databricks_workspace.workflows.job_permissions - name: job_permissions - title: Job_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1jobs~1{job_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1jobs~1{job_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1jobs~1{job_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/job_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/job_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/job_permissions/methods/setpermissions' - delete: [] - exec: [] - job_permission_levels: - id: databricks_workspace.workflows.job_permission_levels - name: job_permission_levels - title: Job_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1jobs~1{job_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/job_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - job_runs: - id: databricks_workspace.workflows.job_runs - name: job_runs - title: Job_runs - methods: - cancelallruns: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1cancel-all/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - cancelrun: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1cancel/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - deleterun: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - exportrun: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1export/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - getrun: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1get/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listruns: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.runs - repairrun: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1repair/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/job_runs/methods/listruns' - - $ref: '#/components/x-stackQL-resources/job_runs/methods/getrun' - insert: [] - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/job_runs/methods/deleterun' - exec: - - $ref: '#/components/x-stackQL-resources/job_runs/methods/repairrun' - - $ref: '#/components/x-stackQL-resources/job_runs/methods/exportrun' - - $ref: '#/components/x-stackQL-resources/job_runs/methods/cancelrun' - - $ref: '#/components/x-stackQL-resources/job_runs/methods/cancelallruns' - job_run_output: - id: databricks_workspace.workflows.job_run_output - name: job_run_output - title: Job_run_output - methods: - getrunoutput: - operation: - $ref: '#/paths/~1api~12.2~1jobs~1runs~1get-output/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/job_run_output/methods/getrunoutput' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] diff --git a/providers/src/databricks_workspace/v00.00.00000/services/workspace.yaml b/providers/src/databricks_workspace/v00.00.00000/services/workspace.yaml index d9c75d98..73f97dbc 100644 --- a/providers/src/databricks_workspace/v00.00.00000/services/workspace.yaml +++ b/providers/src/databricks_workspace/v00.00.00000/services/workspace.yaml @@ -1,3401 +1,3424 @@ -openapi: 3.0.0 -info: - version: 2024-12-19-stackql-generated - contact: - name: StackQL Studios - url: https://stackql.io/ - email: info@stackql.io - title: Databricks Workspace API - description: Workspace -servers: -- url: https://{deployment_name}.cloud.databricks.com - variables: - deployment_name: - description: The Databricks Workspace Deployment Name - default: dbc-abcd0123-a1bc -paths: - /api/2.0/settings/types/automatic_cluster_update/names/default: - get: - operationId: automatic-cluster-update-get - externalDocs: - url: https://docs.databricks.com/api/workspace/automaticclusterupdate/get - x-stackQL-resource: automatic_cluster_update - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - automatic_cluster_update_workspace: - type: object - properties: - enabled: - type: boolean - can_toggle: - type: boolean - maintenance_window: - type: object - properties: - week_day_based_schedule: - type: object - properties: - frequency: - type: string - day_of_week: - type: string - window_start_time: - type: object - properties: - hours: - type: integer - minutes: - type: integer - enablement_details: - type: object - properties: - unavailable_for_non_enterprise_tier: - type: boolean - unavailable_for_disabled_entitlement: - type: boolean - forced_for_compliance_mode: - type: boolean - restart_even_if_no_updates_available: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the automatic cluster update setting. - patch: - operationId: automatic-cluster-update-update - externalDocs: - url: https://docs.databricks.com/api/workspace/automaticclusterupdate/update - x-stackQL-resource: automatic_cluster_update - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - automatic_cluster_update_workspace: - type: object - properties: - enabled: - type: boolean - can_toggle: - type: boolean - maintenance_window: - type: object - properties: - week_day_based_schedule: - type: object - properties: - frequency: - type: string - day_of_week: - type: string - window_start_time: - type: object - properties: - hours: - type: integer - minutes: - type: integer - enablement_details: - type: object - properties: - unavailable_for_non_enterprise_tier: - type: boolean - unavailable_for_disabled_entitlement: - type: boolean - forced_for_compliance_mode: - type: boolean - restart_even_if_no_updates_available: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - automatic_cluster_update_workspace: - enabled: true - can_toggle: true - maintenance_window: - week_day_based_schedule: - frequency: FIRST_OF_MONTH - day_of_week: MONDAY - window_start_time: - hours: 0 - minutes: 0 - enablement_details: - unavailable_for_non_enterprise_tier: true - unavailable_for_disabled_entitlement: true - forced_for_compliance_mode: true - restart_even_if_no_updates_available: true - field_mask: string - description: Details required to update a setting. - description: Updates the automatic cluster update setting for the workspace. - A fresh etag needs to be provided in - /api/2.0/settings/types/shield_csp_enablement_ws_db/names/default: - get: - operationId: compliance-security-profile-get - externalDocs: - url: https://docs.databricks.com/api/workspace/compliancesecurityprofile/get - x-stackQL-resource: compliance_security_profile - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - compliance_security_profile_workspace: - type: object - properties: - is_enabled: - type: boolean - compliance_standards: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the compliance security profile setting. - patch: - operationId: compliance-security-profile-update - externalDocs: - url: https://docs.databricks.com/api/workspace/compliancesecurityprofile/update - x-stackQL-resource: compliance_security_profile - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - compliance_security_profile_workspace: - type: object - properties: - is_enabled: - type: boolean - compliance_standards: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - compliance_security_profile_workspace: - is_enabled: true - compliance_standards: - - NONE - field_mask: string - description: Details required to update a setting. - description: Updates the compliance security profile setting for the workspace. - A fresh etag needs to be provided in - /api/2.0/settings/types/aibi_dash_embed_ws_acc_policy/names/default: - delete: - operationId: dashboard-embedding-access-policy-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/aibidashboardembeddingaccesspolicy/delete - x-stackQL-resource: dashboard_embedding_access_policy - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Delete the AI/BI dashboard embedding access policy, reverting back - to the default. - get: - operationId: dashboard-embedding-access-policy-get - externalDocs: - url: https://docs.databricks.com/api/workspace/aibidashboardembeddingaccesspolicy/get - x-stackQL-resource: dashboard_embedding_access_policy - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - aibi_dashboard_embedding_access_policy: - type: object - properties: - access_policy_type: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves the AI/BI dashboard embedding access policy. The default - setting is ALLOW_APPROVED_DOMAINS, permitting AI/BI dashboards to be embedded - on approved domains. - patch: - operationId: dashboard-embedding-access-policy-update - externalDocs: - url: https://docs.databricks.com/api/workspace/aibidashboardembeddingaccesspolicy/update - x-stackQL-resource: dashboard_embedding_access_policy - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - aibi_dashboard_embedding_access_policy: - type: object - properties: - access_policy_type: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - aibi_dashboard_embedding_access_policy: - access_policy_type: ALLOW_ALL_DOMAINS - field_mask: string - description: Details required to update a setting. - description: Updates the AI/BI dashboard embedding access policy at the workspace - level. - /api/2.0/settings/types/aibi_dash_embed_ws_apprvd_domains/names/default: - delete: - operationId: dashboard-embedding-approved-domains-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/aibidashboardembeddingapproveddomains/delete - x-stackQL-resource: dashboard_embedding_approved_domains - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Delete the list of domains approved to host embedded AI/BI dashboards, - reverting back to the default empty list. - get: - operationId: dashboard-embedding-approved-domains-get - externalDocs: - url: https://docs.databricks.com/api/workspace/aibidashboardembeddingapproveddomains/get - x-stackQL-resource: dashboard_embedding_approved_domains - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - aibi_dashboard_embedding_approved_domains: - type: object - properties: - approved_domains: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Retrieves the list of domains approved to host embedded AI/BI dashboards. - patch: - operationId: dashboard-embedding-approved-domains-update - externalDocs: - url: https://docs.databricks.com/api/workspace/aibidashboardembeddingapproveddomains/update - x-stackQL-resource: dashboard_embedding_approved_domains - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - aibi_dashboard_embedding_approved_domains: - type: object - properties: - approved_domains: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - aibi_dashboard_embedding_approved_domains: - approved_domains: - - string - field_mask: string - description: Details required to update a setting. - description: Updates the list of domains approved to host embedded AI/BI dashboards. - This update will fail if the current workspace access policy is not ALLOW_APPROVED_DOMAINS. - /api/2.0/workspace/list: - get: - operationId: directories-list - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/list - x-stackQL-resource: directories - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: path - required: true - description: string - in: query - - name: notebooks_modified_after - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - objects: - type: array - items: - type: object - properties: - path: - type: string - object_type: - type: string - object_id: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.objects - description: Lists the contents of a directory, or the object if it is not a - directory. If the input path does not exist, this call returns an error - /api/2.0/workspace/mkdirs: - post: - operationId: directories-mkdirs - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/mkdirs - x-stackQL-resource: directories - x-stackQL-method: mkdirs - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - path: - type: required - example: - path: /Users/user@example.com/project - description: Creates the specified directory (and necessary parent directories - if they do not exist). If there is an object (not a directory) at any prefix - of the input path, this call returns an error - /api/2.0/settings/types/shield_esm_enablement_ws_db/names/default: - get: - operationId: enhanced-security-monitoring-get - externalDocs: - url: https://docs.databricks.com/api/workspace/enhancedsecuritymonitoring/get - x-stackQL-resource: enhanced_security_monitoring - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - enhanced_security_monitoring_workspace: - type: object - properties: - is_enabled: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the enhanced security monitoring setting. - patch: - operationId: enhanced-security-monitoring-update - externalDocs: - url: https://docs.databricks.com/api/workspace/enhancedsecuritymonitoring/update - x-stackQL-resource: enhanced_security_monitoring - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - enhanced_security_monitoring_workspace: - type: object - properties: - is_enabled: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - enhanced_security_monitoring_workspace: - is_enabled: true - field_mask: string - description: Details required to update a setting. - description: Updates the enhanced security monitoring setting for the workspace. - A fresh etag needs to be provided in - /api/2.0/workspace-conf: - get: - operationId: feature-status-getstatus - externalDocs: - url: https://docs.databricks.com/api/workspace/workspaceconf/getstatus - x-stackQL-resource: feature_status - x-stackQL-method: getstatus - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: keys - required: true - description: string - in: query - responses: - '200': - description: Status was returned successfully. - content: - application/json: - schema: - type: object - properties: - property1: - type: string - property2: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Enables or disables a specified feature for a workspace. - patch: - operationId: feature-status-setstatus - externalDocs: - url: https://docs.databricks.com/api/workspace/workspaceconf/setstatus - x-stackQL-resource: feature_status - x-stackQL-method: setstatus - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '204': - description: Enabling or disabling feature was successful. - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Enables or disables a specified feature for a workspace. - /api/2.0/ip-access-lists: - post: - operationId: ip-access-lists-create - externalDocs: - url: https://docs.databricks.com/api/workspace/ipaccesslists/create - x-stackQL-resource: ip_access_lists - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - ip_access_list: - type: object - properties: - list_id: - type: string - label: - type: string - ip_addresses: - type: array - items: - type: string - address_count: - type: integer - list_type: - type: string - created_at: - type: integer - created_by: - type: integer - updated_at: - type: integer - updated_by: - type: integer - enabled: - type: boolean - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - label: - type: required - list_type: - type: string - ip_addresses: - type: required - example: - label: Office VPN - list_type: ALLOW - ip_addresses: - - 192.168.100.0/22 - description: Details required to configure a block list or allow list. - description: Creates an IP access list for this workspace. - get: - operationId: ip-access-lists-list - externalDocs: - url: https://docs.databricks.com/api/workspace/ipaccesslists/list - x-stackQL-resource: ip_access_lists - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - ip_access_lists: - type: array - items: - type: object - properties: - list_id: - type: string - label: - type: string - ip_addresses: - type: array - items: - type: string - address_count: - type: integer - list_type: - type: string - created_at: - type: integer - created_by: - type: integer - updated_at: - type: integer - updated_by: - type: integer - enabled: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.ip_access_lists - description: Gets all IP access lists for the specified workspace. - /api/2.0/ip-access-lists/{ip_access_list_id}: - delete: - operationId: ip-access-lists-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/ipaccesslists/delete - x-stackQL-resource: ip_access_lists - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes an IP access list, specified by its list ID. - get: - operationId: ip-access-lists-get - externalDocs: - url: https://docs.databricks.com/api/workspace/ipaccesslists/get - x-stackQL-resource: ip_access_lists - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - ip_access_list: - type: object - properties: - list_id: - type: string - label: - type: string - ip_addresses: - type: array - items: - type: string - address_count: - type: integer - list_type: - type: string - created_at: - type: integer - created_by: - type: integer - updated_at: - type: integer - updated_by: - type: integer - enabled: - type: boolean - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.ip_access_list - description: Gets an IP access list, specified by its list ID. - put: - operationId: ip-access-lists-replace - externalDocs: - url: https://docs.databricks.com/api/workspace/ipaccesslists/replace - x-stackQL-resource: ip_access_lists - x-stackQL-method: replace - x-stackQL-verb: replace - x-numReqParams: 1 - parameters: - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - label: - type: required - list_type: - type: string - ip_addresses: - type: required - enabled: - type: string - example: - label: Office VPN - list_type: ALLOW - ip_addresses: - - 192.168.100.0/22 - enabled: true - description: Details required to replace an IP access list. - description: Replaces an IP access list, specified by its ID. - patch: - operationId: ip-access-lists-update - externalDocs: - url: https://docs.databricks.com/api/workspace/ipaccesslists/update - x-stackQL-resource: ip_access_lists - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: ip_access_list_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - label: - type: string - list_type: - type: string - ip_addresses: - type: Array of string - enabled: - type: boolean - example: - label: Office VPN - list_type: ALLOW - ip_addresses: - - 192.168.100.0/22 - enabled: true - description: Details required to update an IP access list. - description: Updates an existing IP access list, specified by its ID. - /api/2.0/notification-destinations: - post: - operationId: notification-destinations-create - externalDocs: - url: https://docs.databricks.com/api/workspace/notificationdestinations/create - x-stackQL-resource: notification_destinations - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - config: - type: object - properties: - generic_webhook: - type: object - properties: - password_set: - type: boolean - url_set: - type: boolean - username_set: - type: boolean - destination_type: - type: string - display_name: - type: string - id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - display_name: - type: string - config: - type: object - example: - config: - generic_webhook: - password: my-password - url: https://my-webhook.com - username: my-username - display_name: My webhook destination - description: Creates a notification destination. Requires workspace admin permissions. - get: - operationId: notification-destinations-list - externalDocs: - url: https://docs.databricks.com/api/workspace/notificationdestinations/list - x-stackQL-resource: notification_destinations - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: page_token - in: query - - name: page_size - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - next_page_token: - type: string - results: - type: array - items: - type: object - properties: - destination_type: - type: string - display_name: - type: string - id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.results - description: Lists notification destinations. - /api/2.0/notification-destinations/{id}: - delete: - operationId: notification-destinations-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/notificationdestinations/delete - x-stackQL-resource: notification_destinations - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a notification destination. Requires workspace admin permissions. - get: - operationId: notification-destinations-get - externalDocs: - url: https://docs.databricks.com/api/workspace/notificationdestinations/get - x-stackQL-resource: notification_destinations - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - destination_type: - type: string - display_name: - type: string - id: - type: string - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets a notification destination. - patch: - operationId: notification-destinations-update - externalDocs: - url: https://docs.databricks.com/api/workspace/notificationdestinations/update - x-stackQL-resource: notification_destinations - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 1 - parameters: - - name: id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - config: - type: object - properties: - slack: - type: object - properties: - url_set: - type: boolean - destination_type: - type: string - display_name: - type: string - id: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - display_name: - type: string - config: - type: object - example: - config: - slack: - url: https://hooks.slack.com/services/T00000000/B00000000/AAAAAAAAAAAAAAAA - display_name: My slack destination updated - id: fe25e731-92f2-4838-9fb2-1ca364320a3d - description: Updates a notification destination. Requires workspace admin permissions. - At least one field is required in the request body. - /api/2.0/workspace/delete: - post: - operationId: objects-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/delete - x-stackQL-resource: objects - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - path: - type: required - recursive: - type: string - example: - path: /Users/me@example.com/MyFolder - recursive: true - description: Deletes an object or a directory (and optionally recursively deletes - all objects in the directory). - /api/2.0/workspace/export: - get: - operationId: objects-export - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/export - x-stackQL-resource: objects - x-stackQL-method: export - x-stackQL-verb: exec - x-numReqParams: 1 - parameters: - - name: path - required: true - description: string - in: query - - name: format - in: query - - name: direct_download - in: query - responses: - '200': - description: The request field `direct_download` determines whether a JSON - response or binary contents are returned by this endpoint. - content: - application/json: - schema: - type: object - properties: - content: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Exports an object or the contents of an entire directory. - /api/2.0/workspace/import: - post: - operationId: objects-import - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/import - x-stackQL-resource: objects - x-stackQL-method: import - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: {} - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - path: - type: required - format: - type: string - language: - type: string - content: - type: string - overwrite: - type: string - example: - format: SOURCE - path: /Users/user@example.com/project/ScalaExampleNotebook - language: SCALA - content: MSsx - overwrite: true - description: Imports a workspace object (for example, a notebook or file) or - the contents of an entire directory. If - /api/2.0/permissions/{workspace_object_type}/{workspace_object_id}: - get: - operationId: object-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/getpermissions - x-stackQL-resource: object_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: workspace_object_type - required: true - description: string - in: path - - name: workspace_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of a workspace object. Workspace objects can - inherit permissions from their parent objects or root object. - put: - operationId: object-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/setpermissions - x-stackQL-resource: object_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 2 - parameters: - - name: workspace_object_type - required: true - description: string - in: path - - name: workspace_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their parent objects or root object. - patch: - operationId: object-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/updatepermissions - x-stackQL-resource: object_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 2 - parameters: - - name: workspace_object_type - required: true - description: string - in: path - - name: workspace_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_MANAGE - description: Updates the permissions on a workspace object. Workspace objects - can inherit permissions from their parent objects or root object. - /api/2.0/permissions/{workspace_object_type}/{workspace_object_id}/permissionLevels: - get: - operationId: object-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/getpermissionlevels - x-stackQL-resource: object_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 2 - parameters: - - name: workspace_object_type - required: true - description: string - in: path - - name: workspace_object_id - required: true - description: string - in: path - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/workspace/get-status: - get: - operationId: object-status-getstatus - externalDocs: - url: https://docs.databricks.com/api/workspace/workspace/getstatus - x-stackQL-resource: object_status - x-stackQL-method: getstatus - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: path - required: true - description: string - in: query - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - path: - type: string - language: - type: string - object_type: - type: string - object_id: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the status of an object or a directory. If - /api/2.0/settings/types/default_namespace_ws/names/default: - delete: - operationId: permissions-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/defaultnamespace/delete - x-stackQL-resource: permissions - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Deletes the default namespace setting for the workspace. A fresh - etag needs to be provided in - get: - operationId: permissions-get - externalDocs: - url: https://docs.databricks.com/api/workspace/defaultnamespace/get - x-stackQL-resource: permissions - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - namespace: - type: object - properties: - value: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the default namespace setting. - patch: - operationId: permissions-update - externalDocs: - url: https://docs.databricks.com/api/workspace/defaultnamespace/update - x-stackQL-resource: permissions - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - namespace: - type: object - properties: - value: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - namespace: - value: string - field_mask: string - description: Details required to update a setting. - description: Updates the default namespace setting for the workspace. A fresh - etag needs to be provided in - /api/2.0/settings/types/restrict_workspace_admins/names/default: - delete: - operationId: restrict-admins-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/restrictworkspaceadmins/delete - x-stackQL-resource: restrict_admins - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - description: Reverts the restrict workspace admins setting status for the workspace. - A fresh etag needs to be provided in - get: - operationId: restrict-admins-get - externalDocs: - url: https://docs.databricks.com/api/workspace/restrictworkspaceadmins/get - x-stackQL-resource: restrict_admins - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: etag - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - restrict_workspace_admins: - type: object - properties: - status: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the restrict workspace admins setting. - patch: - operationId: restrict-admins-update - externalDocs: - url: https://docs.databricks.com/api/workspace/restrictworkspaceadmins/update - x-stackQL-resource: restrict_admins - x-stackQL-method: update - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - etag: - type: string - setting_name: - type: string - restrict_workspace_admins: - type: object - properties: - status: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '409': - description: Request was rejected due a conflict with an existing resource. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - allow_missing: - type: required - field_mask: - type: boolean - setting: - type: object - example: - allow_missing: true - setting: - etag: string - setting_name: string - restrict_workspace_admins: - status: ALLOW_ALL - field_mask: string - description: Details required to update a setting. - description: Updates the restrict workspace admins setting for the workspace. - A fresh etag needs to be provided in - /api/2.0/token-management/on-behalf-of/tokens: - post: - operationId: tokens-createobotoken - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/createobotoken - x-stackQL-resource: tokens - x-stackQL-method: createobotoken - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - token_value: - type: string - token_info: - type: object - properties: - token_id: - type: string - creation_time: - type: integer - expiry_time: - type: integer - comment: - type: string - created_by_id: - type: integer - created_by_username: - type: string - owner_id: - type: integer - workspace_id: - type: integer - last_used_day: - type: integer - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - application_id: - type: required - lifetime_seconds: - type: string - comment: - type: integer - format: int64 - example: - application_id: 6f5ccf28-d83a-4957-9bfb-5bbfac551410 - lifetime_seconds: 3600 - comment: This is for the ABC department automation scripts. - description: Configuration details for creating on-behalf tokens. - description: Creates a token on behalf of a service principal. - /api/2.0/token-management/tokens/{token_id}: - delete: - operationId: tokens-delete - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/delete - x-stackQL-resource: tokens - x-stackQL-method: delete - x-stackQL-verb: delete - x-numReqParams: 1 - parameters: - - name: token_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Deletes a token, specified by its ID. - get: - operationId: tokens-get - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/get - x-stackQL-resource: tokens - x-stackQL-method: get - x-stackQL-verb: select - x-numReqParams: 1 - parameters: - - name: token_id - required: true - description: string - in: path - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - token_info: - type: object - properties: - token_id: - type: string - creation_time: - type: integer - expiry_time: - type: integer - comment: - type: string - created_by_id: - type: integer - created_by_username: - type: string - owner_id: - type: integer - workspace_id: - type: integer - last_used_day: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.token_info - description: Gets information about a token, specified by its ID. - /api/2.0/token-management/tokens: - get: - operationId: tokens-list - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/list - x-stackQL-resource: tokens - x-stackQL-method: list - x-stackQL-verb: select - x-numReqParams: 0 - parameters: - - name: created_by_id - in: query - - name: created_by_username - in: query - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - token_infos: - type: array - items: - type: object - properties: - token_id: - type: string - creation_time: - type: integer - expiry_time: - type: integer - comment: - type: string - created_by_id: - type: integer - created_by_username: - type: string - owner_id: - type: integer - workspace_id: - type: integer - last_used_day: - type: integer - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '500': - description: Internal error. - x-stackQL-objectKey: $.token_infos - description: Lists all tokens associated with the specified workspace or user. - /api/2.0/permissions/authorization/tokens: - get: - operationId: token-permissions-getpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/getpermissions - x-stackQL-resource: token_permissions - x-stackQL-method: getpermissions - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - description: Gets the permissions of all tokens. Tokens can inherit permissions - from their root object. - put: - operationId: token-permissions-setpermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/setpermissions - x-stackQL-resource: token_permissions - x-stackQL-method: setpermissions - x-stackQL-verb: replace - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_USE - description: Sets permissions on an object, replacing existing permissions if - they exist. Deletes all direct permissions if none are specified. Objects - can inherit permissions from their root object. - patch: - operationId: token-permissions-updatepermissions - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/updatepermissions - x-stackQL-resource: token_permissions - x-stackQL-method: updatepermissions - x-stackQL-verb: update - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - object_id: - type: string - object_type: - type: string - access_control_list: - type: array - items: - type: object - properties: - user_name: - type: string - group_name: - type: string - service_principal_name: - type: string - display_name: - type: string - all_permissions: - type: array - items: - type: object - properties: - permission_level: - type: string - inherited: - type: boolean - inherited_from_object: - type: array - items: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - access_control_list: - type: Array of object - example: - access_control_list: - - user_name: string - group_name: string - service_principal_name: string - permission_level: CAN_USE - description: Updates the permissions on all tokens. Tokens can inherit permissions - from their root object. - /api/2.0/permissions/authorization/tokens/permissionLevels: - get: - operationId: token-permission-levels-getpermissionlevels - externalDocs: - url: https://docs.databricks.com/api/workspace/tokenmanagement/getpermissionlevels - x-stackQL-resource: token_permission_levels - x-stackQL-method: getpermissionlevels - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: '' - content: - application/json: - schema: - type: object - properties: - permission_levels: - type: array - items: - type: object - properties: - permission_level: - type: string - description: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '403': - description: Caller does not have permission to execute the specified operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - x-stackQL-objectKey: $.permission_levels - description: Gets the permission levels that a user can have on an object. - /api/2.0/token/create: - post: - operationId: user-tokens-create - externalDocs: - url: https://docs.databricks.com/api/workspace/tokens/create - x-stackQL-resource: user_tokens - x-stackQL-method: create - x-stackQL-verb: insert - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - token_value: - type: string - token_info: - type: object - properties: - token_id: - type: string - creation_time: - type: integer - expiry_time: - type: integer - comment: - type: string - '400': - description: Request is invalid or malformed. - '401': - description: The request does not have valid authentication credentials - for the operation. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - lifetime_seconds: - type: integer - format: int64 - comment: - type: string - example: - lifetime_seconds: 0 - comment: string - description: Creates and returns a token for a user. If this call is made through - token authentication, it creates a token with the same client ID as the authenticated - token. If the user's token quota is exceeded, this call returns an error - /api/2.0/token/list: - get: - operationId: user-tokens-listtokens - externalDocs: - url: https://docs.databricks.com/api/workspace/tokens/listtokens - x-stackQL-resource: user_tokens - x-stackQL-method: listtokens - x-stackQL-verb: select - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: - token_infos: - type: array - items: - type: object - properties: - token_id: - type: string - creation_time: - type: integer - expiry_time: - type: integer - comment: - type: string - x-stackQL-objectKey: $.token_infos - description: Lists all the valid tokens for a user-workspace pair. - /api/2.0/token/delete: - post: - operationId: user-tokens-revoketoken - externalDocs: - url: https://docs.databricks.com/api/workspace/tokens/revoketoken - x-stackQL-resource: user_tokens - x-stackQL-method: revoketoken - x-stackQL-verb: exec - x-numReqParams: 0 - parameters: [] - responses: - '200': - description: Request completed successfully. - content: - application/json: - schema: - type: object - properties: {} - '401': - description: The request does not have valid authentication credentials - for the operation. - '404': - description: Operation was performed on a resource that does not exist. - '500': - description: Internal error. - requestBody: - required: true - content: - application/json: - schema: - type: object - properties: - token_id: - type: required - example: - token_id: string - description: Revokes an access token. -components: - x-stackQL-resources: - automatic_cluster_update: - id: databricks_workspace.workspace.automatic_cluster_update - name: automatic_cluster_update - title: Automatic_cluster_update - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1automatic_cluster_update~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1automatic_cluster_update~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/automatic_cluster_update/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/automatic_cluster_update/methods/update' - replace: [] - delete: [] - exec: [] - compliance_security_profile: - id: databricks_workspace.workspace.compliance_security_profile - name: compliance_security_profile - title: Compliance_security_profile - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1shield_csp_enablement_ws_db~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1shield_csp_enablement_ws_db~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/compliance_security_profile/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/compliance_security_profile/methods/update' - replace: [] - delete: [] - exec: [] - dashboard_embedding_access_policy: - id: databricks_workspace.workspace.dashboard_embedding_access_policy - name: dashboard_embedding_access_policy - title: Dashboard_embedding_access_policy - methods: - delete: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_acc_policy~1names~1default/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_acc_policy~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_acc_policy~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dashboard_embedding_access_policy/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/dashboard_embedding_access_policy/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dashboard_embedding_access_policy/methods/delete' - exec: [] - dashboard_embedding_approved_domains: - id: databricks_workspace.workspace.dashboard_embedding_approved_domains - name: dashboard_embedding_approved_domains - title: Dashboard_embedding_approved_domains - methods: - delete: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_apprvd_domains~1names~1default/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_apprvd_domains~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1aibi_dash_embed_ws_apprvd_domains~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/dashboard_embedding_approved_domains/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/dashboard_embedding_approved_domains/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/dashboard_embedding_approved_domains/methods/delete' - exec: [] - directories: - id: databricks_workspace.workspace.directories - name: directories - title: Directories - methods: - list: - operation: - $ref: '#/paths/~1api~12.0~1workspace~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.objects - mkdirs: - operation: - $ref: '#/paths/~1api~12.0~1workspace~1mkdirs/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/directories/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/directories/methods/mkdirs' - update: [] - replace: [] - delete: [] - exec: [] - enhanced_security_monitoring: - id: databricks_workspace.workspace.enhanced_security_monitoring - name: enhanced_security_monitoring - title: Enhanced_security_monitoring - methods: - get: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1shield_esm_enablement_ws_db~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1shield_esm_enablement_ws_db~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/enhanced_security_monitoring/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/enhanced_security_monitoring/methods/update' - replace: [] - delete: [] - exec: [] - feature_status: - id: databricks_workspace.workspace.feature_status - name: feature_status - title: Feature_status - methods: - getstatus: - operation: - $ref: '#/paths/~1api~12.0~1workspace-conf/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setstatus: - operation: - $ref: '#/paths/~1api~12.0~1workspace-conf/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/feature_status/methods/getstatus' - insert: [] - update: [] - replace: - - $ref: '#/components/x-stackQL-resources/feature_status/methods/setstatus' - delete: [] - exec: [] - ip_access_lists: - id: databricks_workspace.workspace.ip_access_lists - name: ip_access_lists - title: Ip_access_lists - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1ip-access-lists/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1ip-access-lists/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.ip_access_lists - delete: - operation: - $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.ip_access_list - replace: - operation: - $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1ip-access-lists~1{ip_access_list_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/get' - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/update' - replace: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/replace' - delete: - - $ref: '#/components/x-stackQL-resources/ip_access_lists/methods/delete' - exec: [] - notification_destinations: - id: databricks_workspace.workspace.notification_destinations - name: notification_destinations - title: Notification_destinations - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1notification-destinations/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - list: - operation: - $ref: '#/paths/~1api~12.0~1notification-destinations/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.results - delete: - operation: - $ref: '#/paths/~1api~12.0~1notification-destinations~1{id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1notification-destinations~1{id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1notification-destinations~1{id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/notification_destinations/methods/get' - - $ref: '#/components/x-stackQL-resources/notification_destinations/methods/list' - insert: - - $ref: '#/components/x-stackQL-resources/notification_destinations/methods/create' - update: - - $ref: '#/components/x-stackQL-resources/notification_destinations/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/notification_destinations/methods/delete' - exec: [] - objects: - id: databricks_workspace.workspace.objects - name: objects - title: Objects - methods: - delete: - operation: - $ref: '#/paths/~1api~12.0~1workspace~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - export: - operation: - $ref: '#/paths/~1api~12.0~1workspace~1export/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - import: - operation: - $ref: '#/paths/~1api~12.0~1workspace~1import/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: [] - insert: [] - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/objects/methods/delete' - exec: - - $ref: '#/components/x-stackQL-resources/objects/methods/import' - - $ref: '#/components/x-stackQL-resources/objects/methods/export' - object_permissions: - id: databricks_workspace.workspace.object_permissions - name: object_permissions - title: Object_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/object_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/object_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/object_permissions/methods/setpermissions' - delete: [] - exec: [] - object_permission_levels: - id: databricks_workspace.workspace.object_permission_levels - name: object_permission_levels - title: Object_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/object_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - object_status: - id: databricks_workspace.workspace.object_status - name: object_status - title: Object_status - methods: - getstatus: - operation: - $ref: '#/paths/~1api~12.0~1workspace~1get-status/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/object_status/methods/getstatus' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - permissions: - id: databricks_workspace.workspace.permissions - name: permissions - title: Permissions - methods: - delete: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1default_namespace_ws~1names~1default/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1default_namespace_ws~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1default_namespace_ws~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/permissions/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/permissions/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/permissions/methods/delete' - exec: [] - restrict_admins: - id: databricks_workspace.workspace.restrict_admins - name: restrict_admins - title: Restrict_admins - methods: - delete: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1restrict_workspace_admins~1names~1default/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1restrict_workspace_admins~1names~1default/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - update: - operation: - $ref: '#/paths/~1api~12.0~1settings~1types~1restrict_workspace_admins~1names~1default/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/restrict_admins/methods/get' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/restrict_admins/methods/update' - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/restrict_admins/methods/delete' - exec: [] - tokens: - id: databricks_workspace.workspace.tokens - name: tokens - title: Tokens - methods: - createobotoken: - operation: - $ref: '#/paths/~1api~12.0~1token-management~1on-behalf-of~1tokens/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - delete: - operation: - $ref: '#/paths/~1api~12.0~1token-management~1tokens~1{token_id}/delete' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - get: - operation: - $ref: '#/paths/~1api~12.0~1token-management~1tokens~1{token_id}/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.token_info - list: - operation: - $ref: '#/paths/~1api~12.0~1token-management~1tokens/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.token_infos - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/tokens/methods/list' - - $ref: '#/components/x-stackQL-resources/tokens/methods/get' - insert: - - $ref: '#/components/x-stackQL-resources/tokens/methods/createobotoken' - update: [] - replace: [] - delete: - - $ref: '#/components/x-stackQL-resources/tokens/methods/delete' - exec: [] - token_permissions: - id: databricks_workspace.workspace.token_permissions - name: token_permissions - title: Token_permissions - methods: - getpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1tokens/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - setpermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1tokens/put' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - updatepermissions: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1tokens/patch' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/token_permissions/methods/getpermissions' - insert: [] - update: - - $ref: '#/components/x-stackQL-resources/token_permissions/methods/updatepermissions' - replace: - - $ref: '#/components/x-stackQL-resources/token_permissions/methods/setpermissions' - delete: [] - exec: [] - token_permission_levels: - id: databricks_workspace.workspace.token_permission_levels - name: token_permission_levels - title: Token_permission_levels - methods: - getpermissionlevels: - operation: - $ref: '#/paths/~1api~12.0~1permissions~1authorization~1tokens~1permissionLevels/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.permission_levels - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/token_permission_levels/methods/getpermissionlevels' - insert: [] - update: [] - replace: [] - delete: [] - exec: [] - user_tokens: - id: databricks_workspace.workspace.user_tokens - name: user_tokens - title: User_tokens - methods: - create: - operation: - $ref: '#/paths/~1api~12.0~1token~1create/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - listtokens: - operation: - $ref: '#/paths/~1api~12.0~1token~1list/get' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - objectKey: $.token_infos - revoketoken: - operation: - $ref: '#/paths/~1api~12.0~1token~1delete/post' - response: - mediaType: application/json - overrideMediaType: application/json - openAPIDocKey: '200' - sqlVerbs: - select: - - $ref: '#/components/x-stackQL-resources/user_tokens/methods/listtokens' - insert: - - $ref: '#/components/x-stackQL-resources/user_tokens/methods/create' - update: [] - replace: [] - delete: [] - exec: - - $ref: '#/components/x-stackQL-resources/user_tokens/methods/revoketoken' +openapi: 3.0.0 +info: + title: Databricks Workspace API (workspace) + description: >- + OpenAPI specification for the Databricks workspace service (workspace-level + APIs), generated from the Databricks Python SDK. + version: 0.1.0 + x-stackql-sdk-version: 0.86.0 + x-stackql-date-generated: '2026-02-12' + x-stackql-sdk-namespace: databricks.sdk.service.workspace +servers: + - url: https://{deployment_name}.cloud.databricks.com + variables: + deployment_name: + description: The Databricks Workspace Deployment Name + default: dbc-abcd0123-a1bc +paths: + /api/2.0/git-credentials: + post: + operationId: git_credentials_create + summary: >- + Creates a Git credential entry for the user. Only one Git credential per + user is supported, so any + tags: + - workspace + - git_credentials + description: >- + Creates a Git credential entry for the user. Only one Git credential per + user is supported, so any + + attempts to create credentials if an entry already exists will fail. Use + the PATCH endpoint to update + + existing credentials, or the DELETE endpoint to delete existing + credentials. + + + :param git_provider: str + Git provider. This field is case-insensitive. The available Git providers are `gitHub`, + `bitbucketCloud`, `gitLab`, `azureDevOpsServices`, `gitHubEnterprise`, `bitbucketServer`, + `gitLabEnterpriseEdition` and `awsCodeCommit`. + :param git_email: str (optional) + The authenticating email associated with your Git provider user account. Used for authentication + with the remote repository and also sets the author & committer identity for commits. Required for + most Git providers except AWS CodeCommit. Learn more at + https://docs.databricks.com/aws/en/repos/get-access-tokens-from-git-provider + :param git_username: str (optional) + The username provided with your Git provider account and associated with the credential. For most + Git providers it is only used to set the Git committer & author names for commits, however it may be + required for authentication depending on your Git provider / token requirements. Required for AWS + CodeCommit. + :param is_default_for_provider: bool (optional) + if the credential is the default for the given provider + :param name: str (optional) + the name of the git credential, used for identification and ease of lookup + :param personal_access_token: str (optional) + The personal access token used to authenticate to the corresponding Git provider. For certain + providers, support may exist for other types of scoped access tokens. [Learn more]. + + [Learn more]: https://docs.databricks.com/repos/get-access-tokens-from-git-provider.html + :param principal_id: int (optional) + The ID of the service principal whose credentials will be modified. Only service principal managers + can perform this action. + + :returns: :class:`CreateCredentialsResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + git_provider: + type: string + description: >- + Git provider. This field is case-insensitive. The available + Git providers are `gitHub`, `bitbucketCloud`, `gitLab`, + `azureDevOpsServices`, `gitHubEnterprise`, + `bitbucketServer`, `gitLabEnterpriseEdition` and + `awsCodeCommit`. + git_email: + type: string + description: >- + The authenticating email associated with your Git provider + user account. Used for authentication with the remote + repository and also sets the author & committer identity for + commits. Required for most Git providers except AWS + CodeCommit. Learn more at + https://docs.databricks.com/aws/en/repos/get-access-tokens-from-git-provider + git_username: + type: string + description: >- + The username provided with your Git provider account and + associated with the credential. For most Git providers it is + only used to set the Git committer & author names for + commits, however it may be required for authentication + depending on your Git provider / token requirements. + Required for AWS CodeCommit. + is_default_for_provider: + type: string + description: if the credential is the default for the given provider + name: + type: string + description: >- + the name of the git credential, used for identification and + ease of lookup + personal_access_token: + type: string + description: >- + The personal access token used to authenticate to the + corresponding Git provider. For certain providers, support + may exist for other types of scoped access tokens. [Learn + more]. [Learn more]: + https://docs.databricks.com/repos/get-access-tokens-from-git-provider.html + principal_id: + type: string + description: >- + The ID of the service principal whose credentials will be + modified. Only service principal managers can perform this + action. + required: + - git_provider + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateCredentialsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GitCredentialsAPI + get: + operationId: git_credentials_list + summary: Lists the calling user's Git credentials. + tags: + - workspace + - git_credentials + description: |- + Lists the calling user's Git credentials. + + :param principal_id: int (optional) + The ID of the service principal whose credentials will be listed. Only service principal managers + can perform this action. + + :returns: Iterator over :class:`CredentialInfo` + parameters: + - name: principal_id + in: query + required: false + schema: + type: string + description: >- + The ID of the service principal whose credentials will be listed. + Only service principal managers can perform this action. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListCredentialsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GitCredentialsAPI + /api/2.0/git-credentials/{credential_id}: + delete: + operationId: git_credentials_delete + summary: Deletes the specified Git credential. + tags: + - workspace + - git_credentials + description: |- + Deletes the specified Git credential. + + :param credential_id: int + The ID for the corresponding credential to access. + :param principal_id: int (optional) + The ID of the service principal whose credentials will be modified. Only service principal managers + can perform this action. + parameters: + - name: credential_id + in: path + required: true + schema: + type: integer + description: The ID for the corresponding credential to access. + - name: principal_id + in: query + required: false + schema: + type: string + description: >- + The ID of the service principal whose credentials will be modified. + Only service principal managers can perform this action. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GitCredentialsAPI + get: + operationId: git_credentials_get + summary: Gets the Git credential with the specified credential ID. + tags: + - workspace + - git_credentials + description: |- + Gets the Git credential with the specified credential ID. + + :param credential_id: int + The ID for the corresponding credential to access. + :param principal_id: int (optional) + The ID of the service principal whose credentials will be modified. Only service principal managers + can perform this action. + + :returns: :class:`GetCredentialsResponse` + parameters: + - name: credential_id + in: path + required: true + schema: + type: integer + description: The ID for the corresponding credential to access. + - name: principal_id + in: query + required: false + schema: + type: string + description: >- + The ID of the service principal whose credentials will be modified. + Only service principal managers can perform this action. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetCredentialsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GitCredentialsAPI + patch: + operationId: git_credentials_update + summary: Updates the specified Git credential. + tags: + - workspace + - git_credentials + description: |- + Updates the specified Git credential. + + :param credential_id: int + The ID for the corresponding credential to access. + :param git_provider: str + Git provider. This field is case-insensitive. The available Git providers are `gitHub`, + `bitbucketCloud`, `gitLab`, `azureDevOpsServices`, `gitHubEnterprise`, `bitbucketServer`, + `gitLabEnterpriseEdition` and `awsCodeCommit`. + :param git_email: str (optional) + The authenticating email associated with your Git provider user account. Used for authentication + with the remote repository and also sets the author & committer identity for commits. Required for + most Git providers except AWS CodeCommit. Learn more at + https://docs.databricks.com/aws/en/repos/get-access-tokens-from-git-provider + :param git_username: str (optional) + The username provided with your Git provider account and associated with the credential. For most + Git providers it is only used to set the Git committer & author names for commits, however it may be + required for authentication depending on your Git provider / token requirements. Required for AWS + CodeCommit. + :param is_default_for_provider: bool (optional) + if the credential is the default for the given provider + :param name: str (optional) + the name of the git credential, used for identification and ease of lookup + :param personal_access_token: str (optional) + The personal access token used to authenticate to the corresponding Git provider. For certain + providers, support may exist for other types of scoped access tokens. [Learn more]. + + [Learn more]: https://docs.databricks.com/repos/get-access-tokens-from-git-provider.html + :param principal_id: int (optional) + The ID of the service principal whose credentials will be modified. Only service principal managers + can perform this action. + parameters: + - name: credential_id + in: path + required: true + schema: + type: integer + description: The ID for the corresponding credential to access. + requestBody: + content: + application/json: + schema: + type: object + properties: + git_provider: + type: string + description: >- + Git provider. This field is case-insensitive. The available + Git providers are `gitHub`, `bitbucketCloud`, `gitLab`, + `azureDevOpsServices`, `gitHubEnterprise`, + `bitbucketServer`, `gitLabEnterpriseEdition` and + `awsCodeCommit`. + git_email: + type: string + description: >- + The authenticating email associated with your Git provider + user account. Used for authentication with the remote + repository and also sets the author & committer identity for + commits. Required for most Git providers except AWS + CodeCommit. Learn more at + https://docs.databricks.com/aws/en/repos/get-access-tokens-from-git-provider + git_username: + type: string + description: >- + The username provided with your Git provider account and + associated with the credential. For most Git providers it is + only used to set the Git committer & author names for + commits, however it may be required for authentication + depending on your Git provider / token requirements. + Required for AWS CodeCommit. + is_default_for_provider: + type: string + description: if the credential is the default for the given provider + name: + type: string + description: >- + the name of the git credential, used for identification and + ease of lookup + personal_access_token: + type: string + description: >- + The personal access token used to authenticate to the + corresponding Git provider. For certain providers, support + may exist for other types of scoped access tokens. [Learn + more]. [Learn more]: + https://docs.databricks.com/repos/get-access-tokens-from-git-provider.html + principal_id: + type: string + description: >- + The ID of the service principal whose credentials will be + modified. Only service principal managers can perform this + action. + required: + - git_provider + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: GitCredentialsAPI + /api/2.0/repos: + post: + operationId: repos_create + summary: >- + Creates a repo in the workspace and links it to the remote Git repo + specified. Note that repos created + tags: + - workspace + - repos + description: >- + Creates a repo in the workspace and links it to the remote Git repo + specified. Note that repos created + + programmatically must be linked to a remote Git repo, unlike repos + created in the browser. + + + :param url: str + URL of the Git repository to be linked. + :param provider: str + Git provider. This field is case-insensitive. The available Git providers are `gitHub`, + `bitbucketCloud`, `gitLab`, `azureDevOpsServices`, `gitHubEnterprise`, `bitbucketServer`, + `gitLabEnterpriseEdition` and `awsCodeCommit`. + :param path: str (optional) + Desired path for the repo in the workspace. Almost any path in the workspace can be chosen. If repo + is created in `/Repos`, path must be in the format `/Repos/{folder}/{repo-name}`. + :param sparse_checkout: :class:`SparseCheckout` (optional) + If specified, the repo will be created with sparse checkout enabled. You cannot enable/disable + sparse checkout after the repo is created. + + :returns: :class:`CreateRepoResponse` + requestBody: + content: + application/json: + schema: + type: object + properties: + url: + type: string + description: URL of the Git repository to be linked. + provider: + type: string + description: >- + Git provider. This field is case-insensitive. The available + Git providers are `gitHub`, `bitbucketCloud`, `gitLab`, + `azureDevOpsServices`, `gitHubEnterprise`, + `bitbucketServer`, `gitLabEnterpriseEdition` and + `awsCodeCommit`. + path: + type: string + description: >- + Desired path for the repo in the workspace. Almost any path + in the workspace can be chosen. If repo is created in + `/Repos`, path must be in the format + `/Repos/{folder}/{repo-name}`. + sparse_checkout: + type: string + description: >- + If specified, the repo will be created with sparse checkout + enabled. You cannot enable/disable sparse checkout after the + repo is created. + required: + - url + - provider + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/CreateRepoResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + get: + operationId: repos_list + summary: >- + Returns repos that the calling user has Manage permissions on. Use + `next_page_token` to iterate + tags: + - workspace + - repos + description: >- + Returns repos that the calling user has Manage permissions on. Use + `next_page_token` to iterate + + through additional pages. + + + :param next_page_token: str (optional) + Token used to get the next page of results. If not specified, returns the first page of results as + well as a next page token if there are more results. + :param path_prefix: str (optional) + Filters repos that have paths starting with the given path prefix. If not provided or when provided + an effectively empty prefix (`/` or `/Workspace`) Git folders (repos) from `/Workspace/Repos` will + be served. + + :returns: Iterator over :class:`RepoInfo` + parameters: + - name: next_page_token + in: query + required: false + schema: + type: string + description: >- + Token used to get the next page of results. If not specified, + returns the first page of results as well as a next page token if + there are more results. + - name: path_prefix + in: query + required: false + schema: + type: string + description: >- + Filters repos that have paths starting with the given path prefix. + If not provided or when provided an effectively empty prefix (`/` or + `/Workspace`) Git folders (repos) from `/Workspace/Repos` will be + served. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListReposResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + /api/2.0/repos/{repo_id}: + delete: + operationId: repos_delete + summary: Deletes the specified repo. + tags: + - workspace + - repos + description: |- + Deletes the specified repo. + + :param repo_id: int + The ID for the corresponding repo to delete. + parameters: + - name: repo_id + in: path + required: true + schema: + type: integer + description: The ID for the corresponding repo to delete. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + get: + operationId: repos_get + summary: Returns the repo with the given repo ID. + tags: + - workspace + - repos + description: |- + Returns the repo with the given repo ID. + + :param repo_id: int + ID of the Git folder (repo) object in the workspace. + + :returns: :class:`GetRepoResponse` + parameters: + - name: repo_id + in: path + required: true + schema: + type: integer + description: ID of the Git folder (repo) object in the workspace. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetRepoResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + patch: + operationId: repos_update + summary: >- + Updates the repo to a different branch or tag, or updates the repo to + the latest commit on the same + tags: + - workspace + - repos + description: >- + Updates the repo to a different branch or tag, or updates the repo to + the latest commit on the same + + branch. + + + :param repo_id: int + ID of the Git folder (repo) object in the workspace. + :param branch: str (optional) + Branch that the local version of the repo is checked out to. + :param sparse_checkout: :class:`SparseCheckoutUpdate` (optional) + If specified, update the sparse checkout settings. The update will fail if sparse checkout is not + enabled for the repo. + :param tag: str (optional) + Tag that the local version of the repo is checked out to. Updating the repo to a tag puts the repo + in a detached HEAD state. Before committing new changes, you must update the repo to a branch + instead of the detached HEAD. + parameters: + - name: repo_id + in: path + required: true + schema: + type: integer + description: ID of the Git folder (repo) object in the workspace. + requestBody: + content: + application/json: + schema: + type: object + properties: + branch: + type: string + description: Branch that the local version of the repo is checked out to. + sparse_checkout: + type: string + description: >- + If specified, update the sparse checkout settings. The + update will fail if sparse checkout is not enabled for the + repo. + tag: + type: string + description: >- + Tag that the local version of the repo is checked out to. + Updating the repo to a tag puts the repo in a detached HEAD + state. Before committing new changes, you must update the + repo to a branch instead of the detached HEAD. + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + /api/2.0/permissions/repos/{repo_id}/permissionLevels: + get: + operationId: repos_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - workspace + - repos + description: |- + Gets the permission levels that a user can have on an object. + + :param repo_id: str + The repo for which to get or manage permissions. + + :returns: :class:`GetRepoPermissionLevelsResponse` + parameters: + - name: repo_id + in: path + required: true + schema: + type: string + description: The repo for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetRepoPermissionLevelsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + /api/2.0/permissions/repos/{repo_id}: + get: + operationId: repos_get_permissions + summary: >- + Gets the permissions of a repo. Repos can inherit permissions from their + root object. + tags: + - workspace + - repos + description: >- + Gets the permissions of a repo. Repos can inherit permissions from their + root object. + + + :param repo_id: str + The repo for which to get or manage permissions. + + :returns: :class:`RepoPermissions` + parameters: + - name: repo_id + in: path + required: true + schema: + type: string + description: The repo for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RepoPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + put: + operationId: repos_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - workspace + - repos + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their root object. + + + :param repo_id: str + The repo for which to get or manage permissions. + :param access_control_list: List[:class:`RepoAccessControlRequest`] + (optional) + + + :returns: :class:`RepoPermissions` + parameters: + - name: repo_id + in: path + required: true + schema: + type: string + description: The repo for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`RepoPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RepoPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + patch: + operationId: repos_update_permissions + summary: >- + Updates the permissions on a repo. Repos can inherit permissions from + their root object. + tags: + - workspace + - repos + description: >- + Updates the permissions on a repo. Repos can inherit permissions from + their root object. + + + :param repo_id: str + The repo for which to get or manage permissions. + :param access_control_list: List[:class:`RepoAccessControlRequest`] + (optional) + + + :returns: :class:`RepoPermissions` + parameters: + - name: repo_id + in: path + required: true + schema: + type: string + description: The repo for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`RepoPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/RepoPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: ReposAPI + /api/2.0/secrets/scopes/create: + post: + operationId: secrets_create_scope + summary: Creates a new secret scope. + tags: + - workspace + - secrets + description: >- + Creates a new secret scope. + + + The scope name must consist of alphanumeric characters, dashes, + underscores, and periods, and may not + + exceed 128 characters. + + + Example request: + + + .. code:: + + + { "scope": "my-simple-databricks-scope", "initial_manage_principal": + "users" "scope_backend_type": + + "databricks|azure_keyvault", # below is only required if scope type is + azure_keyvault + + "backend_azure_keyvault": { "resource_id": + + "/subscriptions/xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx/resourceGroups/xxxx/providers/Microsoft.KeyVault/vaults/xxxx", + + "tenant_id": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx", "dns_name": + "https://xxxx.vault.azure.net/", } } + + + If ``initial_manage_principal`` is specified, the initial ACL applied to + the scope is applied to the + + supplied principal (user or group) with ``MANAGE`` permissions. The only + supported principal for this + + option is the group ``users``, which contains all users in the + workspace. If + + ``initial_manage_principal`` is not specified, the initial ACL with + ``MANAGE`` permission applied to + + the scope is assigned to the API request issuer's user identity. + + + If ``scope_backend_type`` is ``azure_keyvault``, a secret scope is + created with secrets from a given + + Azure KeyVault. The caller must provide the keyvault_resource_id and the + tenant_id for the key vault. + + If ``scope_backend_type`` is ``databricks`` or is unspecified, an empty + secret scope is created and + + stored in Databricks's own storage. + + + Throws ``RESOURCE_ALREADY_EXISTS`` if a scope with the given name + already exists. Throws + + ``RESOURCE_LIMIT_EXCEEDED`` if maximum number of scopes in the workspace + is exceeded. Throws + + ``INVALID_PARAMETER_VALUE`` if the scope name is invalid. Throws + ``BAD_REQUEST`` if request violated + + constraints. Throws ``CUSTOMER_UNAUTHORIZED`` if normal user attempts to + create a scope with name + + reserved for databricks internal usage. Throws ``UNAUTHENTICATED`` if + unable to verify user access + + permission on Azure KeyVault + + + :param scope: str + Scope name requested by the user. Scope names are unique. + :param backend_azure_keyvault: :class:`AzureKeyVaultSecretScopeMetadata` + (optional) + The metadata for the secret scope if the type is ``AZURE_KEYVAULT`` + :param initial_manage_principal: str (optional) + The principal that is initially granted ``MANAGE`` permission to the created scope. + :param scope_backend_type: :class:`ScopeBackendType` (optional) + The backend type the scope will be created with. If not specified, will default to ``DATABRICKS`` + requestBody: + content: + application/json: + schema: + type: object + properties: + scope: + type: string + description: Scope name requested by the user. Scope names are unique. + backend_azure_keyvault: + type: string + description: >- + The metadata for the secret scope if the type is + ``AZURE_KEYVAULT`` + initial_manage_principal: + type: string + description: >- + The principal that is initially granted ``MANAGE`` + permission to the created scope. + scope_backend_type: + type: string + description: >- + The backend type the scope will be created with. If not + specified, will default to ``DATABRICKS`` + required: + - scope + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/acls/delete: + post: + operationId: secrets_delete_acl + summary: Deletes the given ACL on the given scope. + tags: + - workspace + - secrets + description: >- + Deletes the given ACL on the given scope. + + + Users must have the ``MANAGE`` permission to invoke this API. + + + Example request: + + + .. code:: + + + { "scope": "my-secret-scope", "principal": "data-scientists" } + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope, principal, + or ACL exists. Throws + + ``PERMISSION_DENIED`` if the user does not have permission to make this + API call. Throws + + ``INVALID_PARAMETER_VALUE`` if the permission or principal is invalid. + + + :param scope: str + The name of the scope to remove permissions from. + :param principal: str + The principal to remove an existing ACL from. + requestBody: + content: + application/json: + schema: + type: object + properties: + scope: + type: string + description: The name of the scope to remove permissions from. + principal: + type: string + description: The principal to remove an existing ACL from. + required: + - scope + - principal + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/scopes/delete: + post: + operationId: secrets_delete_scope + summary: Deletes a secret scope. + tags: + - workspace + - secrets + description: >- + Deletes a secret scope. + + + Example request: + + + .. code:: + + + { "scope": "my-secret-scope" } + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if the scope does not exist. Throws + ``PERMISSION_DENIED`` if the + + user does not have permission to make this API call. Throws + ``BAD_REQUEST`` if system user attempts to + + delete internal secret scope. + + + :param scope: str + Name of the scope to delete. + requestBody: + content: + application/json: + schema: + type: object + properties: + scope: + type: string + description: Name of the scope to delete. + required: + - scope + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/delete: + post: + operationId: secrets_delete_secret + summary: >- + Deletes the secret stored in this secret scope. You must have ``WRITE`` + or ``MANAGE`` permission on + tags: + - workspace + - secrets + description: >- + Deletes the secret stored in this secret scope. You must have ``WRITE`` + or ``MANAGE`` permission on + + the Secret Scope. + + + Example request: + + + .. code:: + + + { "scope": "my-secret-scope", "key": "my-secret-key" } + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope or secret + exists. Throws + + ``PERMISSION_DENIED`` if the user does not have permission to make this + API call. Throws + + ``BAD_REQUEST`` if system user attempts to delete an internal secret, or + request is made against Azure + + KeyVault backed scope. + + + :param scope: str + The name of the scope that contains the secret to delete. + :param key: str + Name of the secret to delete. + requestBody: + content: + application/json: + schema: + type: object + properties: + scope: + type: string + description: The name of the scope that contains the secret to delete. + key: + type: string + description: Name of the secret to delete. + required: + - scope + - key + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/acls/get: + get: + operationId: secrets_get_acl + summary: >- + Describes the details about the given ACL, such as the group and + permission. + tags: + - workspace + - secrets + description: >- + Describes the details about the given ACL, such as the group and + permission. + + + Users must have the ``MANAGE`` permission to invoke this API. + + + Example response: + + + .. code:: + + + { "principal": "data-scientists", "permission": "READ" } + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope exists. + Throws ``PERMISSION_DENIED`` if the + + user does not have permission to make this API call. Throws + ``INVALID_PARAMETER_VALUE`` if the + + permission or principal is invalid. + + + :param scope: str + The name of the scope to fetch ACL information from. + :param principal: str + The principal to fetch ACL information for. + + :returns: :class:`AclItem` + parameters: + - name: scope + in: query + required: true + schema: + type: string + description: The name of the scope to fetch ACL information from. + - name: principal + in: query + required: true + schema: + type: string + description: The principal to fetch ACL information for. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/AclItem' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/get: + get: + operationId: secrets_get_secret + summary: >- + Gets a secret for a given key and scope. This API can only be called + from the DBUtils interface. Users + tags: + - workspace + - secrets + description: >- + Gets a secret for a given key and scope. This API can only be called + from the DBUtils interface. Users + + need the READ permission to make this call. + + + Example response: + + + .. code:: + + + { "key": "my-string-key", "value": } + + + Note that the secret value returned is in bytes. The interpretation of + the bytes is determined by the + + caller in DBUtils and the type the data is decoded into. + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret or secret scope + exists. Throws + + ``PERMISSION_DENIED`` if the user does not have permission to make this + API call. + + + Note: This is explicitly an undocumented API. It also doesn't need to be + supported for the /preview + + prefix, because it's not a customer-facing API (i.e. only used for + DBUtils SecretUtils to fetch + + secrets). + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope or secret + exists. Throws ``BAD_REQUEST`` if + + normal user calls get secret outside of a notebook. AKV specific errors: + Throws + + ``INVALID_PARAMETER_VALUE`` if secret name is not alphanumeric or too + long. Throws + + ``PERMISSION_DENIED`` if secret manager cannot access AKV with 403 error + Throws ``MALFORMED_REQUEST`` + + if secret manager cannot access AKV with any other 4xx error + + + :param scope: str + The name of the scope that contains the secret. + :param key: str + Name of the secret to fetch value information. + + :returns: :class:`GetSecretResponse` + parameters: + - name: scope + in: query + required: true + schema: + type: string + description: The name of the scope that contains the secret. + - name: key + in: query + required: true + schema: + type: string + description: Name of the secret to fetch value information. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/GetSecretResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/acls/list: + get: + operationId: secrets_list_acls + summary: Lists the ACLs set on the given scope. + tags: + - workspace + - secrets + description: >- + Lists the ACLs set on the given scope. + + + Users must have the ``MANAGE`` permission to invoke this API. + + + Example response: + + + .. code:: + + + { "acls": [{ "principal": "admins", "permission": "MANAGE" },{ + "principal": "data-scientists", + + "permission": "READ" }] } + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope exists. + Throws ``PERMISSION_DENIED`` if the + + user does not have permission to make this API call. + + + :param scope: str + The name of the scope to fetch ACL information from. + + :returns: Iterator over :class:`AclItem` + parameters: + - name: scope + in: query + required: true + schema: + type: string + description: The name of the scope to fetch ACL information from. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListAclsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/scopes/list: + get: + operationId: secrets_list_scopes + summary: Lists all secret scopes available in the workspace. + tags: + - workspace + - secrets + description: >- + Lists all secret scopes available in the workspace. + + + Example response: + + + .. code:: + + + { "scopes": [{ "name": "my-databricks-scope", "backend_type": + "DATABRICKS" },{ "name": "mount-points", + + "backend_type": "DATABRICKS" }] } + + + Throws ``PERMISSION_DENIED`` if the user does not have permission to + make this API call. + + + + :returns: Iterator over :class:`SecretScope` + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListScopesResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/list: + get: + operationId: secrets_list_secrets + summary: >- + Lists the secret keys that are stored at this scope. This is a + metadata-only operation; secret data + tags: + - workspace + - secrets + description: >- + Lists the secret keys that are stored at this scope. This is a + metadata-only operation; secret data + + cannot be retrieved using this API. Users need the READ permission to + make this call. + + + Example response: + + + .. code:: + + + { "secrets": [ { "key": "my-string-key"", "last_updated_timestamp": + "1520467595000" }, { "key": + + "my-byte-key", "last_updated_timestamp": "1520467595000" }, ] } + + + The lastUpdatedTimestamp returned is in milliseconds since epoch. + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope exists. + Throws ``PERMISSION_DENIED`` if the + + user does not have permission to make this API call. + + + :param scope: str + The name of the scope to list secrets within. + + :returns: Iterator over :class:`SecretMetadata` + parameters: + - name: scope + in: query + required: true + schema: + type: string + description: The name of the scope to list secrets within. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListSecretsResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/acls/put: + post: + operationId: secrets_put_acl + summary: >- + Creates or overwrites the ACL associated with the given principal (user + or group) on the specified + tags: + - workspace + - secrets + description: >- + Creates or overwrites the ACL associated with the given principal (user + or group) on the specified + + scope point. In general, a user or group will use the most powerful + permission available to them, and + + permissions are ordered as follows: + + + * ``MANAGE`` - Allowed to change ACLs, and read and write to this secret + scope. * ``WRITE`` - Allowed + + to read and write to this secret scope. * ``READ`` - Allowed to read + this secret scope and list what + + secrets are available. + + + Note that in general, secret values can only be read from within a + command on a cluster (for example, + + through a notebook). There is no API to read the actual secret value + material outside of a cluster. + + However, the user's permission will be applied based on who is executing + the command, and they must + + have at least READ permission. + + + Users must have the ``MANAGE`` permission to invoke this API. + + + Example request: + + + .. code:: + + + { "scope": "my-secret-scope", "principal": "data-scientists", + "permission": "READ" } + + + The principal is a user or group name corresponding to an existing + Databricks principal to be granted + + or revoked access. + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope exists. + Throws ``RESOURCE_ALREADY_EXISTS`` + + if a permission for the principal already exists. Throws + ``INVALID_PARAMETER_VALUE`` if the permission + + or principal is invalid. Throws ``PERMISSION_DENIED`` if the user does + not have permission to make + + this API call. + + + :param scope: str + The name of the scope to apply permissions to. + :param principal: str + The principal in which the permission is applied. + :param permission: :class:`AclPermission` + The permission level applied to the principal. + requestBody: + content: + application/json: + schema: + type: object + properties: + scope: + type: string + description: The name of the scope to apply permissions to. + principal: + type: string + description: The principal in which the permission is applied. + permission: + type: string + description: The permission level applied to the principal. + required: + - scope + - principal + - permission + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/secrets/put: + post: + operationId: secrets_put_secret + summary: >- + Inserts a secret under the provided scope with the given name. If a + secret already exists with the + tags: + - workspace + - secrets + description: >- + Inserts a secret under the provided scope with the given name. If a + secret already exists with the + + same name, this command overwrites the existing secret's value. The + server encrypts the secret using + + the secret scope's encryption settings before storing it. You must have + ``WRITE`` or ``MANAGE`` + + permission on the secret scope. + + + The secret key must consist of alphanumeric characters, dashes, + underscores, and periods, and cannot + + exceed 128 characters. The maximum allowed secret value size is 128 KB. + The maximum number of secrets + + in a given scope is 1000. + + + Example request: + + + .. code:: + + + { "scope": "my-databricks-scope", "key": "my-string-key", + "string_value": "foobar" } + + + The input fields "string_value" or "bytes_value" specify the type of the + secret, which will determine + + the value returned when the secret value is requested. Exactly one must + be specified. + + + Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret scope exists. + Throws ``RESOURCE_LIMIT_EXCEEDED`` + + if maximum number of secrets in scope is exceeded. Throws + ``INVALID_PARAMETER_VALUE`` if the request + + parameters are invalid. Throws ``PERMISSION_DENIED`` if the user does + not have permission to make this + + API call. Throws ``MALFORMED_REQUEST`` if request is incorrectly + formatted or conflicting. Throws + + ``BAD_REQUEST`` if request is made against Azure KeyVault backed scope. + + + :param scope: str + The name of the scope to which the secret will be associated with. + :param key: str + A unique name to identify the secret. + :param bytes_value: str (optional) + If specified, value will be stored as bytes. + :param string_value: str (optional) + If specified, note that the value will be stored in UTF-8 (MB4) form. + requestBody: + content: + application/json: + schema: + type: object + properties: + scope: + type: string + description: >- + The name of the scope to which the secret will be associated + with. + key: + type: string + description: A unique name to identify the secret. + bytes_value: + type: string + description: If specified, value will be stored as bytes. + string_value: + type: string + description: >- + If specified, note that the value will be stored in UTF-8 + (MB4) form. + required: + - scope + - key + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: SecretsAPI + /api/2.0/workspace/delete: + post: + operationId: workspace_delete + summary: >- + Deletes an object or a directory (and optionally recursively deletes all + objects in the directory). * + tags: + - workspace + description: >- + Deletes an object or a directory (and optionally recursively deletes all + objects in the directory). * + + If `path` does not exist, this call returns an error + `RESOURCE_DOES_NOT_EXIST`. * If `path` is a + + non-empty directory and `recursive` is set to `false`, this call returns + an error + + `DIRECTORY_NOT_EMPTY`. + + + Object deletion cannot be undone and deleting a directory recursively is + not atomic. + + + :param path: str + The absolute path of the notebook or directory. + :param recursive: bool (optional) + The flag that specifies whether to delete the object recursively. It is `false` by default. Please + note this deleting directory is not atomic. If it fails in the middle, some of objects under this + directory may be deleted and cannot be undone. + requestBody: + content: + application/json: + schema: + type: object + properties: + path: + type: string + description: The absolute path of the notebook or directory. + recursive: + type: string + description: >- + The flag that specifies whether to delete the object + recursively. It is `false` by default. Please note this + deleting directory is not atomic. If it fails in the middle, + some of objects under this directory may be deleted and + cannot be undone. + required: + - path + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + /api/2.0/workspace/export: + get: + operationId: workspace_export + summary: Exports an object or the contents of an entire directory. + tags: + - workspace + description: >- + Exports an object or the contents of an entire directory. + + + If `path` does not exist, this call returns an error + `RESOURCE_DOES_NOT_EXIST`. + + + If the exported data would exceed size limit, this call returns + `MAX_NOTEBOOK_SIZE_EXCEEDED`. + + Currently, this API does not support exporting a library. + + + :param path: str + The absolute path of the object or directory. Exporting a directory is only supported for the `DBC`, + `SOURCE`, and `AUTO` format. + :param format: :class:`ExportFormat` (optional) + This specifies the format of the exported file. By default, this is `SOURCE`. + + The value is case sensitive. + + - `SOURCE`: The notebook is exported as source code. Directory exports will not include non-notebook + entries. - `HTML`: The notebook is exported as an HTML file. - `JUPYTER`: The notebook is exported + as a Jupyter/IPython Notebook file. - `DBC`: The notebook is exported in Databricks archive format. + Directory exports will not include non-notebook entries. - `R_MARKDOWN`: The notebook is exported to + R Markdown format. - `AUTO`: The object or directory is exported depending on the objects type. + Directory exports will include notebooks and workspace files. + + :returns: :class:`ExportResponse` + parameters: + - name: path + in: query + required: true + schema: + type: string + description: >- + The absolute path of the object or directory. Exporting a directory + is only supported for the `DBC`, `SOURCE`, and `AUTO` format. + - name: format + in: query + required: false + schema: + type: string + description: >- + This specifies the format of the exported file. By default, this is + `SOURCE`. The value is case sensitive. - `SOURCE`: The notebook is + exported as source code. Directory exports will not include + non-notebook entries. - `HTML`: The notebook is exported as an HTML + file. - `JUPYTER`: The notebook is exported as a Jupyter/IPython + Notebook file. - `DBC`: The notebook is exported in Databricks + archive format. Directory exports will not include non-notebook + entries. - `R_MARKDOWN`: The notebook is exported to R Markdown + format. - `AUTO`: The object or directory is exported depending on + the objects type. Directory exports will include notebooks and + workspace files. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ExportResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + /api/2.0/permissions/{workspace_object_type}/{workspace_object_id}/permissionLevels: + get: + operationId: workspace_get_permission_levels + summary: Gets the permission levels that a user can have on an object. + tags: + - workspace + description: |- + Gets the permission levels that a user can have on an object. + + :param workspace_object_type: str + The workspace object type for which to get or manage permissions. Could be one of the following: + alerts, alertsv2, dashboards, dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + :param workspace_object_id: str + The workspace object for which to get or manage permissions. + + :returns: :class:`GetWorkspaceObjectPermissionLevelsResponse` + parameters: + - name: workspace_object_type + in: path + required: true + schema: + type: string + description: >- + The workspace object type for which to get or manage permissions. + Could be one of the following: alerts, alertsv2, dashboards, + dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + - name: workspace_object_id + in: path + required: true + schema: + type: string + description: The workspace object for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: >- + #/components/schemas/GetWorkspaceObjectPermissionLevelsResponse + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + /api/2.0/permissions/{workspace_object_type}/{workspace_object_id}: + get: + operationId: workspace_get_permissions + summary: >- + Gets the permissions of a workspace object. Workspace objects can + inherit permissions from their + tags: + - workspace + description: >- + Gets the permissions of a workspace object. Workspace objects can + inherit permissions from their + + parent objects or root object. + + + :param workspace_object_type: str + The workspace object type for which to get or manage permissions. Could be one of the following: + alerts, alertsv2, dashboards, dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + :param workspace_object_id: str + The workspace object for which to get or manage permissions. + + :returns: :class:`WorkspaceObjectPermissions` + parameters: + - name: workspace_object_type + in: path + required: true + schema: + type: string + description: >- + The workspace object type for which to get or manage permissions. + Could be one of the following: alerts, alertsv2, dashboards, + dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + - name: workspace_object_id + in: path + required: true + schema: + type: string + description: The workspace object for which to get or manage permissions. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspaceObjectPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + put: + operationId: workspace_set_permissions + summary: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + tags: + - workspace + description: >- + Sets permissions on an object, replacing existing permissions if they + exist. Deletes all direct + + permissions if none are specified. Objects can inherit permissions from + their parent objects or root + + object. + + + :param workspace_object_type: str + The workspace object type for which to get or manage permissions. Could be one of the following: + alerts, alertsv2, dashboards, dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + :param workspace_object_id: str + The workspace object for which to get or manage permissions. + :param access_control_list: + List[:class:`WorkspaceObjectAccessControlRequest`] (optional) + + + :returns: :class:`WorkspaceObjectPermissions` + parameters: + - name: workspace_object_type + in: path + required: true + schema: + type: string + description: >- + The workspace object type for which to get or manage permissions. + Could be one of the following: alerts, alertsv2, dashboards, + dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + - name: workspace_object_id + in: path + required: true + schema: + type: string + description: The workspace object for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`WorkspaceObjectPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspaceObjectPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + patch: + operationId: workspace_update_permissions + summary: >- + Updates the permissions on a workspace object. Workspace objects can + inherit permissions from their + tags: + - workspace + description: >- + Updates the permissions on a workspace object. Workspace objects can + inherit permissions from their + + parent objects or root object. + + + :param workspace_object_type: str + The workspace object type for which to get or manage permissions. Could be one of the following: + alerts, alertsv2, dashboards, dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + :param workspace_object_id: str + The workspace object for which to get or manage permissions. + :param access_control_list: + List[:class:`WorkspaceObjectAccessControlRequest`] (optional) + + + :returns: :class:`WorkspaceObjectPermissions` + parameters: + - name: workspace_object_type + in: path + required: true + schema: + type: string + description: >- + The workspace object type for which to get or manage permissions. + Could be one of the following: alerts, alertsv2, dashboards, + dbsql-dashboards, directories, experiments, files, genie, notebooks, + queries + - name: workspace_object_id + in: path + required: true + schema: + type: string + description: The workspace object for which to get or manage permissions. + requestBody: + content: + application/json: + schema: + type: object + properties: + access_control_list: + type: string + description: ':returns: :class:`WorkspaceObjectPermissions`' + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/WorkspaceObjectPermissions' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + /api/2.0/workspace/get-status: + get: + operationId: workspace_get_status + summary: >- + Gets the status of an object or a directory. If `path` does not exist, + this call returns an error + tags: + - workspace + description: >- + Gets the status of an object or a directory. If `path` does not exist, + this call returns an error + + `RESOURCE_DOES_NOT_EXIST`. + + + :param path: str + The absolute path of the notebook or directory. + + :returns: :class:`ObjectInfo` + parameters: + - name: path + in: query + required: true + schema: + type: string + description: The absolute path of the notebook or directory. + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ObjectInfo' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + /api/2.0/workspace/import: + post: + operationId: workspace_import_ + summary: >- + Imports a workspace object (for example, a notebook or file) or the + contents of an entire directory. + tags: + - workspace + description: >- + Imports a workspace object (for example, a notebook or file) or the + contents of an entire directory. + + If `path` already exists and `overwrite` is set to `false`, this call + returns an error + + `RESOURCE_ALREADY_EXISTS`. To import a directory, you can use either the + `DBC` format or the `SOURCE` + + format with the `language` field unset. To import a single file as + `SOURCE`, you must set the + + `language` field. Zip files within directories are not supported. + + + :param path: str + The absolute path of the object or directory. Importing a directory is only supported for the `DBC` + and `SOURCE` formats. + :param content: str (optional) + The base64-encoded content. This has a limit of 10 MB. + + If the limit (10MB) is exceeded, exception with error code **MAX_NOTEBOOK_SIZE_EXCEEDED** is thrown. + This parameter might be absent, and instead a posted file is used. + :param format: :class:`ImportFormat` (optional) + This specifies the format of the file to be imported. + + The value is case sensitive. + + - `AUTO`: The item is imported depending on an analysis of the item's extension and the header + content provided in the request. If the item is imported as a notebook, then the item's extension is + automatically removed. - `SOURCE`: The notebook or directory is imported as source code. - `HTML`: + The notebook is imported as an HTML file. - `JUPYTER`: The notebook is imported as a Jupyter/IPython + Notebook file. - `DBC`: The notebook is imported in Databricks archive format. Required for + directories. - `R_MARKDOWN`: The notebook is imported from R Markdown format. + :param language: :class:`Language` (optional) + The language of the object. This value is set only if the object type is `NOTEBOOK`. + :param overwrite: bool (optional) + The flag that specifies whether to overwrite existing object. It is `false` by default. For `DBC` + format, `overwrite` is not supported since it may contain a directory. + requestBody: + content: + application/json: + schema: + type: object + properties: + path: + type: string + description: >- + The absolute path of the object or directory. Importing a + directory is only supported for the `DBC` and `SOURCE` + formats. + content: + type: string + description: >- + The base64-encoded content. This has a limit of 10 MB. If + the limit (10MB) is exceeded, exception with error code + **MAX_NOTEBOOK_SIZE_EXCEEDED** is thrown. This parameter + might be absent, and instead a posted file is used. + format: + type: string + description: >- + This specifies the format of the file to be imported. The + value is case sensitive. - `AUTO`: The item is imported + depending on an analysis of the item's extension and the + header content provided in the request. If the item is + imported as a notebook, then the item's extension is + automatically removed. - `SOURCE`: The notebook or directory + is imported as source code. - `HTML`: The notebook is + imported as an HTML file. - `JUPYTER`: The notebook is + imported as a Jupyter/IPython Notebook file. - `DBC`: The + notebook is imported in Databricks archive format. Required + for directories. - `R_MARKDOWN`: The notebook is imported + from R Markdown format. + language: + type: string + description: >- + The language of the object. This value is set only if the + object type is `NOTEBOOK`. + overwrite: + type: string + description: >- + The flag that specifies whether to overwrite existing + object. It is `false` by default. For `DBC` format, + `overwrite` is not supported since it may contain a + directory. + required: + - path + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + /api/2.0/workspace/list: + get: + operationId: workspace_list + summary: >- + Lists the contents of a directory, or the object if it is not a + directory. If the input path does not + tags: + - workspace + description: >- + Lists the contents of a directory, or the object if it is not a + directory. If the input path does not + + exist, this call returns an error `RESOURCE_DOES_NOT_EXIST`. + + + :param path: str + The absolute path of the notebook or directory. + :param notebooks_modified_after: int (optional) + UTC timestamp in milliseconds + + :returns: Iterator over :class:`ObjectInfo` + parameters: + - name: path + in: query + required: true + schema: + type: string + description: The absolute path of the notebook or directory. + - name: notebooks_modified_after + in: query + required: false + schema: + type: string + description: UTC timestamp in milliseconds + responses: + '200': + description: Success + content: + application/json: + schema: + $ref: '#/components/schemas/ListResponse' + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI + /api/2.0/workspace/mkdirs: + post: + operationId: workspace_mkdirs + summary: >- + Creates the specified directory (and necessary parent directories if + they do not exist). If there is + tags: + - workspace + description: >- + Creates the specified directory (and necessary parent directories if + they do not exist). If there is + + an object (not a directory) at any prefix of the input path, this call + returns an error + + `RESOURCE_ALREADY_EXISTS`. + + + Note that if this operation fails it may have succeeded in creating some + of the necessary parent + + directories. + + + :param path: str + The absolute path of the directory. If the parent directories do not exist, it will also create + them. If the directory already exists, this command will do nothing and succeed. + requestBody: + content: + application/json: + schema: + type: object + properties: + path: + type: string + description: >- + The absolute path of the directory. If the parent + directories do not exist, it will also create them. If the + directory already exists, this command will do nothing and + succeed. + required: + - path + responses: + '200': + description: Success + default: + description: Error response + content: + application/json: + schema: + type: object + properties: + error: + type: string + message: + type: string + x-stackql-sdk-source: WorkspaceAPI +components: + schemas: + AclItem: + type: object + properties: + principal: + type: string + description: The principal in which the permission is applied. + permission: + $ref: '#/components/schemas/AclPermission' + description: The permission level applied to the principal. + required: + - principal + - permission + description: >- + An item representing an ACL rule applied to the given principal (user or + group) on the + associated scope point. + AzureKeyVaultSecretScopeMetadata: + type: object + properties: + resource_id: + type: string + description: >- + The resource id of the azure KeyVault that user wants to associate + the scope with. + dns_name: + type: string + description: The DNS of the KeyVault + required: + - resource_id + - dns_name + description: >- + The metadata of the Azure KeyVault for a secret scope of type + `AZURE_KEYVAULT` + CreateCredentialsResponse: + type: object + properties: + credential_id: + type: integer + git_provider: + type: string + description: The Git provider associated with the credential. + git_email: + type: string + description: >- + The authenticating email associated with your Git provider user + account. Used for authentication with the remote repository and also + sets the author & committer identity for commits. Required for most + Git providers except AWS CodeCommit. Learn more at + https://docs.databricks.com/aws/en/repos/get-access-tokens-from-git-provider + git_username: + type: string + description: >- + The username provided with your Git provider account and associated + with the credential. For most Git providers it is only used to set + the Git committer & author names for commits, however it may be + required for authentication depending on your Git provider / token + requirements. Required for AWS CodeCommit. + is_default_for_provider: + type: boolean + description: if the credential is the default for the given provider + name: + type: string + description: >- + the name of the git credential, used for identification and ease of + lookup + required: + - credential_id + - git_provider + CreateRepoResponse: + type: object + properties: + branch: + type: string + head_commit_id: + type: string + description: >- + SHA-1 hash representing the commit ID of the current HEAD of the Git + folder (repo). + id: + type: integer + description: ID of the Git folder (repo) object in the workspace. + path: + type: string + description: Path of the Git folder (repo) in the workspace. + provider: + type: string + description: Git provider of the linked Git repository. + sparse_checkout: + $ref: '#/components/schemas/SparseCheckout' + description: Sparse checkout settings for the Git folder (repo). + url: + type: string + description: URL of the linked Git repository. + CredentialInfo: + type: object + properties: + credential_id: + type: integer + git_email: + type: string + description: >- + The authenticating email associated with your Git provider user + account. Used for authentication with the remote repository and also + sets the author & committer identity for commits. Required for most + Git providers except AWS CodeCommit. Learn more at + https://docs.databricks.com/aws/en/repos/get-access-tokens-from-git-provider + git_provider: + type: string + description: The Git provider associated with the credential. + git_username: + type: string + description: >- + The username provided with your Git provider account and associated + with the credential. For most Git providers it is only used to set + the Git committer & author names for commits, however it may be + required for authentication depending on your Git provider / token + requirements. Required for AWS CodeCommit. + is_default_for_provider: + type: boolean + description: if the credential is the default for the given provider + name: + type: string + description: >- + the name of the git credential, used for identification and ease of + lookup + required: + - credential_id + DeleteCredentialsResponse: + type: object + properties: {} + DeleteRepoResponse: + type: object + properties: {} + DeleteResponse: + type: object + properties: {} + DeleteSecretResponse: + type: object + properties: {} + ExportResponse: + type: object + properties: + content: + type: string + description: >- + The base64-encoded content. If the limit (10MB) is exceeded, + exception with error code **MAX_NOTEBOOK_SIZE_EXCEEDED** is thrown. + file_type: + type: string + description: The file type of the exported file. + description: >- + The request field `direct_download` determines whether a JSON response + or binary contents are + returned by this endpoint. + GetCredentialsResponse: + type: object + properties: + credential_id: + type: integer + git_email: + type: string + description: >- + The authenticating email associated with your Git provider user + account. Used for authentication with the remote repository and also + sets the author & committer identity for commits. Required for most + Git providers except AWS CodeCommit. Learn more at + https://docs.databricks.com/aws/en/repos/get-access-tokens-from-git-provider + git_provider: + type: string + description: The Git provider associated with the credential. + git_username: + type: string + description: >- + The username provided with your Git provider account and associated + with the credential. For most Git providers it is only used to set + the Git committer & author names for commits, however it may be + required for authentication depending on your Git provider / token + requirements. Required for AWS CodeCommit. + is_default_for_provider: + type: boolean + description: if the credential is the default for the given provider + name: + type: string + description: >- + the name of the git credential, used for identification and ease of + lookup + required: + - credential_id + GetRepoPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/RepoPermissionsDescription' + GetRepoResponse: + type: object + properties: + branch: + type: string + head_commit_id: + type: string + description: >- + SHA-1 hash representing the commit ID of the current HEAD of the + repo. + id: + type: integer + description: ID of the Git folder (repo) object in the workspace. + path: + type: string + description: Path of the Git folder (repo) in the workspace. + provider: + type: string + description: Git provider of the linked Git repository. + sparse_checkout: + $ref: '#/components/schemas/SparseCheckout' + description: Sparse checkout settings for the Git folder (repo). + url: + type: string + description: URL of the linked Git repository. + GetSecretResponse: + type: object + properties: + key: + type: string + value: + type: string + description: The value of the secret in its byte representation. + GetWorkspaceObjectPermissionLevelsResponse: + type: object + properties: + permission_levels: + type: array + items: + $ref: '#/components/schemas/WorkspaceObjectPermissionsDescription' + ImportResponse: + type: object + properties: {} + ListAclsResponse: + type: object + properties: + items: + type: array + items: + $ref: '#/components/schemas/AclItem' + ListCredentialsResponse: + type: object + properties: + credentials: + type: array + items: + $ref: '#/components/schemas/CredentialInfo' + ListReposResponse: + type: object + properties: + next_page_token: + type: string + repos: + type: array + items: + $ref: '#/components/schemas/RepoInfo' + description: List of Git folders (repos). + ListResponse: + type: object + properties: + objects: + type: array + items: + $ref: '#/components/schemas/ObjectInfo' + ListScopesResponse: + type: object + properties: + scopes: + type: array + items: + $ref: '#/components/schemas/SecretScope' + ListSecretsResponse: + type: object + properties: + secrets: + type: array + items: + $ref: '#/components/schemas/SecretMetadata' + MkdirsResponse: + type: object + properties: {} + ObjectInfo: + type: object + properties: + created_at: + type: integer + description: Only applicable to files. The creation UTC timestamp. + language: + $ref: '#/components/schemas/Language' + description: >- + The language of the object. This value is set only if the object + type is ``NOTEBOOK``. + modified_at: + type: integer + description: Only applicable to files, the last modified UTC timestamp. + object_id: + type: integer + description: Unique identifier for the object. + object_type: + $ref: '#/components/schemas/ObjectType' + description: >- + The type of the object in workspace. - `NOTEBOOK`: document that + contains runnable code, visualizations, and explanatory text. - + `DIRECTORY`: directory - `LIBRARY`: library - `FILE`: file - `REPO`: + repository - `DASHBOARD`: Lakeview dashboard + path: + type: string + description: The absolute path of the object. + resource_id: + type: string + description: >- + A unique identifier for the object that is consistent across all + Databricks APIs. + size: + type: integer + description: Only applicable to files. The file size in bytes can be returned. + description: >- + The information of the object in workspace. It will be returned by + ``list`` and ``get-status``. + RepoAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/RepoPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + RepoAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/RepoPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + RepoInfo: + type: object + properties: + branch: + type: string + description: Name of the current git branch of the git folder (repo). + head_commit_id: + type: string + description: Current git commit id of the git folder (repo). + id: + type: integer + description: Id of the git folder (repo) in the Workspace. + path: + type: string + description: Root path of the git folder (repo) in the Workspace. + provider: + type: string + description: Git provider of the remote git repository, e.g. `gitHub`. + sparse_checkout: + $ref: '#/components/schemas/SparseCheckout' + description: Sparse checkout config for the git folder (repo). + url: + type: string + description: URL of the remote git repository. + description: Git folder (repo) information. + RepoPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/RepoPermissionLevel' + RepoPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/RepoAccessControlResponse' + object_id: + type: string + object_type: + type: string + RepoPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/RepoPermissionLevel' + SecretMetadata: + type: object + properties: + key: + type: string + description: A unique name to identify the secret. + last_updated_timestamp: + type: integer + description: The last updated timestamp (in milliseconds) for the secret. + description: >- + The metadata about a secret. Returned when listing secrets. Does not + contain the actual secret + value. + SecretScope: + type: object + properties: + backend_type: + $ref: '#/components/schemas/ScopeBackendType' + description: The type of secret scope backend. + keyvault_metadata: + $ref: '#/components/schemas/AzureKeyVaultSecretScopeMetadata' + description: The metadata for the secret scope if the type is ``AZURE_KEYVAULT`` + name: + type: string + description: A unique name to identify the secret scope. + description: >- + An organizational resource for storing secrets. Secret scopes can be + different types + (Databricks-managed, Azure KeyVault backed, etc), and ACLs can be applied to control permissions + for all secrets within a scope. + SparseCheckout: + type: object + properties: + patterns: + type: array + items: + type: string + description: >- + List of sparse checkout cone patterns, see [cone mode handling] for + details. [cone mode handling]: + https://git-scm.com/docs/git-sparse-checkout#_internalscone_mode_handling + description: Sparse checkout configuration, it contains options like cone patterns. + SparseCheckoutUpdate: + type: object + properties: + patterns: + type: array + items: + type: string + description: >- + List of sparse checkout cone patterns, see [cone mode handling] for + details. [cone mode handling]: + https://git-scm.com/docs/git-sparse-checkout#_internalscone_mode_handling + description: Sparse checkout configuration, it contains options like cone patterns. + UpdateCredentialsResponse: + type: object + properties: {} + UpdateRepoResponse: + type: object + properties: {} + WorkspaceObjectAccessControlRequest: + type: object + properties: + group_name: + type: string + permission_level: + $ref: '#/components/schemas/WorkspaceObjectPermissionLevel' + service_principal_name: + type: string + description: application ID of a service principal + user_name: + type: string + description: name of the user + WorkspaceObjectAccessControlResponse: + type: object + properties: + all_permissions: + type: array + items: + $ref: '#/components/schemas/WorkspaceObjectPermission' + display_name: + type: string + description: Display name of the user or service principal. + group_name: + type: string + description: name of the group + service_principal_name: + type: string + description: Name of the service principal. + user_name: + type: string + description: name of the user + WorkspaceObjectPermission: + type: object + properties: + inherited: + type: boolean + inherited_from_object: + type: array + items: + type: string + permission_level: + $ref: '#/components/schemas/WorkspaceObjectPermissionLevel' + WorkspaceObjectPermissions: + type: object + properties: + access_control_list: + type: array + items: + $ref: '#/components/schemas/WorkspaceObjectAccessControlResponse' + object_id: + type: string + object_type: + type: string + WorkspaceObjectPermissionsDescription: + type: object + properties: + description: + type: string + permission_level: + $ref: '#/components/schemas/WorkspaceObjectPermissionLevel' + AclPermission: + type: string + enum: + - MANAGE + - READ + - WRITE + description: The ACL permission levels for Secret ACLs applied to secret scopes. + ExportFormat: + type: string + enum: + - AUTO + - DBC + - HTML + - JUPYTER + - RAW + - R_MARKDOWN + - SOURCE + description: The format for workspace import and export. + ImportFormat: + type: string + enum: + - AUTO + - DBC + - HTML + - JUPYTER + - RAW + - R_MARKDOWN + - SOURCE + description: The format for workspace import and export. + Language: + type: string + enum: + - PYTHON + - R + - SCALA + - SQL + description: The language of notebook. + ObjectType: + type: string + enum: + - DASHBOARD + - DIRECTORY + - FILE + - LIBRARY + - NOTEBOOK + - REPO + description: The type of the object in workspace. + RepoPermissionLevel: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_READ + - CAN_RUN + description: Permission level + ScopeBackendType: + type: string + enum: + - AZURE_KEYVAULT + - DATABRICKS + description: >- + The types of secret scope backends in the Secret Manager. Azure KeyVault + backed secret scopes + + will be supported in a later release. + WorkspaceObjectPermissionLevel: + type: string + enum: + - CAN_EDIT + - CAN_MANAGE + - CAN_READ + - CAN_RUN + description: Permission level + x-stackQL-resources: + git_credentials: + id: databricks_workspace.workspace.git_credentials + name: git_credentials + title: Git Credentials + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1git-credentials/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1git-credentials/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.credentials + delete: + operation: + $ref: '#/paths/~1api~12.0~1git-credentials~1{credential_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1git-credentials~1{credential_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1git-credentials~1{credential_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/git_credentials/methods/get' + - $ref: '#/components/x-stackQL-resources/git_credentials/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/git_credentials/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/git_credentials/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/git_credentials/methods/delete' + replace: [] + repos: + id: databricks_workspace.workspace.repos + name: repos + title: Repos + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1repos/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1repos/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.repos + delete: + operation: + $ref: '#/paths/~1api~12.0~1repos~1{repo_id}/delete' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1repos~1{repo_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1repos~1{repo_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/repos/methods/get' + - $ref: '#/components/x-stackQL-resources/repos/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/repos/methods/create' + update: + - $ref: '#/components/x-stackQL-resources/repos/methods/update' + delete: + - $ref: '#/components/x-stackQL-resources/repos/methods/delete' + replace: [] + repo_permission_levels: + id: databricks_workspace.workspace.repo_permission_levels + name: repo_permission_levels + title: Repo Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1repos~1{repo_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/repo_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + repo_permissions: + id: databricks_workspace.workspace.repo_permissions + name: repo_permissions + title: Repo Permissions + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1repos~1{repo_id}/get' + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1repos~1{repo_id}/put' + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: '#/paths/~1api~12.0~1permissions~1repos~1{repo_id}/patch' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/repo_permissions/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/repo_permissions/methods/update' + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/repo_permissions/methods/set' + secret_scopes: + id: databricks_workspace.workspace.secret_scopes + name: secret_scopes + title: Secret Scopes + methods: + create: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1scopes~1create/post' + response: + mediaType: application/json + openAPIDocKey: '200' + delete: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1scopes~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1scopes~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.scopes + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/secret_scopes/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/secret_scopes/methods/create' + update: [] + delete: [] + replace: [] + secret_acls: + id: databricks_workspace.workspace.secret_acls + name: secret_acls + title: Secret Acls + methods: + delete: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1acls~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1acls~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1acls~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.items + put: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1acls~1put/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/secret_acls/methods/get' + - $ref: '#/components/x-stackQL-resources/secret_acls/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/secret_acls/methods/put' + update: [] + delete: [] + replace: [] + secrets: + id: databricks_workspace.workspace.secrets + name: secrets + title: Secrets + methods: + delete: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + get: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1get/get' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.secrets + put: + operation: + $ref: '#/paths/~1api~12.0~1secrets~1put/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/secrets/methods/get' + - $ref: '#/components/x-stackQL-resources/secrets/methods/list' + insert: + - $ref: '#/components/x-stackQL-resources/secrets/methods/put' + update: [] + delete: [] + replace: [] + workspace: + id: databricks_workspace.workspace.workspace + name: workspace + title: Workspace + methods: + delete: + operation: + $ref: '#/paths/~1api~12.0~1workspace~1delete/post' + response: + mediaType: application/json + openAPIDocKey: '200' + export: + operation: + $ref: '#/paths/~1api~12.0~1workspace~1export/get' + response: + mediaType: application/json + openAPIDocKey: '200' + import: + operation: + $ref: '#/paths/~1api~12.0~1workspace~1import/post' + response: + mediaType: application/json + openAPIDocKey: '200' + list: + operation: + $ref: '#/paths/~1api~12.0~1workspace~1list/get' + response: + mediaType: application/json + openAPIDocKey: '200' + objectKey: $.objects + mkdirs: + operation: + $ref: '#/paths/~1api~12.0~1workspace~1mkdirs/post' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/workspace/methods/list' + insert: [] + update: [] + delete: [] + replace: [] + object_permission_levels: + id: databricks_workspace.workspace.object_permission_levels + name: object_permission_levels + title: Object Permission Levels + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}~1permissionLevels/get + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: >- + #/components/x-stackQL-resources/object_permission_levels/methods/get + insert: [] + update: [] + delete: [] + replace: [] + object_permissions: + id: databricks_workspace.workspace.object_permissions + name: object_permissions + title: Object Permissions + methods: + get: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}/get + response: + mediaType: application/json + openAPIDocKey: '200' + set: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}/put + response: + mediaType: application/json + openAPIDocKey: '200' + update: + operation: + $ref: >- + #/paths/~1api~12.0~1permissions~1{workspace_object_type}~1{workspace_object_id}/patch + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/object_permissions/methods/get' + insert: [] + update: + - $ref: '#/components/x-stackQL-resources/object_permissions/methods/update' + delete: [] + replace: + - $ref: '#/components/x-stackQL-resources/object_permissions/methods/set' + workspace_status: + id: databricks_workspace.workspace.workspace_status + name: workspace_status + title: Workspace Status + methods: + get: + operation: + $ref: '#/paths/~1api~12.0~1workspace~1get-status/get' + response: + mediaType: application/json + openAPIDocKey: '200' + sqlVerbs: + select: + - $ref: '#/components/x-stackQL-resources/workspace_status/methods/get' + insert: [] + update: [] + delete: [] + replace: []