Refactor reading env vars (#888) #1356
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Integration Tests | |
on: | |
push: | |
paths-ignore: | |
- "**.MD" | |
- "**.md" | |
- "adapters/databricks/__version__.py" | |
- "tests/unit/**" | |
- ".github/workflows/main.yml" | |
- ".github/workflows/stale.yml" | |
concurrency: | |
group: ${{ github.workflow }}-${{ github.ref }} | |
cancel-in-progress: true | |
jobs: | |
run-uc-cluster-e2e-tests: | |
runs-on: ubuntu-latest | |
environment: azure-prod | |
env: | |
DBT_DATABRICKS_HOST_NAME: ${{ secrets.DATABRICKS_HOST }} | |
DBT_DATABRICKS_CLIENT_ID: ${{ secrets.TEST_PECO_SP_ID }} | |
DBT_DATABRICKS_CLIENT_SECRET: ${{ secrets.TEST_PECO_SP_SECRET }} | |
DBT_DATABRICKS_UC_INITIAL_CATALOG: peco | |
DBT_DATABRICKS_LOCATION_ROOT: ${{ secrets.TEST_PECO_EXTERNAL_LOCATION }}test | |
TEST_PECO_UC_CLUSTER_ID: ${{ secrets.TEST_PECO_UC_CLUSTER_ID }} | |
steps: | |
- name: Check out repository | |
uses: actions/checkout@v4 | |
- name: Set up python | |
id: setup-python | |
uses: actions/setup-python@v5 | |
with: | |
python-version: "3.9" | |
- name: Get http path from environment | |
run: python .github/workflows/build_cluster_http_path.py | |
shell: sh | |
- name: Install Hatch | |
id: install-dependencies | |
uses: pypa/hatch@install | |
- name: Run UC Cluster Functional Tests | |
run: [email protected] DBT_DATABRICKS_LOCATION_ROOT=$DBT_DATABRICKS_LOCATION_ROOT DBT_DATABRICKS_HOST_NAME=$DBT_DATABRICKS_HOST_NAME DBT_DATABRICKS_UC_CLUSTER_HTTP_PATH=$DBT_DATABRICKS_UC_CLUSTER_HTTP_PATH DBT_DATABRICKS_CLIENT_ID=$DBT_DATABRICKS_CLIENT_ID DBT_DATABRICKS_CLIENT_SECRET=$DBT_DATABRICKS_CLIENT_SECRET hatch -v run uc-cluster-e2e | |
run-sqlwarehouse-e2e-tests: | |
runs-on: ubuntu-latest | |
environment: azure-prod | |
env: | |
DBT_DATABRICKS_HOST_NAME: ${{ secrets.DATABRICKS_HOST }} | |
DBT_DATABRICKS_CLIENT_ID: ${{ secrets.TEST_PECO_SP_ID }} | |
DBT_DATABRICKS_CLIENT_SECRET: ${{ secrets.TEST_PECO_SP_SECRET }} | |
DBT_DATABRICKS_HTTP_PATH: ${{ secrets.TEST_PECO_WAREHOUSE_HTTP_PATH }} | |
DBT_DATABRICKS_UC_INITIAL_CATALOG: peco | |
DBT_DATABRICKS_LOCATION_ROOT: ${{ secrets.TEST_PECO_EXTERNAL_LOCATION }}test | |
TEST_PECO_UC_CLUSTER_ID: ${{ secrets.TEST_PECO_UC_CLUSTER_ID }} | |
steps: | |
- name: Check out repository | |
uses: actions/checkout@v4 | |
- name: Set up python | |
id: setup-python | |
uses: actions/setup-python@v5 | |
with: | |
python-version: "3.9" | |
- name: Get http path from environment | |
run: python .github/workflows/build_cluster_http_path.py | |
shell: sh | |
- name: Install Hatch | |
id: install-dependencies | |
uses: pypa/hatch@install | |
- name: Run Sql Endpoint Functional Tests | |
run: [email protected] DBT_DATABRICKS_LOCATION_ROOT=$DBT_DATABRICKS_LOCATION_ROOT DBT_DATABRICKS_HOST_NAME=$DBT_DATABRICKS_HOST_NAME DBT_DATABRICKS_UC_CLUSTER_HTTP_PATH=$DBT_DATABRICKS_UC_CLUSTER_HTTP_PATH DBT_DATABRICKS_CLIENT_ID=$DBT_DATABRICKS_CLIENT_ID DBT_DATABRICKS_CLIENT_SECRET=$DBT_DATABRICKS_CLIENT_SECRET hatch -v run sqlw-e2e | |
run-cluster-e2e-tests: | |
runs-on: ubuntu-latest | |
environment: azure-prod | |
env: | |
DBT_DATABRICKS_HOST_NAME: ${{ secrets.DATABRICKS_HOST }} | |
DBT_DATABRICKS_TOKEN: ${{ secrets.DATABRICKS_TOKEN }} | |
TEST_PECO_CLUSTER_ID: ${{ secrets.TEST_PECO_CLUSTER_ID }} | |
DBT_DATABRICKS_LOCATION_ROOT: ${{ secrets.TEST_PECO_EXTERNAL_LOCATION }}test | |
steps: | |
- name: Check out repository | |
uses: actions/checkout@v4 | |
- name: Set up python | |
id: setup-python | |
uses: actions/setup-python@v5 | |
with: | |
python-version: "3.9" | |
- name: Get http path from environment | |
run: python .github/workflows/build_cluster_http_path.py | |
shell: sh | |
- name: Install Hatch | |
id: install-dependencies | |
uses: pypa/hatch@install | |
- name: Run Cluster Functional Tests | |
run: [email protected] DBT_DATABRICKS_LOCATION_ROOT=$DBT_DATABRICKS_LOCATION_ROOT DBT_DATABRICKS_HOST_NAME=$DBT_DATABRICKS_HOST_NAME DBT_DATABRICKS_HTTP_PATH=$DBT_DATABRICKS_CLUSTER_HTTP_PATH DBT_DATABRICKS_CLIENT_ID=$DBT_DATABRICKS_CLIENT_ID DBT_DATABRICKS_CLIENT_SECRET=$DBT_DATABRICKS_CLIENT_SECRET hatch -v run cluster-e2e |