Skip to content
This repository has been archived by the owner on Apr 8, 2024. It is now read-only.

chore: remove aws tests #898

Merged
merged 1 commit into from
Aug 31, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
30 changes: 0 additions & 30 deletions .github/workflows/test_integration_adapter.yml
Original file line number Diff line number Diff line change
Expand Up @@ -28,9 +28,7 @@ jobs:
- postgres
- bigquery
- snowflake
- redshift
# TODO: enable as 1.5 becomes available for following adapters
# - athena
# - trino
# - duckdb
# - sqlserver
Expand All @@ -45,8 +43,6 @@ jobs:
- profile: snowflake
teleport: true
cloud: true
# - profile: redshift
# cloud: true
- profile: bigquery
cloud: true

Expand Down Expand Up @@ -93,10 +89,6 @@ jobs:
pip install --upgrade pip

ADAPTER_PACKAGE="dbt-${{ matrix.profile }}"
if [[ '${{ matrix.profile }}' == 'athena' ]]
then
ADAPTER_PACKAGE="dbt-athena-community"
fi

if [[ -n '${{ matrix.dbt_version }}' ]]
then
Expand Down Expand Up @@ -152,20 +144,6 @@ jobs:
SF_SCHEMA: ${{ secrets.SF_SCHEMA }}
# Duckdb
DB_PATH: ${{ github.workspace }}/duck.db
# Redshift
RS_HOST: ${{ secrets.RS_HOST }}
RS_PORT: ${{ secrets.RS_PORT }}
RS_USER: ${{ secrets.RS_USER }}
RS_PASSWORD: ${{ secrets.RS_PASSWORD }}
RS_DBNAME: ${{ secrets.RS_DBNAME }}
RS_SCHEMA: ${{ secrets.RS_SCHEMA }}
# Athena
ATHENA_S3_STAGING_DIR: ${{ secrets.ATHENA_S3_STAGING_DIR }}
ATHENA_REGION: ${{ secrets.ATHENA_REGION }}
ATHENA_WORK_GROUP: ${{ secrets.ATHENA_WORK_GROUP }}
ATHENA_PROFILE: ${{ secrets.ATHENA_PROFILE }}
ATHENA_DATABASE: ${{ secrets.ATHENA_DATABASE }}
ATHENA_SCHEMA: ${{ secrets.ATHENA_SCHEMA }}
run: |
source .venv/bin/activate

Expand All @@ -185,14 +163,6 @@ jobs:
# schema: "{{ env_var('DBT_SCHEMA', 'dbt_fal') }}"
export DBT_DATABASE="" DBT_SCHEMA=""
fi
if [[ '${{ matrix.profile }}' == "redshift" ]]
then
export DBT_DATABASE="$RS_DBNAME" DBT_SCHEMA="$RS_SCHEMA"
fi
if [[ '${{ matrix.profile }}' == "athena" ]]
then
export DBT_DATABASE="$ATHENA_DATABASE" DBT_SCHEMA="$ATHENA_SCHEMA"
fi

if [[ '${{ matrix.profile }}' == "bigquery" ]]
then
Expand Down
39 changes: 1 addition & 38 deletions .github/workflows/test_integration_cli.yml
Original file line number Diff line number Diff line change
Expand Up @@ -29,9 +29,7 @@ on:
- postgres
- bigquery
- snowflake
- redshift
- duckdb
- athena
- fal

python:
Expand Down Expand Up @@ -72,9 +70,7 @@ jobs:
'fal'
]
EXTRA_OPTIONS = [
'redshift',
'duckdb',
'athena',
]
OUTPUT = OPTIONS

Expand Down Expand Up @@ -211,19 +207,9 @@ jobs:

- name: Start Docker database
working-directory: projects/adapter/cli_tests
# TODO: make redshift use a real connection
if: contains(fromJSON('["redshift", "postgres", "fal"]'), matrix.profile)
if: contains(fromJSON('["postgres", "fal"]'), matrix.profile)
run: docker-compose up -d

- name: Configure AWS Credentials
uses: aws-actions/configure-aws-credentials@v1
# TODO: make redshift use a real connection
if: contains(fromJSON('["redshift", "athena"]'), matrix.profile)
with:
aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }}
aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
aws-region: us-east-1

- name: Install conda
uses: s-weigand/setup-conda@v1
with:
Expand All @@ -245,14 +231,6 @@ jobs:
KEYFILE: ${{ secrets.GCP_SA_KEY }}
GCLOUD_PROJECT: ${{ secrets.GCP_PROJECT_ID }}
BQ_DATASET: ${{ secrets.BQ_DATASET }}
# Redshift
# TODO: change to {{ secrets.* }} when testing with real Redshift
RS_DB_NAME: test
RS_SCHEMA: dbt_fal
RS_HOST: localhost
RS_PASSWORD: pass
RS_PORT: 5432
RS_USER: pguser
# Snowflake
SF_ACCOUNT: ${{ secrets.SF_ACCOUNT }}
SF_USER: ${{ secrets.SF_USER }}
Expand All @@ -263,23 +241,12 @@ jobs:
SF_SCHEMA: ${{ secrets.SF_SCHEMA }}
# Duckdb
DB_PATH: ${{ github.workspace }}/duck.db
# Athena
ATHENA_S3_STAGING_DIR: ${{ secrets.ATHENA_S3_STAGING_DIR }}
ATHENA_REGION: ${{ secrets.ATHENA_REGION }}
ATHENA_WORK_GROUP: ${{ secrets.ATHENA_WORK_GROUP }}
ATHENA_PROFILE: ${{ secrets.ATHENA_PROFILE }}
ATHENA_DATABASE: ${{ secrets.ATHENA_DATABASE }}
ATHENA_SCHEMA: ${{ secrets.ATHENA_SCHEMA }}
run: |
# Database and schema setup for sources
if [[ '${{ matrix.profile }}' == "bigquery" ]]
then
export DBT_DATABASE="$GCLOUD_PROJECT" DBT_SCHEMA="$BQ_DATASET"
fi
if [[ '${{ matrix.profile }}' == "redshift" ]]
then
export DBT_DATABASE="$RS_DB_NAME" DBT_SCHEMA="$RS_SCHEMA"
fi
if [[ '${{ matrix.profile }}' == "snowflake" ]]
then
export DBT_DATABASE="$SF_DATABASE" DBT_SCHEMA="$SF_SCHEMA"
Expand All @@ -291,10 +258,6 @@ jobs:
# schema: "{{ env_var('DBT_SCHEMA', 'dbt_fal') }}"
export DBT_DATABASE="" DBT_SCHEMA=""
fi
if [[ '${{ matrix.profile }}' == "athena" ]]
then
export DBT_DATABASE="$ATHENA_DATABASE" DBT_SCHEMA="$ATHENA_SCHEMA"
fi

if [[ '${{ matrix.profile }}' == "bigquery" ]]
then
Expand Down
Loading