Permalink
Cannot retrieve contributors at this time
96 lines (86 sloc)
4.64 KB
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# ,-. ;-. ,--. . . . , ,--. ,-. ,-. ,--. ,-. ,-. . . ,--. , ,-. # | |
# / \ | ) | |\ | | / | | ) ( ` | / / \ |\ | | | / # | |
# | | |-' |- | \| | / |- |-< `-. |- | | | | \| |- | | -. # | |
# \ / | | | | |/ | | \ . ) | \ \ / | | | | \ | # | |
# `-' ' `--' ' ' ' `--' ' ' `-' `--' `-' `-' ' ' ' ' `-' # | |
######################################################################################## | |
# Airflow Settings | |
######################################################################################## | |
# Some brand-based suggestions: #C52B9B (pink), #FFE033 (yellow) | |
AIRFLOW__WEBSERVER__NAVBAR_COLOR="#FFF" | |
# Disabled by default to make development easier | |
# (enabled on prod for security) | |
AIRFLOW__CORE__HIDE_SENSITIVE_VAR_CONN_FIELDS=False | |
# Use the following python code to generate a fernet key for production | |
# python -c "import base64, os; print(base64.urlsafe_b64encode(os.urandom(32)).decode())" | |
# AIRFLOW__CORE__FERNET_KEY= | |
# Executor to use | |
AIRFLOW__CORE__EXECUTOR=LocalExecutor | |
# Environment this instance is being run in | |
AIRFLOW_VAR_ENVIRONMENT=dev | |
######################################################################################## | |
# API Keys | |
######################################################################################## | |
# See: https://airflow.apache.org/docs/apache-airflow/stable/howto/variable.html#storing-variables-in-environment-variables | |
AIRFLOW_VAR_API_KEY_WALTERS_ART_MUSEUM=not_set | |
AIRFLOW_VAR_API_KEY_BROOKLYN_MUSEUM=not_set | |
AIRFLOW_VAR_API_KEY_DATA_GOV=not_set | |
AIRFLOW_VAR_API_KEY_EUROPEANA=not_set | |
AIRFLOW_VAR_API_KEY_FLICKR=not_set | |
AIRFLOW_VAR_API_KEY_JAMENDO=not_set | |
AIRFLOW_VAR_API_KEY_NYPL=not_set | |
AIRFLOW_VAR_API_KEY_THINGIVERSE=not_set | |
AIRFLOW_VAR_API_KEY_FREESOUND=not_set | |
######################################################################################## | |
# Connection/Variable info | |
######################################################################################## | |
# Airflow primary metadata database | |
# Change the following line in prod to use the appropriate DB | |
AIRFLOW__CORE__SQL_ALCHEMY_CONN=postgresql+psycopg2://airflow:airflow@postgres:5432/airflow | |
# Remote logging connection ID | |
# Replace "access_key" and "secret+key" with the real values. Secret key must be URL-encoded | |
AIRFLOW_CONN_AWS_DEFAULT=aws://test_key:test_secret@?region_name=us-east-1&host=http://s3:5000 | |
# Change the following line in prod to use the appropriate DB | |
AIRFLOW_CONN_POSTGRES_OPENLEDGER_UPSTREAM=postgres://deploy:deploy@postgres:5432/openledger | |
AIRFLOW_CONN_POSTGRES_OPENLEDGER_TESTING=postgres://deploy:deploy@postgres:5432/openledger | |
# Slack webhook connection info (this must be URL encoded, with a | |
# pre-pended https://, e.g. "https://https%3A%2F%2Fhooks.slack.com%2Fservices%2everythingelse") | |
# A distinction is made here between "notifications" and "alerts", the former being | |
# useful updates and the latter being alarms or actionable errors. | |
AIRFLOW_CONN_SLACK_NOTIFICATIONS=https://slack | |
AIRFLOW_CONN_SLACK_ALERTS=https://slack | |
OPENLEDGER_CONN_ID=postgres_openledger_upstream | |
TEST_CONN_ID=postgres_openledger_testing | |
S3_LOCAL_ENDPOINT=http://s3:5000 | |
AWS_CONN_ID=aws_default | |
AIRFLOW_CONN_EMR_EMPTY=emr:// | |
AIRFLOW_CONN_EMR_TEST=emr://?host=http://s3:5000 | |
EMR_CONN_ID=emr_empty | |
EMR_TEST_CONN_ID=emr_test | |
# Connection to the Ingestion Server, used for managing data refreshes. (This must be URL encoded, | |
# with a pre-pended https://, e.g. "https://https%3A%2F%2Fhost.docker.internal%3A8001")) | |
AIRFLOW_CONN_DATA_REFRESH=not_set | |
######################################################################################## | |
# Other config | |
######################################################################################## | |
# Version of the catalog docker image to use. Defaults to `latest` and is not used for | |
# local development (since the image is always built locally). See available tags at | |
# https://ghcr.io/wordpress/openverse-catalog | |
DOCKER_IMAGE_TAG=latest | |
# External port airflow will be mounted to | |
AIRFLOW_PORT=9090 | |
# Minutes to wait until processing a file that hasn't been modified | |
LOADER_FILE_AGE=1 | |
# Contact email for any APIs | |
[email protected] | |
# AWS/S3 configuration - does not need to be changed for development | |
AWS_ACCESS_KEY=test_key | |
AWS_SECRET_KEY=test_secret | |
# General bucket used for TSV->DB ingestion and logging | |
OPENVERSE_BUCKET=openverse-storage | |
# Used only for commoncrawl parsing | |
S3_BUCKET=not_set | |
COMMONCRAWL_BUCKET=not_set | |
# Seconds to wait before poking for availability of the data refresh pool when running a data_refresh | |
# DAG. Used to shorten the time for testing purposes. | |
DATA_REFRESH_POKE_INTERVAL=5 |