diff --git a/data-serving/scripts/export-data/setup_country_export.sh b/data-serving/scripts/export-data/setup_country_export.sh index 18c75211c..582e574c0 100755 --- a/data-serving/scripts/export-data/setup_country_export.sh +++ b/data-serving/scripts/export-data/setup_country_export.sh @@ -11,7 +11,7 @@ require_env "${BUCKET:-}" "Specify S3 bucket to output files in BUCKET" echo "Setting up country export job definitions for environment {ENV}..." -CASECOUNT_URL=${CASECOUNT_URL:-https://covid-19-aggregates.s3.amazonaws.com/country/latest.json} +CASECOUNT_URL=${CASECOUNT_URL:-https://covid-19-aggregates-eu.s3.amazonaws.com/country/latest.json} # mongoexport rate in cases/s # actual rate is higher, but this allows some wiggle room # in calculation of Batch job timeouts diff --git a/docs/data_landscape.md b/docs/data_landscape.md index ca64903c3..07b6d82b3 100644 --- a/docs/data_landscape.md +++ b/docs/data_landscape.md @@ -23,8 +23,8 @@ Various buckets (data containers) are used for both temporary and long-term stor Aggregated data from the line list used by the map visualisation. -* covid-19-aggregates -* covid-19-aggregates-dev +* covid-19-aggregates-eu +* covid-19-aggregates-dev-eu ### Export diff --git a/ingestion/monitoring/completeness.py b/ingestion/monitoring/completeness.py index ae018c669..c9f951870 100644 --- a/ingestion/monitoring/completeness.py +++ b/ingestion/monitoring/completeness.py @@ -114,8 +114,8 @@ def setup_logger(): setup_logger() endpoint_url = os.getenv("ENDPOINT_URL") objects = data_files( - os.getenv("COUNTRY_EXPORT_BUCKET", "covid-19-country-export"), + os.getenv("COUNTRY_EXPORT_BUCKET", "covid-19-country-export-eu"), endpoint_url=endpoint_url, ) data = completeness_s3_many(objects, endpoint_url) - upload(data, os.getenv("METRICS_BUCKET", "covid-19-aggregates"), endpoint_url) + upload(data, os.getenv("METRICS_BUCKET", "covid-19-aggregates-eu"), endpoint_url) diff --git a/ingestion/monitoring/daily_metrics.py b/ingestion/monitoring/daily_metrics.py index a4d64b512..0854c64c9 100644 --- a/ingestion/monitoring/daily_metrics.py +++ b/ingestion/monitoring/daily_metrics.py @@ -11,7 +11,7 @@ import boto3 -BUCKET = "covid-19-aggregates" +BUCKET = "covid-19-aggregates-eu" WEBHOOK_URL = os.environ.get("SLACK_WEBHOOK_METRICS_URL", None) logger = logging.getLogger(__name__) diff --git a/ingestion/monitoring/freshness.py b/ingestion/monitoring/freshness.py index 3e629d799..bb2b8d7ca 100644 --- a/ingestion/monitoring/freshness.py +++ b/ingestion/monitoring/freshness.py @@ -91,7 +91,7 @@ def setup_logger(): setup_logger() if not (api_key := os.getenv("GDH_API_KEY")): raise ValueError("Set GDH_API_KEY to your Global.health API key") - bucket = os.getenv("BUCKET", "covid-19-aggregates") + bucket = os.getenv("BUCKET", "covid-19-aggregates-eu") s3_endpoint = os.getenv("S3_ENDPOINT") instance = os.getenv("GDH_URL", DEFAULT_INSTANCE) if sources := fetch_sources(api_key, instance):