diff --git a/extract/s3_extract/push_proxy_job.py b/extract/s3_extract/push_proxy_job.py deleted file mode 100755 index 8847214ef..000000000 --- a/extract/s3_extract/push_proxy_job.py +++ /dev/null @@ -1,12 +0,0 @@ -import argparse - -from extract.s3_extract.stage_import import push_proxy_import - -parser = argparse.ArgumentParser() -parser.add_argument("log_type", help="Type to import, can be US, TEST, or DE") -parser.add_argument("date", help="Date to execute import for") - -if __name__ == "__main__": - args = parser.parse_args() - - push_proxy_import(args.log_type, args.date) diff --git a/extract/s3_extract/stage_import.py b/extract/s3_extract/stage_import.py index ce9df5ad6..f5ee4c4f8 100644 --- a/extract/s3_extract/stage_import.py +++ b/extract/s3_extract/stage_import.py @@ -54,28 +54,6 @@ def diagnostics_import(import_date): ) -def push_proxy_import(log_type, import_date): - """ - Function to load data from a previously set up Snowflake stage - for push proxy data from AWS ELB - - @log_type: Str with valid values of US, TEST, or DE - @import_date: Date with format "%Y/%m/%d" - """ - loc = PUSH_PROXY_LOCATIONS[log_type] - aws_account_id = os.getenv('AWS_ACCOUNT_ID') - az = loc['az'] - - extract_from_stage( - loc['table'], - loc['stage'], - 'push_proxy', - get_path(aws_account_id, az), - get_push_proxy_pattern(import_date), - os.environ.copy(), - ) - - def get_push_proxy_pattern(import_date): date = import_date.replace('/', '\\/') return f".*{date}\\/.*"