Ce 2642 adding s3 bucket for mwaa etl scripts#2064
Conversation
|
There was a problem hiding this comment.
@LBH-wgreeff Looks good. Ready to approve, but I suggest renaming mwaa_etl_scripts_bucket to ecs_etl_scripts_bucket. Since most of the ETL scripts are in the airflow/dags folder, Only the etl scripts requiring ECS compute resources should be placed here. Using mwaa_etl_scripts_bucket might confuse users, making them unsure where to store their ETL scripts.
Additionally, I will rename the etl_scripts folder in the dap-airflow repo to ecs_etl_scripts shortly. Apologies for the inconvenience.
|
I'm inclined to go a step further and say the name should be service agnostic. The logical grouping is the task / workflow not the service the script is going to run on in my mind. |
|
Hey Tim, do you mean keeping the name of the |
|
I meant the bucket name - I can quite easily envisage a world where either:
I think, to your point about not confusing users by how things are named, it's much easier to have a series of scripts that are in a structure similar to:
Rather than having to sync them all to different buckets. This is particularly true while people are manually uploading files to stg for testing. |
|
@timburke-hackit sounds good. What you’ve said makes sense to me |
|
Just to add, the Python script, you can specify the bucket and an optional object within the bucket, aka folder in the bucket. |

Adding new s3 bucket for ETL Scripts