Hey, I’m starting to setup Dagster in ECS for our team so we can hopefully migrate from airflow, but I’m unsure on how to proceed in fetching updates from different repositories, here’s what I roughly planned:
ECS Cluster running a service with dagit, daemon, and a gRPC for each part of our user code as a standalone repository (one for our internal ETL tool, one for the transformation part, etc). With the base scripts for them, as well as schedules/jobs/etc defined within each of these repos. Now, as far as I can tell this would require us to restart the service so the latest version of the scripts is fetched every time whichever one of these components is fetched. This seems to me like a bit of an overkill and I’m assuming I’m overlooking a different way to do it a little more cleanly, but couldn’t find any exact information as I scoured the web. Any ideas of how best to do it/how your teams implemented it?
This exact problem was a non-issue when running Airflow as we’re outsourcing all the computing to separate ECS tasks/lambdas, so we can deploy the code independently of airflow and it would just execute the lambda on schedule.