VxD
07/22/2022, 5:07 AMmax_concurrent_runs
setting currently being set to 36, we are never seeing more than 20 jobs dispatched concurrently by the celery_executor
with most celery workers just sitting idle.
What could be the bottleneck and how can we diagnose this? We were hoping to raise this setting to several hundreds as we add more hardware to the cluster.
Could it be that the gRPC server is not cutting it? We are hosting one instance for the whole cluster. Shall we be running more?VxD
07/22/2022, 5:27 AMVxD
07/22/2022, 5:37 AMdaniel
07/22/2022, 12:02 PMVxD
07/22/2022, 12:04 PMVxD
07/22/2022, 12:08 PMVxD
07/22/2022, 12:09 PMVxD
07/22/2022, 12:10 PMdaniel
07/22/2022, 12:11 PMdaniel
07/22/2022, 12:11 PMVxD
07/22/2022, 12:11 PMVxD
07/22/2022, 12:12 PMdaniel
07/22/2022, 12:15 PMdaniel
07/22/2022, 12:15 PMVxD
07/22/2022, 12:16 PMdaniel
07/22/2022, 12:17 PMdaniel
07/22/2022, 12:17 PMVxD
07/22/2022, 12:18 PMVxD
07/26/2022, 1:33 AMRunRequests
is to work around the fact that Dagster subgraphs cannot branch into parallel operations (see here).
As a result, we are unable to use dynamic graphs here and have to rely on a sensor to dispatch our subgraphs as parallel jobs. 😞daniel
07/26/2022, 1:35 AMVxD
07/26/2022, 1:37 AMVxD
07/26/2022, 1:38 AMVxD
07/26/2022, 1:40 AMexecute_in_process
or something. 😅