Divyansh Tripathi
08/13/2023, 7:50 PMDagsterExecutionInterruptedError
error. This frequency has increased many folds the moment we switched to Jobs.
• The RunRetry exception when thrown after this happens [have written a wrapper function which throws this exception], Is not honoured and the run fails completely.
Is there any different logic for as to how retries are handled in job-solids vs pipeline-solidsprha
08/14/2023, 5:51 PM0.14.14
? Or are you testing a newer version of dagster? Do you have a run launcher configured, or are you using the default?
Are the newer runs being executed with the multiprocess executor? I believe that with jobs, the default executor switched to the multi-process executor, rather than the in-process executor (the default with pipelines). When using this with the default run launcher, you might be aggressively spawning new processes, causing some sort of termination event to be sent. To test if this is the issue, you can override your executor in your job, to use the in_process_executor.
@job(executor_def=in_process_executor)
def my_job():
...
To maintain the process isolation (and increase op parallelization), you keep using the multiprocess exeutor, but configure your jobs to limit the number of concurrent ops executing.