Charles Lariviere
03/30/2023, 2:31 PMdaniel
03/30/2023, 2:32 PMCharles Lariviere
03/30/2023, 3:27 PMdoes this mean within a single run, submitting some ops to one place and other ops to another?Yes -- op/asset-level granularity would be ideal, but I understand that might be more complex to implement and could workaround job-level granularity. One example where this would be useful is where I have a job that runs a training pipeline for an ML model, from feature computation (i.e. k8s), to training (i.e. k8s/ec2 with GPUs), and evaluation (i.e. on-prem with custom inference hardware). I could very well break this out into two different jobs, but it feels like with cloud IO managers, the execution of each asset/op in a job doesn't need to be tied to a single machine or even cluster.
daniel
03/30/2023, 3:29 PMCharles Lariviere
03/30/2023, 4:09 PMWhat type of agent were you imagining the on-prem server would be running?Thinking it would be running a Docker agent
We've definitely talked about ways to intersperse different ops within a single run to run in different places (some ops running as k8s pods, others in local subprocesses, maybe others as an ECS task, etc. - all communicating together through the event log as part of the same run).That sounds great! Curious if you know whether there's an open issue I could track for the status of that?
daniel
03/30/2023, 5:31 PMCharles Lariviere
03/30/2023, 5:42 PMdocker_executor
?daniel
03/30/2023, 5:42 PMCharles Lariviere
03/30/2023, 5:43 PM