martin o leary
09/06/2023, 9:37 AMDockerRunLauncher
and using 2 of my own user code images.
Both are running the gRPC servers fine so I can see the code locations in dagit however one of the images will execute runs fine and the other won't.
The issue is source code related but I have no idea how to access the logs from inside that run instance container.
The logging for the run shows that the we get far enough that the docker launch happens:
[DockerRunLauncher] Launching run in a new container cb20571c08146c030e69fcf058d3de6ca4a22edaffa7cc6c0f641c3c7d963b41 with image <http://ghcr.io/mycompany/my_second_user_code_image:latest|ghcr.io/mycompany/my_second_user_code_image:latest>
Nothing happens in the UI and I need to cancel the job.
I don't see the container running on the host so it obviously fails and gets removed.
I have set the run storage, schedule storage and event log storage to save to postgres and the compute logs go to S3 but I can't seem to figure out where to find the logs from inside that launched container before it exits.
So:
1. What config can i set on the run_launcher so that I can keep the container around after failure and inspect the logs?
2. Where should those logs end up based on my storage setup?martin o leary
09/06/2023, 10:10 AMrun_launcher:
module: dagster_docker
class: DockerRunLauncher
config:
container_kwargs:
auto_remove: false
claire
09/06/2023, 6:50 PMLocalComputeLogManager
they will be persisted to disk.martin o leary
09/06/2023, 8:49 PMclaire
09/06/2023, 9:15 PMmartin o leary
09/07/2023, 6:31 AMJoe Van Drunen
09/07/2023, 5:57 PMS3ComputeLogManager
will only upload on termination, there might have been something preventing it being marked as failed/successful and trigger the log file upload.
You can configure an upload interval if you'd like
compute_logs:
module: dagster_aws.s3.compute_log_manager
class: S3ComputeLogManager
config:
upload_interval: 30
martin o leary
09/07/2023, 5:59 PM