Xinbin Huang
01/26/2021, 6:34 PMFran Sanchez
01/27/2021, 3:00 PM[DockerRunLauncher] Launching run in container five_oracle_pipeline_dagster_oracle_pipelines with ID ceacbfd6707e3cfa09dce15898d1923b3e4edb14de818a9cf39bfc0390a4cfe1
And in the docker-compose output:
dagster_docker_daemon | 2021-01-27 14:56:41 - SchedulerDaemon - INFO - Not checking for any runs since no schedules have been started.
dagster_docker_daemon | 2021-01-27 14:56:41 - SensorDaemon - INFO - Not checking for any runs since no sensors have been started.
dagster_docker_daemon | 2021-01-27 14:56:41 - QueuedRunCoordinatorDaemon - INFO - Retrieved 1 queued runs, checking limits.
dagster_docker_daemon | 2021-01-27 14:56:43 - QueuedRunCoordinatorDaemon - INFO - Launched 1 runs.
dagster_docker_daemon | 2021-01-27 14:56:47 - QueuedRunCoordinatorDaemon - INFO - Poll returned no queued runs.
dagster_docker_daemon | 2021-01-27 14:56:47 - QueuedRunCoordinatorDaemon - INFO - Launched 0 runs.
But nothing actually runs... Any hints what could be happening?Fran Sanchez
01/27/2021, 3:21 PMFran Sanchez
01/27/2021, 5:15 PMtrigger_rule=TriggerRule.ALL_DONE
Blaise Pabon
01/27/2021, 8:41 PMTobias Macey
01/27/2021, 10:01 PMDaniel Kim
01/27/2021, 11:38 PMJosh Taylor
01/28/2021, 11:59 AMdagster pipeline execute
I can do something like this:
dagster pipeline execute --mode prod -c config/prod/resources.yaml -f foo/bar/baz.py
where -c
is the yaml that I want to pass in, which contains information for the pipeline (database configuration, etc).
Is this not a dagster way of doing things? Should each repository/pipeline not have a standard yaml or something? Is it stored elsewhere? Or can you pass it in with dagster daemon?Josh Taylor
01/28/2021, 12:03 PMresources.yaml
is included
edit2: Ah, it is probably presets!Olivier Dupuis
01/28/2021, 6:45 PMPrasad Chalasani
01/28/2021, 7:59 PMdbt
to manage a workflow consisting of transforming a dataset from bigquery, then running ML on it, and putting results back in bigquery. I know this was asked before, but one question — when re-running a pipeline containing heavy computations, are there mechanisms for avoiding re-computation of parts whose inputs haven’t changed? Let’s say the solids’ inputs [outputs] are results read from [dumped to] files Essentially some type of caching/invalidation mechanism.Prasad Chalasani
01/28/2021, 8:17 PMuser
01/29/2021, 1:21 AMsid
01/29/2021, 1:30 AMmrdavidlaing
01/29/2021, 10:21 AMCameron Marquis
01/29/2021, 4:02 PMdagster.core.errors.DagsterInvalidConfigError: Error in config for pipeline recommender_pipeline
Error 1: Invalid scalar at path root:storage:s3:config:s3_bucket. Value "{'env': 'STORAGE_BUCKET'}" of type "<class 'dict'>" is not valid for expected type "String".
make: *** [run] Error 1
with our yaml:
storage:
s3:
config:
s3_bucket:
env: STORAGE_BUCKET
s3_prefix: temp
loggers:
console:
config:
log_level: DEBUG
I can exec into the container and see that the STORAGE_BUCKET
env var is set correctly.
If the answer is we just need to update to the latest version of Dagster, then that's what we can do but ideally we would like to know if there is an easy fix for us in the short-term. Hopefully, I am just missing something very simple due to Friday fried-brain 🙂.Adrian
01/30/2021, 1:33 AMconfig:
s3_bucket: STORAGE_BUCKET
endpoint: MINIO-SERVICE-NAME:PORT
Adrian
01/30/2021, 2:10 AMdagster-dagit-test-connection
always fails (false negative). I don't see a simple way to exclude this podKlaus Stadler
01/30/2021, 10:55 AMKlaus Stadler
01/30/2021, 10:56 AMNoah K
01/30/2021, 10:57 AMNoah K
01/30/2021, 10:57 AMNoah K
01/30/2021, 10:58 AMNoah K
01/30/2021, 10:58 AMNoah K
01/30/2021, 11:00 AMKlaus Stadler
01/30/2021, 11:00 AMKlaus Stadler
01/30/2021, 11:01 AMNoah K
01/30/2021, 11:02 AMNoah K
01/30/2021, 11:02 AMKlaus Stadler
01/30/2021, 11:04 AM