Tobias Macey
09/23/2020, 2:48 PMTobias Macey
09/23/2020, 2:53 PMKevin
09/23/2020, 5:21 PMdagit -f dagstermill_nb_pipeline.py
runs beautifully; it imports my other .py files and no problems all the pipelines from the repository run.
I take this same file and use it with my k8s deployment with grpc and the dagit pod always fails complaining that my .py files weren't found as modules...
I've attached the k8s log associated to the failed dagit podWali
09/23/2020, 6:43 PMuser
09/24/2020, 11:27 PMuser
09/24/2020, 11:49 PMrex
09/25/2020, 12:03 AMdhume
09/25/2020, 2:30 PMszalai1
09/25/2020, 3:55 PMModuleNotFoundError: No module named 'celery.backends.amqp'
I'm getting this error on all of our pipelines. Full error in thread.
(let me know if this should go to Discussions
)Matt Webster
09/26/2020, 7:20 PMRichard Brady
09/28/2020, 9:05 AMTobias Macey
09/28/2020, 7:11 PMmatas
09/29/2020, 9:48 AMVenkkataraman R
09/29/2020, 12:57 PMVenkkataraman R
09/29/2020, 12:58 PMTobias Macey
09/29/2020, 1:02 PMSergii Ivakhno
09/29/2020, 1:39 PM"storage": {"filesystem": {"config": {"base_dir": '/dagster'}}}
, in the run_config. However I get surprising error Undefined field "storage" at the root. Expected:
. Wonder if the storage has been overwritten somewhere? I also set mode as follows local_mode = ModeDefinition(
name='local',
resource_defs={
's3': s3_resource,
'tempfile': tempfile_resource,
'file_cache': fs_file_cache,
'file_manager': local_file_manager,
},
executor_defs=default_executors + [dask_executor],
intermediate_storage_defs=s3_plus_default_intermediate_storage_defs,
system_storage_defs=s3_plus_default_storage_defs,
logger_defs={'local_logger': json_file_logger}
Dagster version 0.9.9. Thanks for suggestions in advance!Rodrigo Pasquale
09/29/2020, 3:01 PMRodrigo Pasquale
09/29/2020, 5:50 PMTobias Macey
09/29/2020, 7:59 PMRodrigo Pasquale
09/30/2020, 12:51 PMSergii Ivakhno
09/30/2020, 4:25 PMdagster.yml
with PostgresRunStorage
as run_storage:
module: dagster_postgres.run_storage
class: PostgresRunStorage
config:
postgres_db:
username:
env: DAGSTER_PG_USERNAME
password:
env: DAGSTER_PG_PASSWORD
hostname:
env: DAGSTER_PG_HOST
db_name:
env: DAGSTER_PG_DB
port: 5432
and similar for events using PostgresEventLogStorage
. I then execute pipeline with dask executor - it runs successfully but gives quite a few errors qlite3.ProgrammingError: SQLite objects created in a thread can only be used in that same thread. The object was created in thread id 123145510014976 and this is thread id 123145494249472.
I am curious why SQLite
is mentioned at all since I use Postgres
for run storage? Thanks for clarifications!Xinbin Huang
09/30/2020, 6:55 PMdhume
09/30/2020, 7:18 PMuser
10/01/2020, 8:09 PMdaniel
10/01/2020, 8:18 PMschrockn
10/01/2020, 8:20 PMAlessandro Marrella
10/02/2020, 4:59 PMdagit
locally connected to a remote Kubernetes cluster with Dagster installed. I'm trying to use the K8sRunLauncher.
Dagit starts fine and I can see my pipeline, but when i trigger it and check the pod logs I see this error:
dagster.check.ParameterCheckError: Param "pipeline_run" is not a PipelineRun. Got None which is type <class 'NoneType'>.
the args of the pod look like this:
api
execute_run_with_structured_logs
{"__class__": "ExecuteRunArgs", "instance_ref": null, "pipeline_origin": {"__class__": "PipelinePythonOrigin", "pipeline_name": "pipeline", "repository_origin": {"__class__": "RepositoryPythonOrigin", "code_pointer": {"__class__": "FileCodePointer", "fn_name": "pipeline", "python_file": "/Users/[redacted]/Projects/ml-platform/dsdk/pipeline.py", "working_directory": null}, "executable_path": "/Users/[redacted]/Projects/ml-platform/dsdk/.venv/bin/python"}}, "pipeline_run_id": "9a037a94-c7cc-4fd0-a737-ab65d46b5ab4"}
The local paths look suspicious but are they the cause for the error?
This is my launcher config but I'm not sure if it's relevant
run_launcher:
module: dagster_k8s.launcher
class: K8sRunLauncher
config:
dagster_home:
env: DAGSTER_HOME
instance_config_map: dagster-instance
service_account_name: dagster
postgres_password_secret: dagster-postgresql-secret
job_image: [REDACTED].<http://dkr.ecr.us-east-1.amazonaws.com/dsdk:dagster|dkr.ecr.us-east-1.amazonaws.com/dsdk:dagster>
job_namespace: datascience
load_incluster_config: false
image_pull_policy: Always
can anyone help me?
Thanks in advance! And thanks again for this great project 🙂Alessandro Marrella
10/02/2020, 5:01 PMuser
10/02/2020, 8:15 PM