George Pearse
07/21/2021, 8:11 AMLinh Nguyen
07/21/2021, 8:16 AMFrancois-DE
07/21/2021, 8:59 AMdagster.serdes.ipc.DagsterIPCProtocolError: Process exited with return code -9 while waiting for events
.
Would anyone be able to explain what this means?Xu Zhang
07/21/2021, 4:06 PM- grpc_server:
host: the.domain.to.the.repo
port: 4266
location_name: "my_grpc_server"
chrispc
07/21/2021, 4:55 PMTraceback (most recent call last):
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\poll_compute_logs.py", line 61, in <module>
execute_polling(sys.argv[1:])
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\poll_compute_logs.py", line 53, in execute_polling
with open(ipc_output_file, "w"):
FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\s4957336\\AppData\\Local\\Temp\\1\\tmpo_zr47w6\\execute-windows-tail-15bc135d998546caa23116bbff00ceb1'
2021-06-14 02:33:51 - dagster - ERROR - Mexico_CTMS_Corp - a1cb3fdf-19ef-4676-a6b0-b3c7610ec0f9 - 22036 - PIPELINE_FAILURE - Execution of pipeline "Mexico_CTMS_Corp" failed. An exception was thrown during execution.
Exception: Timed out waiting for tail process to start
Stack Trace:
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\api.py", line 762, in pipeline_execution_iterator
for event in pipeline_context.executor.execute(pipeline_context, execution_plan):
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\executor\in_process.py", line 38, in execute
yield from iter(
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\api.py", line 841, in __iter__
yield from self.iterator(
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\plan\execute_plan.py", line 72, in inner_plan_execution_iterator
active_execution.verify_complete(pipeline_context, step.key)
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\contextlib.py", line 120, in __exit__
next(self.gen)
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\storage\compute_log_manager.py", line 56, in watch
yield
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\contextlib.py", line 120, in __exit__
next(self.gen)
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\storage\local_compute_log_manager.py", line 51, in _watch_logs
yield
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\contextlib.py", line 120, in __exit__
next(self.gen)
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\compute_logs.py", line 31, in mirror_stream_to_file
yield pids
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\contextlib.py", line 120, in __exit__
next(self.gen)
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\compute_logs.py", line 75, in tail_to_stream
yield pids
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\contextlib.py", line 120, in __exit__
next(self.gen)
File "C:\Users\s4957336\Anaconda3\envs\borrar\lib\site-packages\dagster\core\execution\compute_logs.py", line 104, in execute_windows_tail
raise Exception("Timed out waiting for tail process to start")
I am not sure why dagster is looking into this path FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\s4957336\\AppData\\Local\\Temp\\1\\tmpo_zr47w6\\execute-windows-tail-15bc135d998546caa23116bbff00ceb1'
what could I do? I am using windows 10, 64 bit. Conda enviroment. Dagster 0.12.2 version. I got a timeout several times randomnly. I configured the backend in a postgres Instance in dagster.yml file.Josh Lloyd
07/21/2021, 6:27 PM@pipeline(
mode_defs=[
mode_def_prod,
mode_def_prod_coll,
],
)
def collective_pipeline():
...
@schedule(
pipeline_name="collective_pipeline",
cron_schedule="0 7 * * *",
execution_timezone="UTC",
mode='prod_coll'
)
def coll_daily_schedule():
return {"resources": {"io_manager": {"config": {"s3_bucket": "dagster-wdt", "s3_prefix": "stage/io"}}}}
def test_coll_daily_schedule():
run_config = coll_daily_schedule()
assert validate_run_config(collective_pipeline, run_config)
error when I use pytest
in commandline:
dagster.core.errors.DagsterInvalidConfigError: Error in config for pipeline collective_pipeline
Error 1: Received unexpected config entries "['s3_bucket', 's3_prefix']" at path root:resources:io_manager:config. Expected: "['base_dir']."
One strange solution I have found is to switch the order of the mode definitions in the pipeline so that mode_def_prod_coll
comes first. Then the test passes. But in my mind, this shouldn’t happen because I’m specifically calling out the name of the mode I want to use in the decorator of the schedule prod_coll
Hebo Yang
07/21/2021, 6:28 PMchrispc
07/21/2021, 6:43 PMGeorge Pearse
07/21/2021, 7:50 PMKirk Stennett
07/21/2021, 11:50 PMNo module named dagster.grpc
I installed
dagster==0.12.2 dagit==0.12.2 dagster-graphql==0.12.2
Running python repo.py
works fine but for some reason dagit can't load it. Any ideas why?Gary Kennedy
07/22/2021, 5:32 AMThanabodee Charoenpiriyakij
07/22/2021, 9:53 AMGabe Calvo
07/22/2021, 12:37 PMrunLauncher:
type: CeleryK8sRunLauncher
config:
celeryK8sRunLauncher:
configSource:
redis_password:
valueFrom:
secretKeyRef:
name: wot-dagster-redis
key: redis-password
redis:
enabled: true
internal: true
host: "wot-dagster-redis-master"
by using the redis>interal: true the wot-dagster-redis gets generated automatically as far as I understand and the "redis-password" is set inside.
I'm then trying to give that password to the celeryK8sRunLauncher as otherwise the celery pods return:
[2021-07-22 12:40:27,366: ERROR/MainProcess] consumer: Cannot connect to <redis://wot-dagster-redis-master:6379/0>: Authentication required..
Any help welcome 🙂
GabeWes Roach
07/22/2021, 4:00 PMJosh Lloyd
07/22/2021, 4:07 PMOperation name: RunsRootQuery
Message: list index out of range
Path: ["pipelineRunsOrError","results",2,"canTerminate"]
Locations: [{"line":36,"column":3}]
Stack Trace:
File "/usr/local/lib/python3.7/site-packages/graphql/execution/executor.py", line 452, in resolve_or_error
return executor.execute(resolve_fn, source, info, **args)
File "/usr/local/lib/python3.7/site-packages/graphql/execution/executors/sync.py", line 16, in execute
return fn(*args, **kwargs)
File "/usr/local/lib/python3.7/site-packages/dagster_graphql/schema/pipelines/pipeline.py", line 274, in resolve_canTerminate
return graphene_info.context.instance.run_coordinator.can_cancel_run(self.run_id)
File "/usr/local/lib/python3.7/site-packages/dagster/core/run_coordinator/queued_run_coordinator.py", line 105, in can_cancel_run
return self._instance.run_launcher.can_terminate(run_id)
File "/usr/local/lib/python3.7/site-packages/dagster_aws/ecs/launcher.py", line 138, in can_terminate
status = self.ecs.describe_tasks(tasks=[arn], cluster=cluster)["tasks"][0]["lastStatus"]
I am running dagster on the experimental EcsRunLauncher in AWS ECS Fargate. Version 0.12.1. Prior to this error, I terminated a couple of long-running runs that were in the middle of execution but had errored on previous steps.pdpark
07/22/2021, 5:45 PMmapping_key
from context. Getting error:
def get_mapping_key(self) -> Optional[str]:
"""
Which mapping_key this execution is for if downstream of a DynamicOutput, otherwise None.
"""
> return self._step_execution_context.step.get_mapping_key()
E AttributeError: _step_execution_context
Test code is pretty simple:
context = build_solid_context(
config={...})
data = ...
export_data(context, data)
Hebo Yang
07/22/2021, 6:12 PMKirk Stennett
07/22/2021, 11:53 PMdef pipeline():
test = generator() # generates dynamic output
other_solid_output = solid()
output1, output2 = test.map(process1).map(process2)
more_output = new_solid(output1, other_solid_output)
last_process(output2, more_output)
I get the error that Dynamic Output must be mapped or collected. I tried using map with these too but that just returns syntax error. Is there any feasible way to make this work? Without changing the logic of the current solidsArun Kumar
07/23/2021, 12:48 AMThanabodee Charoenpiriyakij
07/23/2021, 8:49 AMNicolas Gaillard
07/23/2021, 9:38 AMdagster instance migrate
. It seems that it can't reindex assets for some reasons. Do you have any idea to fix this ?
root@02e3ef89f539:/home/src/dagster# dagster instance migrate
$DAGSTER_HOME: /home/src/dagster
Updating run storage...
INFO [alembic.runtime.migration] Context impl PostgresqlImpl.
INFO [alembic.runtime.migration] Will assume transactional DDL.
INFO [alembic.runtime.migration] Running upgrade 7cba9eeaaf1d -> ddcc6d7244c6, add column mode
INFO [alembic.runtime.migration] Running upgrade ddcc6d7244c6 -> 7b8304b4429d, extract asset_keys index columns
Starting data migration: add_mode_column
Querying run storage.
0%| | 0/36 [00:00<?, ?it/s]100%|████████████████████████████████████████████████████████████████████████████████████████| 36/36 [00:00<00:00, 445.62it/s]
Finished data migration: add_mode_column
Updating event storage...
INFO [alembic.runtime.migration] Context impl PostgresqlImpl.
INFO [alembic.runtime.migration] Will assume transactional DDL.
Starting reindex: asset_key_index_columns
Querying asset keys.
Found 94 assets to reindex.
0%| | 0/94 [00:00<?, ?it/s]
0%| | 0/94 [00:00<?, ?it/s]
Traceback (most recent call last):
File "/usr/local/bin/dagster", line 8, in <module>
sys.exit(main())
File "/usr/local/lib/python3.7/site-packages/dagster/cli/__init__.py", line 48, in main
cli(auto_envvar_prefix=ENV_PREFIX) # pylint:disable=E1123
File "/usr/local/lib/python3.7/site-packages/click/core.py", line 1137, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.7/site-packages/click/core.py", line 1062, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.7/site-packages/click/core.py", line 1668, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.7/site-packages/click/core.py", line 1668, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.7/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.7/site-packages/click/core.py", line 763, in invoke
return __callback(*args, **kwargs)
File "/usr/local/lib/python3.7/site-packages/dagster/cli/instance.py", line 46, in migrate_command
instance.upgrade(click.echo)
File "/usr/local/lib/python3.7/site-packages/dagster/core/instance/__init__.py", line 557, in upgrade
self._event_storage.reindex_assets(print_fn=print_fn)
File "/usr/local/lib/python3.7/site-packages/dagster/core/storage/event_log/sql_event_log.py", line 434, in reindex_assets
self._apply_migration(migration_name, migration_fn, print_fn, force)
File "/usr/local/lib/python3.7/site-packages/dagster/core/storage/event_log/sql_event_log.py", line 421, in _apply_migration
migration_fn()(self, print_fn)
File "/usr/local/lib/python3.7/site-packages/dagster/core/storage/event_log/migration.py", line 133, in migrate_asset_keys_index_columns
tags = event.dagster_event.step_materialization_data.materialization.tags
AttributeError: 'NoneType' object has no attribute 'dagster_event'
Many thanks and have a nice weekend !
NicolasGabe Calvo
07/23/2021, 11:02 AMimagePullSecrets:
- name: prod-wot-registry-secrets
dagster-user-deployments:
enabled: true
imagePullSecrets:
- name: prod-wot-registry-secrets
deployments:
- name: "wot-dagster-repository"
image:
repository: "<path-to-my-repo-on-the-registry>"
tag: latest
pullPolicy: Always
dagsterApiGrpcArgs:
- "-f"
- "/app/repository/repository.py"
port: 3030
envSecrets:
- name: prod-wot-storage-secrets
runLauncher:
type: CeleryK8sRunLauncher
redis:
enabled: true
internal: true
host: "wot-dagster-redis-master"
usePassword: false
flower:
enabled: true
dagsterDaemon:
runCoordinator:
enabled: true
and get the following error:
dagster.core.errors.DagsterInvalidConfigError: Errors whilst loading configuration for {'instance_config_map': Field(, default=@, is_required=True), 'postgres_password_secret': Field(, default=@, is_required=False), 'dagster_home': Field(, default=/opt/dagster/dagster_home, is_required=False), 'load_incluster_config': Field(, default=True, is_required=False), 'kubeconfig_file': Field(, default=None, is_required=False), 'broker': Field(, default=@, is_required=False), 'backend': Field(, default=rpc://, is_required=False), 'include': Field(, default=@, is_required=False), 'config_source': Field(, default=@, is_required=False), 'retries': Field(, default={'enabled': {}}, is_required=False)}.
Error 1: Post processing at path root:instance_config_map of original value {'env': 'DAGSTER_K8S_INSTANCE_CONFIG_MAP'} failed:
dagster.config.errors.PostProcessingError: You have attempted to fetch the environment variable "DAGSTER_K8S_INSTANCE_CONFIG_MAP" which is not set. In order for this execution to succeed it must be set in this environment.
Stack Trace:
File "/usr/local/lib/python3.9/site-packages/dagster/config/post_process.py", line 77, in _post_process
new_value = context.config_type.post_process(config_value)
File "/usr/local/lib/python3.9/site-packages/dagster/config/source.py", line 42, in post_process
return str(_ensure_env_variable(cfg))
File "/usr/local/lib/python3.9/site-packages/dagster/config/source.py", line 16, in _ensure_env_variable
raise PostProcessingError(
Error 2: Post processing at path root:postgres_password_secret of original value {'env': 'DAGSTER_K8S_PG_PASSWORD_SECRET'} failed:
dagster.config.errors.PostProcessingError: You have attempted to fetch the environment variable "DAGSTER_K8S_PG_PASSWORD_SECRET" which is not set. In order for this execution to succeed it must be set in this environment.
Stack Trace:
File "/usr/local/lib/python3.9/site-packages/dagster/config/post_process.py", line 77, in _post_process
new_value = context.config_type.post_process(config_value)
File "/usr/local/lib/python3.9/site-packages/dagster/config/source.py", line 42, in post_process
return str(_ensure_env_variable(cfg))
File "/usr/local/lib/python3.9/site-packages/dagster/config/source.py", line 16, in _ensure_env_variable
raise PostProcessingError(
David
07/23/2021, 4:09 PMXu Zhang
07/23/2021, 6:51 PMos.environ["PYTHONWARNINGS"] = "ignore"
but i still can see warnings from the log
res = conn.execute(run_group_query)
/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_3b298aacc6b7e5bd29d8bd9bb36dfcb367618b6abfa16d967bb355b20119311b/site-packages/graphql/backend/core.py:32: DeprecationWarning: The 'root' alias has been deprecated. Please use 'root_value' instead.
return execute(schema, document_ast, *args, **kwargs)
/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_3b298aacc6b7e5bd29d8bd9bb36dfcb367618b6abfa16d967bb355b20119311b/site-packages/graphql/backend/core.py:32: DeprecationWarning: The 'context' alias has been deprecated. Please use 'context_value' instead.
return execute(schema, document_ast, *args, **kwargs)
/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_3b298aacc6b7e5bd29d8bd9bb36dfcb367618b6abfa16d967bb355b20119311b/site-packages/graphql/backend/core.py:32: DeprecationWarning: The 'root' alias has been deprecated. Please use 'root_value' instead.
return execute(schema, document_ast, *args, **kwargs)
/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_3b298aacc6b7e5bd29d8bd9bb36dfcb367618b6abfa16d967bb355b20119311b/site-packages/graphql/backend/core.py:32: DeprecationWarning: The 'context' alias has been deprecated. Please use 'context_value' instead.
return execute(schema, document_ast, *args, **kwargs)
/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_3b298aacc6b7e5bd29d8bd9bb36dfcb367618b6abfa16d967bb355b20119311b/site-packages/dagster/core/storage/runs/sql_run_storage.py:407: SADeprecationWarning: Executing a subquery object is deprecated and will raise ObjectNotExecutableError in an upcoming release. Please execute the underlying select() statement directly.
res = conn.execute(run_group_query)
Xu Zhang
07/23/2021, 6:51 PMDavid Y
07/23/2021, 7:15 PMtest.py
script on my local machine and want to hit the dagster instance on my remote host, how do I configure that? I'm sure its something simple that I've missed somewhere in the docs. Thanks!takan
07/24/2021, 5:11 AMXu Zhang
07/24/2021, 7:37 PMGabriel Milan
07/24/2021, 7:51 PMRicardo Azevedo
07/25/2021, 6:34 AMmap() collect()
system? I am thinking of a flow like the following (I know this isnt how map works, just trying to get the point across):
dynamic_out = get_values() # returns a dynamicOutput
val = get_some_value() # returns an Output
incrimented = incriment(dynamic_out.map(), val)