Xu Zhang
07/28/2021, 5:55 AMApplication failed to start after 60 seconds, output: /export/apps/python/3.7/lib/python3.7/importlib/_bootstrap.py:219: RuntimeWarning: greenlet.greenlet size changed, may indicate binary incompatibility. Expected 144 from C header, got 152 from PyObject
return f(*args, **kwds)
/export/apps/python/3.7/lib/python3.7/importlib/_bootstrap.py:219: RuntimeWarning: greenlet.greenlet size changed, may indicate binary incompatibility. Expected 144 from C header, got 152 from PyObject
return f(*args, **kwds)
/export/apps/python/3.7/lib/python3.7/importlib/_bootstrap.py:219: RuntimeWarning: greenlet.greenlet size changed, may indicate binary incompatibility. Expected 144 from C header, got 152 from PyObject
return f(*args, **kwds)
/export/apps/python/3.7/lib/python3.7/importlib/_bootstrap.py:219: RuntimeWarning: greenlet.greenlet size changed, may indicate binary incompatibility. Expected 144 from C header, got 152 from PyObject
return f(*args, **kwds)
/export/apps/python/3.7/lib/python3.7/importlib/_bootstrap.py:219: RuntimeWarning: greenlet.greenlet size changed, may indicate binary incompatibility. Expected 144 from C header, got 152 from PyObject
return f(*args, **kwds)
Xu Zhang
07/28/2021, 5:55 AMEmily
07/28/2021, 1:01 PMmodels
only one time. but in reality, when we try running our pipeline this way we get errors on certain steps (like run-operation
) since macros dont take models
as an argument. so in practice, i had to take models
out of the config here and pass it separately in the run
and test
solids. it would be awesome (if possible) to be able to pass models
in the resource config so we only have to set it once, but only have it pass into the dbt commands where its expected as an input (like run
and test
but not run-operation
)
"resources": {
"dbt": {
"config": {
"target": DBT_TARGET,
"profiles_dir": DBT_DIR,
"project_dir": DBT_DIR,
"models": ["*"],
"ignore_handled_error": True,
}
},
},
Deveshi
07/28/2021, 2:50 PMWilliam Reed
07/28/2021, 5:18 PMXu Zhang
07/28/2021, 6:03 PM[Errno 30] Read-only file system: '/export/content/lid/apps/dagster-web'
error when I re-execute a pipeline, but not when I click Launch Execution
in playground page.
dagster.core.errors.DagsterLaunchFailedError: Error during RPC setup for executing run: OSError: [Errno 30] Read-only file system: '/export/content/lid/apps/dagster-web'
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster_graphql/implementation/utils.py", line 26, in _fn
return fn(*args, **kwargs)
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster_graphql/implementation/execution/launch_execution.py", line 11, in launch_pipeline_reexecution
return _launch_pipeline_execution(graphene_info, execution_params, is_reexecuted=True)
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster_graphql/implementation/execution/launch_execution.py", line 50, in _launch_pipeline_execution
run = do_launch(graphene_info, execution_params, is_reexecuted)
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster_graphql/implementation/execution/launch_execution.py", line 38, in do_launch
workspace=graphene_info.context,
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster/core/instance/__init__.py", line 1350, in submit_run
SubmitRunContext(run, workspace=workspace)
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster/core/run_coordinator/default_run_coordinator.py", line 32, in submit_run
self._instance.launch_run(pipeline_run.run_id, context.workspace)
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster/core/instance/__init__.py", line 1409, in launch_run
self._run_launcher.launch_run(LaunchRunContext(pipeline_run=run, workspace=workspace))
File "/export/content/lid/apps/dagster-web/i001/libexec/dagster-web_e6feb50d76cd5e604738da124e6cd04deae3ce84b1e12b8646074e6845081e63/site-packages/dagster/core/launcher/default_run_launcher.py", line 97, in launch_run
res.message, serializable_error_info=res.serializable_error_info
I think it failed due to the same reason that compute_log_manager
had the other day (and you helped me). I looked into the default_run_coordinator
doc, but there is no base_dir etc. to set..Xu Zhang
07/28/2021, 7:20 PMRe-execute All
button:
I have two pipelines A, B with the same name
but at different locations.
When I trigger the run for B using the Playground, the selector sent back to the Dagster is B, which is expected; however, when I click the Re-execute All
button to re-execute the B pipeline, the selector sent back to Dagster is actually A.Xu Zhang
07/28/2021, 7:33 PMXu Zhang
07/28/2021, 7:38 PMArun Kumar
07/29/2021, 2:32 AMtakan
07/29/2021, 9:24 AMimagePullSecrets:
- name: gcr-secret
under dagster-user-deployments
on values.yaml for dagster chart and on values.yaml for dagster-user-deployment chart but no luck so far.
Also, if I need to put my gcr-secret in values.yaml for dagster chart, why do I need it there? Because I separated my deployment of user-code with dagster-user-deployment chart, I thought dagster chart is not responsible handling user-code-deployments anymore. Please help me out to understand it better 🙂pdpark
07/29/2021, 6:37 PM@repository
def prod_repo():
run_config = get_params()
return [
my_graph.to_job(
config={
"solids": {
"do_stuff": {
"config": {
"a_param": run_config.get("a_param")
}
}
}
}
)
Oyin
07/29/2021, 6:39 PMNavneet Sajwan
07/29/2021, 7:24 PMRob Meng
07/29/2021, 9:07 PMMikhail Knyazev
07/29/2021, 9:17 PMdagster-daemon
need to access repo files to run sensors? I run repos in docker containers via gRPC, and it seems dagster-daemon
doesn't pick up any defined sensors...Jean-Pierre M
07/29/2021, 9:54 PMread_bucket
composite_solid.A[file_1]
composite_solid.A[file_2]
composite_solid.A[file_3]
composite_solid.B[file_1]
composite_solid.B[file_2]
composite_solid.B[file_3]
composite_solid.C[file_1]
composite_solid.C[file_2]
composite_solid.C[file_3]
But I would like for it to be:
read_bucket
composite_solid.A[file_1]
composite_solid.B[file_1]
composite_solid.C[file1]
composite_solid.A[file_2]
composite_solid.B[file_2]
composite_solid.C[file2]
composite_solid.A[file_3]
composite_solid.B[file_3]
composite_solid.C[file3]
I have tried playing around with the dagster/priority
tag in the solid definitions but it doesn't seem to do anything in this case.
Thanks in advance!Arun Kumar
07/29/2021, 11:34 PMprint
statements in the repo.py file, how I don't see those in the Kubernetes user deployment pod logs. Everything looks fine in the my local docker environment. Any thoughts on why this could happen?Gabe Calvo
07/30/2021, 12:45 PMAn unexpected exception was thrown. Please file an issue.
ValueError: too many values to unpack (expected 3)
File "c:\users\gcalvo\appdata\local\pypoetry\cache\virtualenvs\9lnrhjfr-py3.9\lib\site-packages\dagster\core\execution\plan\execute_plan.py", line 193, in _dagster_event_sequence_for_step
for step_event in check.generator(step_events):
File "c:\users\gcalvo\appdata\local\pypoetry\cache\virtualenvs\9lnrhjfr-py3.9\lib\site-packages\dagster\core\execution\plan\execute_step.py", line 326, in core_dagster_event_sequence_for_step
for evt in _type_check_and_store_output(step_context, user_event, input_lineage):
File "c:\users\gcalvo\appdata\local\pypoetry\cache\virtualenvs\9lnrhjfr-py3.9\lib\site-packages\dagster\core\execution\plan\execute_step.py", line 380, in _type_check_and_store_output
for evt in _store_output(step_context, step_output_handle, output, input_lineage):
File "c:\users\gcalvo\appdata\local\pypoetry\cache\virtualenvs\9lnrhjfr-py3.9\lib\site-packages\dagster\core\execution\plan\execute_step.py", line 488, in _store_output
handle_output_res = output_manager.handle_output(output_context, output.value)
File "c:\users\gcalvo\appdata\local\pypoetry\cache\virtualenvs\9lnrhjfr-py3.9\lib\site-packages\dagster_azure\adls2\io_manager.py", line 73, in handle_output
key = self._get_path(context)
File "c:\users\gcalvo\appdata\local\pypoetry\cache\virtualenvs\9lnrhjfr-py3.9\lib\site-packages\dagster_azure\adls2\io_manager.py", line 23, in _get_path
run_id, step_key, output_name = context.get_run_scoped_output_identifier()
The code runs fine if I test it locally (in-memory), but returns this error when using the adls2 io manager.
Am I doing something wrong?
Cheers,
GabeFrancois-DE
07/30/2021, 1:16 PMStack Trace:
File "/media/veracrypt1/sandbox/repos/poc-datalake-pipeline/dlpoc-venv/lib/python3.7/site-packages/dagster/core/host_representation/grpc_server_registry.py", line 181, in _get_grpc_endpoint
fixed_server_id=new_server_id,
File "/media/veracrypt1/sandbox/repos/poc-datalake-pipeline/dlpoc-venv/lib/python3.7/site-packages/dagster/grpc/server.py", line 1080, in __init__
fixed_server_id=fixed_server_id,
File "/media/veracrypt1/sandbox/repos/poc-datalake-pipeline/dlpoc-venv/lib/python3.7/site-packages/dagster/grpc/server.py", line 991, in open_server_process
wait_for_grpc_server(server_process, client, subprocess_args)
File "/media/veracrypt1/sandbox/repos/poc-datalake-pipeline/dlpoc-venv/lib/python3.7/site-packages/dagster/grpc/server.py", line 931, in wait_for_grpc_server
f"Timed out waiting for gRPC server to start with arguments: \"{' '.join(subprocess_args)}\""
It is not a resource issue, as I'm running this locally with an i7 10th gen and 32gig RAM.Alexander Jacob
07/30/2021, 1:51 PMenv_vars
.
Is there a way to define additional environment variables in the configuration of celey-docker
i.e. FOO: bar
?
Cheers, AlexJonathan Mak
07/30/2021, 2:15 PM--full-refresh
when using dagster-dbt?
I tried to do something like
context.resources.dbt.run(models=[m], full_refresh = full_refresh)
And got the run command compiled but I get this error when it actually tries to run
dagster_dbt.errors.DagsterDbtCliFatalRuntimeError: Fatal error in the dbt CLI (return code 2)
Wei Mei
07/30/2021, 2:25 PMszalai1
07/30/2021, 5:40 PMkeyError: 'result'
chrispc
07/30/2021, 11:27 PMMakoto
07/30/2021, 11:44 PMchrispc
07/31/2021, 10:39 PMXu Zhang
08/01/2021, 12:18 AMXu Zhang
08/01/2021, 12:19 AMJonathan Mak
08/01/2021, 4:05 AMexecute(_shell_command_ = cmd, _output_logging_ = 'STREAM', _log_ = context.log, _cwd_ = '/my/dir/path')
with dagster-shell but the temporary script directory was still set somewhere else, am I doing anything wrong there?