jorge.arada
08/18/2021, 5:08 PMChris Chan
08/18/2021, 8:00 PMexecute_solid
- I tried to pass in a database password like 'password': {'env': 'DB_PASSWORD'}
but it seems like that isn’t parsed properly - when I inspect context.resource_config
within the resource it just returns the literal dictionary. Is this expected behavior? Should I just use os.environ
? it’s confusing since it works for solid configsDevaraj Nadiger
08/19/2021, 10:17 AMget_s3_keys
api has only implementation for last_run_key
not for last_completion_time
, Am I right?Ming Fang
08/19/2021, 1:08 PM@asset_sensor(asset_key=AssetKey(["sftp", "*"]), pipeline_name="process_file_pipeline")
Simon Späti
08/19/2021, 2:29 PMinput_values
or run_config
parameter for my solid inputs? I wasn't able to make it work. What finally worked if I directly submitted it to the solid itself (see example in thread).Dylan Hunt
08/19/2021, 3:09 PMWilliam Reed
08/19/2021, 4:55 PMDonny Winston
08/19/2021, 5:28 PM@asset_sensor
sensor triggers to run again? I don’t want to have to delete/wipe the asset in dagit and then have to run a prior workflow to regenerate the asset. I want to kick off the sensor rather than the graph/pipeline that it yields a RunRequest
for because the sensor uses asset_event metadata to set the run_config for the RunRequest.Donny Winston
08/19/2021, 5:49 PMJonathan Almanza Camacho
08/19/2021, 6:21 PMNAME READY STATUS RESTARTS AGE
dagster-daemon-f8656bc5c-4sflz 0/1 CrashLoopBackOff 1 9s
dagster-dagit-6b85ddd774-qbxc8 0/1 CrashLoopBackOff 1 9s
dagster-dagster-user-deployments-k8s-example-user-code-1-65wtgz 0/1 CrashLoopBackOff 1 9s
dagster-postgresql-0 1/1 Running 0 84s
And cannot access to the dagster environment.
When I try to deploy without kubernets I can expose it into the port :3000
I'm not sure what is the problem and I can't found something on the web to solve it...sumanta baral
08/19/2021, 8:31 PMjdigs
08/20/2021, 3:28 AMTypeError: unsupported operand type(s) for -: 'tuple' and 'int'
...
File "C:\Users\...\python\drc-fire\fire_module.py", line 53, in prepare_script1
start = self.analysisYear - self.baselineLength
analysisyear is defined as an int in config and is an int, and baselineLength is an int. The script/module runs on it own without error and nothing is in a tuplejdigs
08/20/2021, 3:32 AMself.baselineLength = 3
module
start = self.analysisYear - self.baselineLength
dagster
@solid(config_schema={"year":int,
"region":str,
"cover":str,
"covername":str,
"test":bool})
def run_prep_step_1(context):
year=context.solid_config['year'],
region=ee.FeatureCollection(context.solid_config['region']),
cover=ee.Image(context.solid_config['cover']),
covername=context.solid_config['covername']
a = step1(year,region,cover,covername)
prep_lc = a.prepare_script1()
jdigs
08/20/2021, 4:00 AMSaurav Mittal
08/20/2021, 7:18 AMGreg
08/20/2021, 2:26 PMNathan Saccon
08/20/2021, 3:20 PMXiaosu
08/20/2021, 3:26 PMchrispc
08/20/2021, 4:16 PMAmeen
08/20/2021, 4:20 PMError1: dagster.core.errors.ScheduleExecutionError: Error occurred during the execution of run_config_fn for schedule data_fetch_schedule
The above exception was caused by the following exception:
Error2: TypeError: data_fetch_schedule() takes 0 positional arguments but 1 was given
edit: typoGabriel Milan
08/20/2021, 5:05 PMCarlos Sanoja
08/20/2021, 7:40 PMsean
08/20/2021, 8:29 PMSolidDefinition
, but is there any kind of official API (or one planned?). Thanks.Xu Zhang
08/21/2021, 4:19 PMMikee Jazmines
08/22/2021, 1:13 PMModuleNotFoundError: No module named 'dagster_cron'
dagster.check.CheckError: Failure condition: Couldn't import module dagster_cron.cron_scheduler when attempting to load the configurable class dagster_cron.cron_scheduler.SystemCronScheduler
I tried running it through dagit and dagster pipeline and creating a new virtual environment. I use python 3.7.7. Would love to have any help
Thanks!ismail balaban
08/22/2021, 6:24 PM@solid(
output_defs=[DynamicOutputDefinition(name="num1", dagster_type=int)]
)
def number_generator(_):
for i in range(10):
yield DynamicOutput(output_name="num1", value=i, mapping_key=f"deneme_{i}_1")
@solid(
output_defs=[OutputDefinition(name="power", dagster_type=int)]
)
def return_power(_):
return 2
@solid(
input_defs=[InputDefinition(name="num", dagster_type=int),
InputDefinition(name="power", dagster_type=int)],
output_defs=[OutputDefinition(name="num_power", dagster_type=int)]
)
def num_powers(_, num, power):
return num ** power
@solid()
def print_num(_, num):
print(num)
@pipeline()
def dynamic_pipeline():
numbers = number_generator()
power = return_power()
num_power = numbers.map(lambda num: num_powers(num, power)).collect()
print_num(num_power)
Sample-2
@solid(
output_defs=[DynamicOutputDefinition(name="num1", dagster_type=int),
DynamicOutputDefinition(name="num2", dagster_type=int)]
)
def number_generator(_):
for i in range(10):
yield DynamicOutput(output_name="num1", value=i, mapping_key=f"deneme_{i}_1")
yield DynamicOutput(output_name="num2", value=i, mapping_key=f"deneme_{i}_2")
@solid(
output_defs=[OutputDefinition(name="power", dagster_type=int)]
)
def return_power(_):
return 2
@solid(
input_defs=[InputDefinition(name="num", dagster_type=int),
InputDefinition(name="power", dagster_type=int)],
output_defs=[OutputDefinition(name="num_power", dagster_type=int)]
)
def num_powers(_, num, power):
return num ** power
@solid()
def print_num(_, num):
print(num)
@pipeline()
def dynamic_pipeline():
numbers, _ = number_generator()
power = return_power()
num_power = numbers.map(lambda num: num_powers(num, power)).collect()
print_num(num_power)
Dylan Hunt
08/23/2021, 7:16 AMNilesh Pandey
08/23/2021, 7:57 AMAndrew Parsons
08/23/2021, 3:30 PMconfig_schema
a list of one or more Shapes?
I have Dagster integrated with a Django application (although I'm really only using the Django ORM).
I need to perform a simple Django ORM query using one or more Django ORM filters.
I'm thinking:
Shape(
fields={
'operator': Selector( ... ), # and vs. or
'tags': Field(config=list, ...),
'slice_begin': Field(config=int, ...),
'slice_end': Field(config=int, ...),
},
)
But then... can I define a list of Shapes?Chris Evans
08/23/2021, 3:46 PM"{{ ds }}"
to get the execution timestamp of a DAG. The only way I could come up w/ to something similar was with the below code which requires run storage. I’m curious if there is a simpler way to obtain this information?
from dagster import pipeline, repository, schedule, solid
from dagster.core.storage.pipeline_run import PipelineRunsFilter
@solid
def hello(context):
filter = PipelineRunsFilter(run_ids=[context.run_id])
run_record = context.get_step_execution_context().instance.run_storage.get_run_records(filters=filter)