https://dagster.io/ logo
n

Noah K

10/12/2020, 1:45 AM
Something like that?
f

Frederic Abiven

10/12/2020, 1:46 AM
Copy code
@pipeline(
    mode_defs=[
        ModeDefinition(name="local_dev", resource_defs={"database": mysql_database}),
        ModeDefinition(name="unittest", resource_defs={"database": csv_file_path})
    ],
)
n

Noah K

10/12/2020, 1:48 AM
This isn't something that needs environment-level replacement though
It's just a general purpose parameter
Hence asking about the solid config approach
f

Frederic Abiven

10/12/2020, 1:49 AM
I also thought it would be great having pipeline parameters but I think that the key concept of inputing data to the pipeline is based on the @resource component
n

Noah K

10/12/2020, 1:50 AM
I mean this is not optional, the whole point of this is to be parametereized
That seems like a very common use case
I have N files, they need processing, launch N pipeline runs, one for each.
f

Frederic Abiven

10/12/2020, 1:55 AM
Your resource could be a list of file paths. Then you merge all the data as one dataset and process it. Would that work?
n

Noah K

10/12/2020, 1:55 AM
No, again the whole point of this is one pipeline run per file, because there's an unknown number of them and they have to be updated at different times
Is this not a very common use case?
Pretty standard ETL processing.
f

Frederic Abiven

10/12/2020, 2:02 AM
Not sure then how this could be accomplish. I’m pretty new to Dagster. Sorry
n

Noah K

10/12/2020, 2:03 AM
I mean I laid out 3 different approaches, just trying to work out what the usual one is for the Dagster community.