Simon Späti
02/10/2021, 3:24 PMDagsterDataType
from a PysparkDataFrame
. I have a generic solid load_delta_table_to_df
, but in my Pipeline I'd like to type-check that the returned DataFrame has certain columns (not always the same see example attached). I try to achieve that with custom DagsterType NpsDataFrame
and TagDataFrame
in my pipeline (see attachment), but that will not show the type in Dagit. How could I use a generic solid but returning different typed DataFrames? I'd like to see NpsDataFrame and TagDataFrame instead of generic PySparkDataFrame. Any best practices? Or should I add an additional parameter to load_delta_table_to_df
where I define the output DataFrame? Thanks a lot guys!alex
02/10/2021, 5:31 PMoutput_defs
https://docs.dagster.io/overview/solids-pipelines/solid-factories#mainSimon Späti
02/10/2021, 8:12 PMmrdavidlaing
02/10/2021, 10:18 PMSimon Späti
02/10/2021, 10:32 PM