John Mav
05/27/2020, 6:07 PM@solid
def get_urls(_):
url_list = pd.read_csv('urls_to_update.csv')
# This can be a list of size 0-100
for url in url_list:
yield Output(url, "url")
@solid
def parse_data(_, url):
res = requests.get(url).json()
structured_data = dict(
url=url,
col_a=res["col_a"],
col_b=res["col_b"]
)
yield Output(structured_data, 'result')
@solid
def upsert_data(_, data):
# Some logic to load the data into warehouse
@pipeline
def my_dynamic_pipeline():
upsert_data(
data=parse_data(
url=get_urls()
)
)
Is this just not feasible? Not a huge issue if not I can work around it.max
05/27/2020, 6:08 PMJohn Mav
05/27/2020, 6:09 PMmax
05/27/2020, 6:09 PMJohn Mav
05/27/2020, 6:12 PM