Charm Malubag
11/16/2020, 12:49 PMdagster run wipe
and dagster schedule wipe
, and though these were able to reset the runs and schedules from the schedules page, the DB still seems to contain historical data, and the runs page still takes long to load.sashank
11/16/2020, 1:25 PMCharm Malubag
11/16/2020, 1:36 PMsashank
11/16/2020, 1:41 PMdump-*
folder with the results, feel free to DM it to me and I can take a look.#!/bin/bash
# Dagster 0.7.13 runs query profile
NOW="dump-$(date +%s)"
mkdir $NOW
echo "Dumping RunsRootQueryRun with limit=1..."
time curl <http://localhost:3333/graphql> -X POST -H "Content-type: application/json" -d '{"query":"query RunsRootQuery($limit: Int, $cursor: String, $filter: PipelineRunsFilter!) {\n pipelineRunsOrError(limit: $limit, cursor: $cursor, filter: $filter) {\n ... on PipelineRuns {\n results {\n ...RunTableRunFragment\n __typename\n }\n __typename\n }\n ... on InvalidPipelineRunsFilterError {\n message\n __typename\n }\n ... on PythonError {\n message\n __typename\n }\n __typename\n }\n}\n\nfragment RunTableRunFragment on PipelineRun {\n runId\n status\n stepKeysToExecute\n canCancel\n mode\n rootRunId\n parentRunId\n pipelineSnapshotId\n pipeline {\n __typename\n ... on PipelineReference {\n name\n __typename\n }\n ... on Pipeline {\n pipelineSnapshotId\n solids {\n name\n __typename\n }\n __typename\n }\n }\n stats {\n __typename\n ... on PipelineRunStatsSnapshot {\n stepsSucceeded\n stepsFailed\n startTime\n endTime\n expectations\n materializations\n __typename\n }\n ...PythonErrorFragment\n }\n tags {\n key\n value\n __typename\n }\n __typename\n}\n\nfragment PythonErrorFragment on PythonError {\n __typename\n message\n stack\n cause {\n message\n stack\n __typename\n }\n}\n","variables":{"filter":{},"limit":1}}' > "$NOW/RunsRootQuery1.json"
echo "Dumping RunsRootQueryRun with limit=25..."
time curl <http://localhost:3333/graphql> -X POST -H "Content-type: application/json" -d '{"query":"query RunsRootQuery($limit: Int, $cursor: String, $filter: PipelineRunsFilter!) {\n pipelineRunsOrError(limit: $limit, cursor: $cursor, filter: $filter) {\n ... on PipelineRuns {\n results {\n ...RunTableRunFragment\n __typename\n }\n __typename\n }\n ... on InvalidPipelineRunsFilterError {\n message\n __typename\n }\n ... on PythonError {\n message\n __typename\n }\n __typename\n }\n}\n\nfragment RunTableRunFragment on PipelineRun {\n runId\n status\n stepKeysToExecute\n canCancel\n mode\n rootRunId\n parentRunId\n pipelineSnapshotId\n pipeline {\n __typename\n ... on PipelineReference {\n name\n __typename\n }\n ... on Pipeline {\n pipelineSnapshotId\n solids {\n name\n __typename\n }\n __typename\n }\n }\n stats {\n __typename\n ... on PipelineRunStatsSnapshot {\n stepsSucceeded\n stepsFailed\n startTime\n endTime\n expectations\n materializations\n __typename\n }\n ...PythonErrorFragment\n }\n tags {\n key\n value\n __typename\n }\n __typename\n}\n\nfragment PythonErrorFragment on PythonError {\n __typename\n message\n stack\n cause {\n message\n stack\n __typename\n }\n}\n","variables":{"filter":{},"limit":25}}' > "$NOW/RunsRootQuery25.json"
echo "Done."
Charm Malubag
11/16/2020, 4:15 PM{
"data": {
"pipelineRunsOrError": {
"results": [],
"__typename": "PipelineRuns"
}
}
}
however, when one of the pipelines runs and then I run the script, it takes long to query (limit=1 takes about 11mins). Could this be a performance issue in our DB?