Averell
04/04/2022, 3:48 AMresources/__init__.py
, and my root module is analytx
(so in my code, my imports are like from analytx.assets import ...
)
analytx
├── __init__.py
├── assets
├── hooks
├── jobs
├── partitions
├── repo.py
├── resources
├── schedules
├── sensors
└── utils
The three config params I'm trying to deal with are deploy_local_pipeline_package
, s3_job_package_path
, and local_pipeline_package_path
.
1. with deploy_local_pipeline_package = True
,
a. if I set local_pipeline_package_path
to Path(__file__).parent.parent
, then the zip file code.zip will have all those subdirs of analytx at the root level of the zip file. Then the EMR step will fail with analytx not found
b. if I set local_pipeline_package_path
to Path(__file__).parent.parent.parent
, then the execution stuck at building the zip file
c. with Path(__file__).parent.parent.parent + "/analytx"
, I get the same result as (1a)
2. with deploy_local_pipeline_package = False
, I expected that the path to s3_job_package_path
will be sent in spark-submit, but it actually didn't. Also, no code.zip was built and copied to S3 (which was expected), but the spark-submit still refers to that file, and fails, obviously.Averell
04/04/2022, 12:48 PMAverell
04/04/2022, 12:51 PMclaire
04/04/2022, 9:37 PMclaire
04/04/2022, 9:50 PMAverell
04/05/2022, 3:04 PMclaire
04/05/2022, 6:19 PMPragna
10/27/2022, 11:42 PM