awswrangler.emr.build_spark_step¶
- awswrangler.emr.build_spark_step(path: str, deploy_mode: str = 'cluster', docker_image: Optional[str] = None, name: str = 'my-step', action_on_failure: str = 'CONTINUE', region: Optional[str] = None, boto3_session: Optional[Session] = None) Dict[str, Any] ¶
Build the Step structure (dictionary).
- Parameters
path (str) – Script path. (e.g. s3://bucket/app.py)
deploy_mode (str) – “cluster” | “client”
docker_image (str, optional) – e.g. “{ACCOUNT_ID}.dkr.ecr.{REGION}.amazonaws.com/{IMAGE_NAME}:{TAG}”
name (str, optional) – Step name.
action_on_failure (str) – ‘TERMINATE_JOB_FLOW’, ‘TERMINATE_CLUSTER’, ‘CANCEL_AND_WAIT’, ‘CONTINUE’
region (str, optional) – Region name to not get it from boto3.Session. (e.g. us-east-1)
boto3_session (boto3.Session(), optional) – Boto3 Session. The default boto3 session will be used if boto3_session receive None.
- Returns
Step structure.
- Return type
Dict[str, Any]
Examples
>>> import awswrangler as wr >>> step_id = wr.emr.submit_steps( >>> cluster_id="cluster-id", >>> steps=[ >>> wr.emr.build_spark_step(path="s3://bucket/app.py") >>> ] >>> )