@@ -358,15 +358,15 @@ def test_processing_step(sklearn_processor_fixture, sagemaker_session, sfn_clien
358
358
def test_processing_step_with_placeholders (sklearn_processor_fixture , sagemaker_session , sfn_client , sfn_role_arn ,
359
359
sagemaker_role_arn ):
360
360
region = boto3 .session .Session ().region_name
361
- input_data = ' s3://sagemaker-sample-data-{}/processing/census/census-income.csv' . format ( region )
361
+ input_data = f" s3://sagemaker-sample-data-{ region } /processing/census/census-income.csv"
362
362
363
363
input_s3 = sagemaker_session .upload_data (
364
364
path = os .path .join (DATA_DIR , 'sklearn_processing' ),
365
365
bucket = sagemaker_session .default_bucket (),
366
366
key_prefix = 'integ-test-data/sklearn_processing/code'
367
367
)
368
368
369
- output_s3 = ' s3://' + sagemaker_session .default_bucket () + ' /integ-test-data/sklearn_processing'
369
+ output_s3 = f" s3://{ sagemaker_session .default_bucket ()} /integ-test-data/sklearn_processing"
370
370
371
371
inputs = [
372
372
ProcessingInput (source = input_data , destination = '/opt/ml/processing/input' , input_name = 'input-1' ),
@@ -422,7 +422,6 @@ def test_processing_step_with_placeholders(sklearn_processor_fixture, sagemaker_
422
422
workflow_graph = Chain ([processing_step ])
423
423
424
424
with timeout (minutes = DEFAULT_TIMEOUT_MINUTES ):
425
- # Create workflow and check definition
426
425
workflow = create_workflow_and_check_definition (
427
426
workflow_graph = workflow_graph ,
428
427
workflow_name = unique_name_from_base ("integ-test-processing-step-workflow" ),
@@ -449,4 +448,3 @@ def test_processing_step_with_placeholders(sklearn_processor_fixture, sagemaker_
449
448
450
449
# Cleanup
451
450
state_machine_delete_wait (sfn_client , workflow .state_machine_arn )
452
- # End of Cleanup
0 commit comments