diff --git a/examples/vertex_mlops_enterprise/build/pipeline-deployment-tfx.yaml.TEMPLATE b/examples/vertex_mlops_enterprise/build/pipeline-deployment-tfx.yaml.TEMPLATE index 7e55ed484f9..02f94833ce1 100644 --- a/examples/vertex_mlops_enterprise/build/pipeline-deployment-tfx.yaml.TEMPLATE +++ b/examples/vertex_mlops_enterprise/build/pipeline-deployment-tfx.yaml.TEMPLATE @@ -115,8 +115,8 @@ steps: # Upload compiled pipeline to GCS. -- name: 'gcr.io/cloud-builders/gsutil' - args: ['cp', '$_PIPELINE_NAME.json', '$_PIPELINES_STORE'] +- name: 'gcr.io/cloud-builders/gcloud' + args: ['storage', 'cp', '$_PIPELINE_NAME.json', '$_PIPELINES_STORE'] dir: '$_WORKDIR' id: 'Upload Pipeline to GCS' waitFor: ['Compile Pipeline'] diff --git a/examples/vertex_mlops_enterprise/doc/03-MLOPS.md b/examples/vertex_mlops_enterprise/doc/03-MLOPS.md index c6db45dc094..84cb37ce851 100644 --- a/examples/vertex_mlops_enterprise/doc/03-MLOPS.md +++ b/examples/vertex_mlops_enterprise/doc/03-MLOPS.md @@ -31,7 +31,7 @@ BUCKET=gs://$PROJECT/data/credit_cards* #Extract & Load bq extract --project_id $PROJECT --destination_format PARQUET $SRC_TABLE $BUCKET bq load --project_id $PROJECT --source_format=PARQUET --replace=true $DST_TABLE $BUCKET -gsutil rm $BUCKET +gcloud storage rm $BUCKET ``` As next steps, we will create the base table we will use for the ML process: diff --git a/examples/vertex_mlops_enterprise/doc/create_tables.sh b/examples/vertex_mlops_enterprise/doc/create_tables.sh index 1af9f3e1b01..061bcef9123 100644 --- a/examples/vertex_mlops_enterprise/doc/create_tables.sh +++ b/examples/vertex_mlops_enterprise/doc/create_tables.sh @@ -28,7 +28,7 @@ ENDPOINT="$REGION-aiplatform.googleapis.com" #Extract & Load bq extract --project_id "$PROJECT" --destination_format PARQUET "$SRC_TABLE" "$BUCKET" bq load --project_id "$PROJECT" --source_format=PARQUET --replace=true "$DST_TABLE" "$BUCKET" -gsutil rm "$BUCKET" +gcloud storage rm "$BUCKET" sql_script="CREATE OR REPLACE TABLE \`${PROJECT}.${BQ_DATASET_NAME}.${ML_TABLE}\`