From a8fcec5f702fc2528c87ed6fd698b9cfca8b509e Mon Sep 17 00:00:00 2001 From: Jiaxiao Zheng Date: Thu, 19 Dec 2019 11:39:56 -0800 Subject: [PATCH] Revert "Components - De-hardcoded the UI metadata file path in GCP components (#2697)" (#2759) This reverts commit 08ebd156 --- components/gcp/bigquery/query/component.yaml | 4 ++-- .../python/kfp_component/launcher/__main__.py | 7 ------- components/gcp/dataflow/launch_python/component.yaml | 4 ++-- components/gcp/dataflow/launch_template/component.yaml | 2 +- components/gcp/dataproc/create_cluster/component.yaml | 4 ++-- components/gcp/dataproc/submit_hadoop_job/component.yaml | 4 ++-- components/gcp/dataproc/submit_hive_job/component.yaml | 4 ++-- components/gcp/dataproc/submit_pig_job/component.yaml | 4 ++-- components/gcp/dataproc/submit_pyspark_job/component.yaml | 4 ++-- components/gcp/dataproc/submit_spark_job/component.yaml | 4 ++-- components/gcp/dataproc/submit_sparksql_job/component.yaml | 4 ++-- components/gcp/ml_engine/batch_predict/component.yaml | 4 ++-- components/gcp/ml_engine/deploy/component.yaml | 2 +- components/gcp/ml_engine/train/component.yaml | 4 ++-- 14 files changed, 24 insertions(+), 31 deletions(-) diff --git a/components/gcp/bigquery/query/component.yaml b/components/gcp/bigquery/query/component.yaml index 631fecc2b54..7673fef8c3b 100644 --- a/components/gcp/bigquery/query/component.yaml +++ b/components/gcp/bigquery/query/component.yaml @@ -68,10 +68,10 @@ implementation: --table_id, {inputValue: table_id}, --dataset_location, {inputValue: dataset_location}, --output_gcs_path, {inputValue: output_gcs_path}, - --job_config, {inputValue: job_config}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --job_config, {inputValue: job_config} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: output_gcs_path: /tmp/kfp/output/bigquery/query-output-path.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/container/component_sdk/python/kfp_component/launcher/__main__.py b/components/gcp/container/component_sdk/python/kfp_component/launcher/__main__.py index 0e767679add..7d9937935c9 100644 --- a/components/gcp/container/component_sdk/python/kfp_component/launcher/__main__.py +++ b/components/gcp/container/component_sdk/python/kfp_component/launcher/__main__.py @@ -15,7 +15,6 @@ import argparse import fire import importlib -import os import sys import logging from .launcher import launch @@ -27,14 +26,8 @@ def main(): description='Launch a python module or file.') parser.add_argument('file_or_module', type=str, help='Either a python file path or a module name.') - parser.add_argument('--ui-metadata-path', type=str, - help='Path for the file where the mlpipeline-ui-metadata.json data should be written.') parser.add_argument('args', nargs=argparse.REMAINDER) args = parser.parse_args() - - if args.ui_metadata_path: - os.environ['KFP_UI_METADATA_PATH'] = args.ui_metadata_path - launch(args.file_or_module, args.args) if __name__ == '__main__': diff --git a/components/gcp/dataflow/launch_python/component.yaml b/components/gcp/dataflow/launch_python/component.yaml index 687a21a22b9..74168456ffb 100644 --- a/components/gcp/dataflow/launch_python/component.yaml +++ b/components/gcp/dataflow/launch_python/component.yaml @@ -61,10 +61,10 @@ implementation: --staging_dir, {inputValue: staging_dir}, --requirements_file_path, {inputValue: requirements_file_path}, --args, {inputValue: args}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataflow/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataflow/launch_template/component.yaml b/components/gcp/dataflow/launch_template/component.yaml index 846ba6711d0..ea98c3d2ef8 100644 --- a/components/gcp/dataflow/launch_template/component.yaml +++ b/components/gcp/dataflow/launch_template/component.yaml @@ -73,9 +73,9 @@ implementation: --validate_only, {inputValue: validate_only}, --staging_dir, {inputValue: staging_dir}, --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataflow/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataproc/create_cluster/component.yaml b/components/gcp/dataproc/create_cluster/component.yaml index c8c9bea881b..d625ecbff2a 100644 --- a/components/gcp/dataproc/create_cluster/component.yaml +++ b/components/gcp/dataproc/create_cluster/component.yaml @@ -81,10 +81,10 @@ implementation: --config_bucket, {inputValue: config_bucket}, --image_version, {inputValue: image_version}, --cluster, {inputValue: cluster}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: cluster_name: /tmp/kfp/output/dataproc/cluster_name.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataproc/submit_hadoop_job/component.yaml b/components/gcp/dataproc/submit_hadoop_job/component.yaml index df5f2237038..d8d031cf588 100644 --- a/components/gcp/dataproc/submit_hadoop_job/component.yaml +++ b/components/gcp/dataproc/submit_hadoop_job/component.yaml @@ -91,10 +91,10 @@ implementation: --args, {inputValue: args}, --hadoop_job, {inputValue: hadoop_job}, --job, {inputValue: job}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataproc/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataproc/submit_hive_job/component.yaml b/components/gcp/dataproc/submit_hive_job/component.yaml index 8c5fec19145..837abff02ef 100644 --- a/components/gcp/dataproc/submit_hive_job/component.yaml +++ b/components/gcp/dataproc/submit_hive_job/component.yaml @@ -86,10 +86,10 @@ implementation: --script_variables, {inputValue: script_variables}, --hive_job, {inputValue: hive_job}, --job, {inputValue: job}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataproc/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataproc/submit_pig_job/component.yaml b/components/gcp/dataproc/submit_pig_job/component.yaml index e91336af4b5..b11e716ac3c 100644 --- a/components/gcp/dataproc/submit_pig_job/component.yaml +++ b/components/gcp/dataproc/submit_pig_job/component.yaml @@ -86,10 +86,10 @@ implementation: --script_variables, {inputValue: script_variables}, --pig_job, {inputValue: pig_job}, --job, {inputValue: job}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataproc/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataproc/submit_pyspark_job/component.yaml b/components/gcp/dataproc/submit_pyspark_job/component.yaml index 4e3079cbec1..b440e6a92dc 100644 --- a/components/gcp/dataproc/submit_pyspark_job/component.yaml +++ b/components/gcp/dataproc/submit_pyspark_job/component.yaml @@ -79,10 +79,10 @@ implementation: --args, {inputValue: args}, --pyspark_job, {inputValue: pyspark_job}, --job, {inputValue: job}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataproc/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataproc/submit_spark_job/component.yaml b/components/gcp/dataproc/submit_spark_job/component.yaml index 7fff9592e9e..5bf3d6f6889 100644 --- a/components/gcp/dataproc/submit_spark_job/component.yaml +++ b/components/gcp/dataproc/submit_spark_job/component.yaml @@ -87,10 +87,10 @@ implementation: --args, {inputValue: args}, --spark_job, {inputValue: spark_job}, --job, {inputValue: job}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataproc/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/dataproc/submit_sparksql_job/component.yaml b/components/gcp/dataproc/submit_sparksql_job/component.yaml index e47687381be..c42f5423016 100644 --- a/components/gcp/dataproc/submit_sparksql_job/component.yaml +++ b/components/gcp/dataproc/submit_sparksql_job/component.yaml @@ -86,10 +86,10 @@ implementation: --script_variables, {inputValue: script_variables}, --sparksql_job, {inputValue: sparksql_job}, --job, {inputValue: job}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/dataproc/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/ml_engine/batch_predict/component.yaml b/components/gcp/ml_engine/batch_predict/component.yaml index d12127de365..f0ba536351c 100644 --- a/components/gcp/ml_engine/batch_predict/component.yaml +++ b/components/gcp/ml_engine/batch_predict/component.yaml @@ -81,10 +81,10 @@ implementation: --output_data_format, {inputValue: output_data_format}, --prediction_input, {inputValue: prediction_input}, --job_id_prefix, {inputValue: job_id_prefix}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/ml_engine/job_id.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/ml_engine/deploy/component.yaml b/components/gcp/ml_engine/deploy/component.yaml index 65789e2e409..df0cd8dad64 100644 --- a/components/gcp/ml_engine/deploy/component.yaml +++ b/components/gcp/ml_engine/deploy/component.yaml @@ -109,7 +109,6 @@ implementation: --replace_existing_version, {inputValue: replace_existing_version}, --set_default, {inputValue: set_default}, --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, ] env: KFP_POD_NAME: "{{pod.name}}" @@ -117,3 +116,4 @@ implementation: model_uri: /tmp/kfp/output/ml_engine/model_uri.txt model_name: /tmp/kfp/output/ml_engine/model_name.txt version_name: /tmp/kfp/output/ml_engine/version_name.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json diff --git a/components/gcp/ml_engine/train/component.yaml b/components/gcp/ml_engine/train/component.yaml index 07c811e1b6f..4ca0c556626 100644 --- a/components/gcp/ml_engine/train/component.yaml +++ b/components/gcp/ml_engine/train/component.yaml @@ -118,11 +118,11 @@ implementation: --worker_image_uri, {inputValue: worker_image_uri}, --training_input, {inputValue: training_input}, --job_id_prefix, {inputValue: job_id_prefix}, - --wait_interval, {inputValue: wait_interval}, - --ui-metadata-path, {outputPath: MLPipeline UI metadata}, + --wait_interval, {inputValue: wait_interval} ] env: KFP_POD_NAME: "{{pod.name}}" fileOutputs: job_id: /tmp/kfp/output/ml_engine/job_id.txt job_dir: /tmp/kfp/output/ml_engine/job_dir.txt + MLPipeline UI metadata: /mlpipeline-ui-metadata.json