diff --git a/ods_ci/tests/Tests/400__ods_dashboard/420__model_serving/LLMs/422__model_serving_llm_models.robot b/ods_ci/tests/Tests/400__ods_dashboard/420__model_serving/LLMs/422__model_serving_llm_models.robot index eda7ac3b1..c75d54ed4 100644 --- a/ods_ci/tests/Tests/400__ods_dashboard/420__model_serving/LLMs/422__model_serving_llm_models.robot +++ b/ods_ci/tests/Tests/400__ods_dashboard/420__model_serving/LLMs/422__model_serving_llm_models.robot @@ -36,8 +36,8 @@ Verify User Can Serve And Query A bigscience/mt0-xxl Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=900s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=900s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -82,8 +82,8 @@ Verify User Can Serve And Query A google/flan-t5-xl Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=900s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -128,8 +128,8 @@ Verify User Can Serve And Query A google/flan-t5-xxl Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=900s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -173,8 +173,8 @@ Verify User Can Serve And Query A elyza/elyza-japanese-llama-2-7b-instruct Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=900s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -219,8 +219,8 @@ Verify User Can Serve And Query A ibm/mpt-7b-instruct2 Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=900s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -265,8 +265,8 @@ Verify User Can Serve And Query A google/flan-ul-2 Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=900s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -311,8 +311,8 @@ Verify User Can Serve And Query A codellama/codellama-34b-instruct-hf Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=3000s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=3000s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -348,8 +348,8 @@ Verify User Can Serve And Query A meta-llama/llama-2-13b-chat Model ... limits_dict=${limits} requests_dict=${requests} kserve_mode=${KSERVE_MODE} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=900s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name} @@ -399,8 +399,8 @@ Verify User Can Serve And Query A google/flan-t5-xl Prompt Tuned Model ... overlays=${overlays} Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH} ... namespace=${test_namespace} - Wait For Pods To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} - ... namespace=${test_namespace} timeout=300s + Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name} + ... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=300s ${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name} Run Keyword If "${KSERVE_MODE}"=="RawDeployment" ... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}