Skip to content

Commit

Permalink
remove extra space in llm models suite
Browse files Browse the repository at this point in the history
  • Loading branch information
bdattoma committed Apr 24, 2024
1 parent aa019a7 commit e445249
Showing 1 changed file with 9 additions and 9 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ Verify User Can Serve And Query A bigscience/mt0-xxl Model
... kserve_mode=${KSERVE_MODE}
... AND
... Run Keyword If "${KSERVE_MODE}"=="RawDeployment" Terminate Process llm-query-process kill=true

Verify User Can Serve And Query A google/flan-t5-xl Model
[Documentation] Basic tests for preparing, deploying and querying a LLM model
... using Kserve and TGIS runtime
Expand All @@ -83,7 +83,7 @@ Verify User Can Serve And Query A google/flan-t5-xl Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=900s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down Expand Up @@ -129,7 +129,7 @@ Verify User Can Serve And Query A google/flan-t5-xxl Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=900s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down Expand Up @@ -174,7 +174,7 @@ Verify User Can Serve And Query A elyza/elyza-japanese-llama-2-7b-instruct Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=900s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down Expand Up @@ -220,7 +220,7 @@ Verify User Can Serve And Query A ibm/mpt-7b-instruct2 Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=900s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down Expand Up @@ -266,7 +266,7 @@ Verify User Can Serve And Query A google/flan-ul-2 Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=900s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down Expand Up @@ -312,7 +312,7 @@ Verify User Can Serve And Query A codellama/codellama-34b-instruct-hf Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=3000s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=3000s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down Expand Up @@ -349,7 +349,7 @@ Verify User Can Serve And Query A meta-llama/llama-2-13b-chat Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=900s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=900s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down Expand Up @@ -400,7 +400,7 @@ Verify User Can Serve And Query A google/flan-t5-xl Prompt Tuned Model
Deploy Model Via CLI isvc_filepath=${INFERENCESERVICE_FILLED_FILEPATH}
... namespace=${test_namespace}
Wait For Model KServe Deployment To Be Ready label_selector=serving.kserve.io/inferenceservice=${model_name}
... namespace=${test_namespace} runtime= ${TGIS_RUNTIME_NAME} timeout=300s
... namespace=${test_namespace} runtime=${TGIS_RUNTIME_NAME} timeout=300s
${pod_name}= Get Pod Name namespace=${test_namespace} label_selector=serving.kserve.io/inferenceservice=${model_name}
Run Keyword If "${KSERVE_MODE}"=="RawDeployment"
... Start Port-forwarding namespace=${test_namespace} pod_name=${pod_name}
Expand Down

0 comments on commit e445249

Please sign in to comment.