diff --git a/test/image-blueprints-bootc/layer3-periodic/group1/rhel96-bootc-source-ai-model-serving.containerfile b/test/image-blueprints-bootc/layer3-periodic/group1/rhel96-bootc-source-ai-model-serving.containerfile index 5618d4bc7c..f56166d4a3 100644 --- a/test/image-blueprints-bootc/layer3-periodic/group1/rhel96-bootc-source-ai-model-serving.containerfile +++ b/test/image-blueprints-bootc/layer3-periodic/group1/rhel96-bootc-source-ai-model-serving.containerfile @@ -19,12 +19,8 @@ COPY ./bootc-images/$USHIFT_RPM_REPO_NAME.repo ./bootc-images/microshift-fast-da # vim-common for xxd utility RUN dnf repoinfo --enabled && \ dnf install -y \ - {{ range (env.Getenv "MICROSHIFT_OPTIONAL_RPMS" | strings.Split " ") -}} - "{{ . }}-{{ env.Getenv "SOURCE_VERSION" }}" \ - {{ end -}} - {{ range (env.Getenv "MICROSHIFT_X86_64_RPMS" | strings.Split " ") -}} - "{{ . }}-{{ env.Getenv "SOURCE_VERSION" }}" \ - {{ end -}} + "microshift-ai-model-serving-{{ env.Getenv "SOURCE_VERSION" }}" \ + "microshift-ai-model-serving-release-info-{{ env.Getenv "SOURCE_VERSION" }}" \ vim-common \ qemu-guest-agent && \ rm -vf /etc/yum.repos.d/microshift-*.repo && \ diff --git a/test/scenarios-bootc/periodics/el96-src@ai-model-serving-offline.sh.disabled b/test/scenarios-bootc/periodics/el96-src@ai-model-serving-offline.sh similarity index 100% rename from test/scenarios-bootc/periodics/el96-src@ai-model-serving-offline.sh.disabled rename to test/scenarios-bootc/periodics/el96-src@ai-model-serving-offline.sh diff --git a/test/suites/ai-model-serving/ai-model-serving-offline.robot b/test/suites/ai-model-serving/ai-model-serving-offline.robot index 495056fa1c..c8d1445fa6 100644 --- a/test/suites/ai-model-serving/ai-model-serving-offline.robot +++ b/test/suites/ai-model-serving/ai-model-serving-offline.robot @@ -43,7 +43,14 @@ Model Serving Offline Test Wait For A Deployment [Documentation] Wait for a deployment on offline VM [Arguments] ${namespace} ${name} - offline.Run With Kubeconfig oc rollout status -n\=${namespace} deployment ${name} + offline.Run With Kubeconfig + ... oc + ... rollout + ... status + ... --timeout\=300s + ... -n\=${namespace} + ... deployment + ... ${name} Check If Model Is Ready [Documentation] Ask model server is model is ready for inference.