From 1efb94ddd76e34a4ea0f8502d110c8db9f582a3c Mon Sep 17 00:00:00 2001 From: "Billy D." Date: Mon, 9 Feb 2026 21:16:34 -0500 Subject: [PATCH] fix(strixhalo): add remaining vllm deps for v1.0.24 openai-harmony, llguidance, conch-triton-kernels, model-hosting-container-standards, runai-model-streamer, timm --- dockerfiles/Dockerfile.ray-worker-strixhalo | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/dockerfiles/Dockerfile.ray-worker-strixhalo b/dockerfiles/Dockerfile.ray-worker-strixhalo index 383e1e2..550ff74 100644 --- a/dockerfiles/Dockerfile.ray-worker-strixhalo +++ b/dockerfiles/Dockerfile.ray-worker-strixhalo @@ -215,7 +215,13 @@ RUN --mount=type=cache,target=/root/.cache/uv \ 'grpcio-tools>=1.60.0' \ 'anthropic>=0.20.0' \ 'mcp>=1.0' \ - 'tensorizer>=2.9.0' + 'tensorizer>=2.9.0' \ + 'openai-harmony>=0.0.6' \ + 'llguidance>=1.0' \ + 'conch-triton-kernels>=1.0' \ + 'model-hosting-container-standards>=0.1.0' \ + 'runai-model-streamer>=0.15.0' \ + 'timm>=1.0' # ── Ray Serve application package ────────────────────────────────────── # Baked into the image so the LLM serve app can use the source-built vllm