Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug] vLLM images missing from Docker Hub #961

Open
3 of 6 tasks
eero-t opened this issue Dec 2, 2024 · 3 comments
Open
3 of 6 tasks

[Bug] vLLM images missing from Docker Hub #961

eero-t opened this issue Dec 2, 2024 · 3 comments
Assignees
Labels
bug Something isn't working

Comments

@eero-t
Copy link
Contributor

eero-t commented Dec 2, 2024

Priority

Undecided

OS type

Ubuntu

Hardware type

Gaudi2

Installation method

  • Pull docker images from hub.docker.com
  • Build docker images from source

Deploy method

  • Docker compose
  • Docker
  • Kubernetes
  • Helm

Running nodes

Single Node

What's the version?

latest from DockerHub, with Always pull policy.

Description

DocSum and FaqGen images are missing for vLLM: https://hub.docker.com/u/opea?page=1&search=-vllm

Although they are there for TGI: https://hub.docker.com/u/opea?page=1&search=-tgi

And there are vLLM Dockerfiles for them:

And ChatQnA (text generation) vLLM image is already there: https://hub.docker.com/r/opea/llm-vllm

Because of this, CI fails for: opea-project/GenAIInfra#610

Reproduce steps

docker pull opea/llm-docsum-vllm

Raw log

No response

Attachments

No response

@eero-t eero-t added the bug Something isn't working label Dec 2, 2024
@eero-t eero-t changed the title [Bug] Missing vLLM images [Bug] vLLM images missing from Docker Hub Dec 2, 2024
@eero-t
Copy link
Contributor Author

eero-t commented Dec 2, 2024

Note: alternatively llm-vllm image could support also FaqGen & DocSum, but that would require more work than just uploading the built images to DockerHub.

@feng-intel feng-intel self-assigned this Dec 3, 2024
@feng-intel
Copy link
Collaborator

”opea/llm-docsum-vllm“ image will be added in version 1.1 and it needs OSPDT now.

@eero-t
Copy link
Contributor Author

eero-t commented Dec 3, 2024

Both FaqGen & DocSum image variants have the same difference to (already existing) text-generation vLLM langchain wrapper image deps:

--- requirements.txt	2024-11-12 14:55:59.952764826 +0200
+++ ../../../summarization/vllm/langchain/requirements.txt	2024-11-14 15:05:44.350115888 +0200
@@ -13,4 +13,3 @@
 shortuuid
 transformers
 uvicorn
-vllm

And similar difference in what Dockerfile installs:

 RUN pip install --no-cache-dir --upgrade pip setuptools && \
-    if [ ${ARCH} = "cpu" ]; then \
-      pip install --no-cache-dir --extra-index-url https://download.pytorch.org/whl/cpu -r /home/user/comps/llms/text-generation/vllm/langchain/requirements.txt; \
-    else \
-      pip install --no-cache-dir -r /home/user/comps/llms/text-generation/vllm/langchain/requirements.txt; \
-    fi
+    if [ ${ARCH} = "cpu" ]; then pip install --no-cache-dir torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \
+    pip install --no-cache-dir -r /home/user/comps/llms/summarization/vllm/langchain/requirements.txt

And all of them install the problematic langserve component at runtime, like the already existing TGI variant images:

$ head */*/*/requirements-runtime.txt
==> faq-generation/tgi/langchain/requirements-runtime.txt <==
langserve

==> faq-generation/vllm/langchain/requirements-runtime.txt <==
langserve

==> summarization/tgi/langchain/requirements-runtime.txt <==
langserve

==> summarization/vllm/langchain/requirements-runtime.txt <==
langserve

==> text-generation/ollama/langchain/requirements-runtime.txt <==
langserve

==> text-generation/vllm/langchain/requirements-runtime.txt <==
langserve

==> text-generation/vllm/llama_index/requirements-runtime.txt <==
langserve

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants