-
Notifications
You must be signed in to change notification settings - Fork 202
Issues: QwenLM/Qwen2-VL
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Qwen2-VL-72B vLLM Response Issue with Larger Images: Short and Incomplete Outputs
#551
opened Nov 20, 2024 by
KaranBhuva22
不管使用A6000*4还是A800*4推理2B, 7B还是72B模型,选择device_map=auto都会出现各种奇怪的报错,只有device_map="cuda:x"单卡才能正常使用
#546
opened Nov 19, 2024 by
luosting
用vllm部署Qwen2-VL-7B-Instruct模型,单次请求输入2张图片生成内容不完整,输入单张图片生成内容是完整的
#545
opened Nov 19, 2024 by
xj260098061
Qwen-VL-7B model +vllm assert len(indices) == len(inputs) AssertionError
#542
opened Nov 18, 2024 by
sssunXw
Confusion about the effect of special tokens on model fine-tuning.
#537
opened Nov 18, 2024 by
Davidwhw
is there a way to reduce vRAM usage & speed up vLLM inference for AWQ?
#532
opened Nov 12, 2024 by
mehamednews
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-10-23.