Skip to content

有没有详细一点的web_demo推理部署流程? #92

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
LayKwokMing opened this issue Jan 15, 2025 · 3 comments
Open

有没有详细一点的web_demo推理部署流程? #92

LayKwokMing opened this issue Jan 15, 2025 · 3 comments

Comments

@LayKwokMing
Copy link

有没有详细一点的web_demo推理部署流程?按readme去做,接连报错。
1、先是报 Repo is must be in the form ‘repo_name’ or namespace/repo_name......,排查是readme中“mv demo_VITA_ckpt/config.json demo_VITA_ckpt/origin_config.json”这句话把config.json去掉了,把config.json补上,问题解决。
2、然后遇到 limit_mm_per_prompt is only supported for multimodal models. 把limit_mm_per_prompt={‘image’:256, 'audio':50}注释掉,问题解决。
3、又遇到vllm中['VITAQwen2ForCausalLM'] are not supported for now的问题,至此怀疑自己的操作步骤有问题。
有没有大佬可以解答一下

@lxysl
Copy link
Contributor

lxysl commented Jan 16, 2025

All the issues you encountered are due to not executing the following operations:

# Backup a new weight file
cp -rL  VITA_ckpt/ demo_VITA_ckpt/

mv demo_VITA_ckpt/config.json demo_VITA_ckpt/origin_config.json

cd ./web_demo/vllm_tools
cp -rf qwen2p5_model_weight_file/*  ../../demo_VITA_ckpt/
cp -rf vllm_file/*  your_anaconda/envs/vita_demo/lib/python3.10/site-packages/vllm/model_executor/models/

Our readme about Demo have the following structure:


Demo

readme0

📍 Basic Demo

readme1

📍 Real-Time Interactive Demo

readme2


Readme0 must be executed first before readme1 or readme2 is to be executed afterward.

The instructions here may not be clear, and we will revise them later.

@hezhefly
Copy link

@LayKwokMing 我遇到了与你想通过的问题,就是limit_mm_per_prompt is only supported for multimodal models

反复重试后,发现是自己配置错了:
执行:cp -rf vllm_file/* your_anaconda/envs/vita_demo/lib/python3.10/site-packages/vllm/model_executor/models/代码时,只复制了your_anaconda/envs/vita_demo/lib/python3.10/site-packages/vllm/model_executor/,导致后期所有环节都错误。


I encountered the same problem as you. That is, "limit_mm_per_prompt is only supported for multimodal models". After repeated retries, I found that I had configured it incorrectly. When executing the code "cp -rf vllm_file/* your_anaconda/envs/vita_demo/lib/python3.10/site-packages/vllm/model_executor/models/", I only copied "your_anaconda/envs/vita_demo/lib/python3.10/site-packages/vllm/model_executor/", which led to errors in all subsequent steps.

@hopef
Copy link

hopef commented Mar 8, 2025

你们的demo的使用说明,不能用垃圾来形容,简直无可救药!!!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants