-
Notifications
You must be signed in to change notification settings - Fork 4.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LLaMa-factory 部署Llama-3.2-11B-Vision-Instruct 运行报错 #5549
Comments
"Transformers version: 4.42.3" |
目前看到支持 以下LLAMA-3.2的模型,register_model_group( |
目前看到支持 以下LLAMA-3.2的模型,register_model_group( |
不是,这只是更新了这些模型的下载资源,你要运行起来 3.2 还是要按照官方的要求更新最新的Transformers |
好的,我试试看,谢谢呀 |
是这样的没有支持,我这里有个pr你们可以试试,所有库的版本请确保是最新的,我在a100完成了训练与推理的测试,但是b&b的8bit量化还是有问题4bit就行。目测应该是b&b的问题到时候会解决一下 |
那LLaMA-Factory什么时候会更新支持Llama-3.2-11B-Vision-Instruct的版本呢?预期是多久呢? |
instruct是可以的在我内个分支,只是没处理图像应该关注哪些token的问题因为涉及到batch处理我又时间不多。 |
你好,是下载这个分支的代码嘛https://github.com/marko1616/LLaMA-Factory/tree/feat/llama3.2vl 来运行嘛 |
是这样的 |
下载这个分支运行会报错如下:Traceback (most recent call last):
|
@caijx168 update transformers |
我升级之后报错如下(base) root@root1-System-Product-Name:/home/LLaMA-Factory/LLaMA-Factory-feat-llama3.2vl# tail -f nohup.out |
@caijx168 update llamafactory |
我下载的这个分支就是最新的呢https://github.com/marko1616/LLaMA-Factory/tree/feat/llama3.2vl |
改改代码把src/llamafactory/extras/misc.py的transformers版本改掉,或者降低版本 |
transformers 版本降为4.45.0 运行会报这个错误
|
不支持vllm后端哦 |
|
我之前都是用这个来部署的利用 vLLM 部署 OpenAI API |
把vllm去掉就行 |
fixed |
Reminder
System Info
llamafactory
version: 0.8.3.dev0Reproduction
运行命令如下:
CUDA_VISIBLE_DEVICES=0 API_PORT=8005 nohup python src/api.py
--model_name_or_path /home/Llama-3.2/Llama-3.2-11B-Vision-Instruct
--template llama3
--infer_backend vllm
--vllm_maxlen 8000
--vllm_gpu_util 0.8
--vllm_enforce_eager true &
运行报错如下:[INFO|configuration_utils.py:731] 2024-09-26 13:55:31,084 >> loading configuration file /home/Llama-3.2/Llama-3.2-11B-Vision-Instruct/config.json
Traceback (most recent call last):
File "/root/anaconda3/envs/LLaMA-Factory-main/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 982, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
File "/root/anaconda3/envs/LLaMA-Factory-main/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 684, in getitem
raise KeyError(key)
KeyError: 'mllama'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/LLaMA-Factory-main/src/api.py", line 33, in
main()
File "/home/LLaMA-Factory-main/src/api.py", line 24, in main
chat_model = ChatModel()
File "/home/LLaMA-Factory-main/src/llamafactory/chat/chat_model.py", line 45, in init
self.engine: "BaseEngine" = VllmEngine(model_args, data_args, finetuning_args, generating_args)
File "/home/LLaMA-Factory-main/src/llamafactory/chat/vllm_engine.py", line 55, in init
config = load_config(model_args) # may download model from ms hub
File "/home/LLaMA-Factory-main/src/llamafactory/model/loader.py", line 117, in load_config
return AutoConfig.from_pretrained(model_args.model_name_or_path, **init_kwargs)
File "/root/anaconda3/envs/LLaMA-Factory-main/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 984, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type
mllama
but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.Expected behavior
正确运行Llama-3.2-11B-Vision-Instruct
Others
No response
The text was updated successfully, but these errors were encountered: