You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Traceback (most recent call last):
File "/usr/local/bin/text-generation-server", line 8, in <module>
sys.exit(app())
File "/usr/local/lib/python3.10/dist-packages/text_generation_server/cli.py", line 170, in serve
server.serve(
File "/usr/local/lib/python3.10/dist-packages/text_generation_server/server.py", line 259, in serve
asyncio.run(
File "/usr/lib/python3.10/asyncio/runners.py", line 44, in run
return loop.run_until_complete(main)
File "/usr/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete
return future.result()
File "/usr/local/lib/python3.10/dist-packages/text_generation_server/server.py", line 213, in serve_inner
model = get_model_with_lora_adapters(
File "/usr/local/lib/python3.10/dist-packages/text_generation_server/models/__init__.py", line 227, in get_model_with_lora_adapters
model = get_model(
File "/usr/local/lib/python3.10/dist-packages/text_generation_server/models/__init__.py", line 200, in get_model
return CausalLM(
File "/usr/local/lib/python3.10/dist-packages/text_generation_server/models/causal_lm.py", line 695, in __init__
raise ValueError(f"Model type {model.config.model_type} is not supported!")
ValueError: Model type mllama_text_model is not supported!
Expected behavior
I would expect the server to start successfully, since I see the meta-llama/Llama-3.2-11B-Vision-Instruct model linked on the supported models page as "Mllama".
The text was updated successfully, but these errors were encountered:
System Info
Docker image:
ghcr.io/huggingface/tgi-gaudi:2.3.1
OS: Ubuntu 22.04
HL-SMI Version: hl-1.18.0-fw-53.1.1.1
Driver Version: 1.18.0-ee698fb
Information
Tasks
Reproduction
I'm trying to run tgi-gaudi with
meta-llama/Llama-3.2-11B-Vision-Instruct
. I'm starting the tgi-gaudi server using the 2.3.1 docker image like:I'm seeing the following error in the logs:
Expected behavior
I would expect the server to start successfully, since I see the
meta-llama/Llama-3.2-11B-Vision-Instruct
model linked on the supported models page as "Mllama".The text was updated successfully, but these errors were encountered: