-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Description
As the title states, do we need to set the model loader to ExLlamav2_HF or ExLlamav2?
The documentation says:
**ExLlamav2**
The same as ExLlamav2_HF but using the internal samplers of ExLlamav2 instead of the ones in the Transformers library.
I tried doing just that and got an error message:
11:10:48-588864 INFO Loading "turboderp_Mistral-Large-Instruct-2407-123B-exl2"
11:10:48-593031 ERROR Failed to load the model.
Traceback (most recent call last):
File "/opt/text-generation-webui/modules/ui_model_menu.py", line 231, in load_model_wrapper
shared.model, shared.tokenizer = load_model(selected_model, loader)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/text-generation-webui/modules/models.py", line 93, in load_model
output = load_func_map[loader](model_name)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/text-generation-webui/modules/models.py", line 312, in ExLlamav2_loader
from modules.exllamav2 import Exllamav2Model
File "/opt/text-generation-webui/modules/exllamav2.py", line 13, in <module>
from exllamav2 import (
ImportError: cannot import name 'ExLlamaV2Cache_TP' from 'exllamav2' (/opt/text-generation-webui/installer_files/env/lib/python3.11/site-packages/exllamav2/__init__.py)
And when I selected turboderp_Mistral-Large-Instruct-2407-123B-exl2 in the dropdown menu, it automatically defaulted to ExLlamav2_HF as the loader.
Metadata
Metadata
Assignees
Labels
No labels