Skip to content

Set model loader to ExLlamav2_HF or ExLlamav2? #1

@Kaszebe

Description

@Kaszebe

As the title states, do we need to set the model loader to ExLlamav2_HF or ExLlamav2?

The documentation says:

**ExLlamav2**
The same as ExLlamav2_HF but using the internal samplers of ExLlamav2 instead of the ones in the Transformers library.

I tried doing just that and got an error message:

11:10:48-588864 INFO     Loading "turboderp_Mistral-Large-Instruct-2407-123B-exl2"                        
11:10:48-593031 ERROR    Failed to load the model.                                                        
Traceback (most recent call last):
  File "/opt/text-generation-webui/modules/ui_model_menu.py", line 231, in load_model_wrapper
    shared.model, shared.tokenizer = load_model(selected_model, loader)
                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/text-generation-webui/modules/models.py", line 93, in load_model
    output = load_func_map[loader](model_name)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/text-generation-webui/modules/models.py", line 312, in ExLlamav2_loader
    from modules.exllamav2 import Exllamav2Model
  File "/opt/text-generation-webui/modules/exllamav2.py", line 13, in <module>
    from exllamav2 import (
ImportError: cannot import name 'ExLlamaV2Cache_TP' from 'exllamav2' (/opt/text-generation-webui/installer_files/env/lib/python3.11/site-packages/exllamav2/__init__.py)

And when I selected turboderp_Mistral-Large-Instruct-2407-123B-exl2 in the dropdown menu, it automatically defaulted to ExLlamav2_HF as the loader.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions