We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent a094488 commit 4c96be0Copy full SHA for 4c96be0
server/text_generation_server/inference_engine/hf_custom_tp.py
@@ -85,7 +85,10 @@ def __init__(
85
elif model_type in ["RefinedWeb", "RefinedWebModel", "falcon"]:
86
if sharded and self._config.alibi:
87
raise NotImplementedError("TP is not supported for Falcon models using alibi")
88
- aliases = {"transformer.word_embeddings.weight": ["lm_head.weight"]}
+ aliases = {
89
+ "transformer.word_embeddings.weight": ["lm_head.weight"],
90
+ "lm_head.weight": ["transformer.word_embeddings.weight"],
91
+ }
92
from text_generation_server.models.custom_modeling.flash_rw_modeling import FlashRWForCausalLM
93
model_class = FlashRWForCausalLM
94
0 commit comments