Skip to content
This repository was archived by the owner on Oct 25, 2024. It is now read-only.

Commit 93b12e9

Browse files
authored
Update README.md (#1570)
Signed-off-by: Dong, Bo <bo1.dong@intel.com>
1 parent 0dcee38 commit 93b12e9

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -219,7 +219,7 @@ inputs = tokenizer(prompt, return_tensors="pt").input_ids
219219
model = AutoModelForCausalLM.from_pretrained(model_name, load_in_4bit=True)
220220
outputs = model.generate(inputs)
221221
```
222-
You can also load GGUF format model from Huggingface, and we will use [NeuralSpeed](https://github.com/intel/neural-speed) to accelerate the inference on CPUs.
222+
You can also load GGUF format model from Huggingface, we only support Q4_0 gguf format for now.
223223
```python
224224
from transformers import AutoTokenizer
225225
from intel_extension_for_transformers.transformers import AutoModelForCausalLM

0 commit comments

Comments
 (0)