@@ -236,45 +236,19 @@ for more usage details.
236236
237237### Supported Models
238238
239- ` mlx-lm ` supports thousands of Hugging Face format LLMs. If the model you want to
240- run is not supported, file an
241- [ issue] ( https://github.com/ml-explore/mlx-lm/issues/new ) or better yet,
242- submit a pull request.
243-
244- Here are a few examples of Hugging Face models that work with this example:
245-
246- - [ mistralai/Mistral-7B-v0.1] ( https://huggingface.co/mistralai/Mistral-7B-v0.1 )
247- - [ meta-llama/Llama-2-7b-hf] ( https://huggingface.co/meta-llama/Llama-2-7b-hf )
248- - [ deepseek-ai/deepseek-coder-6.7b-instruct] ( https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct )
249- - [ 01-ai/Yi-6B-Chat] ( https://huggingface.co/01-ai/Yi-6B-Chat )
250- - [ microsoft/phi-2] ( https://huggingface.co/microsoft/phi-2 )
251- - [ mistralai/Mixtral-8x7B-Instruct-v0.1] ( https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1 )
252- - [ Qwen/Qwen-7B] ( https://huggingface.co/Qwen/Qwen-7B )
253- - [ pfnet/plamo-13b] ( https://huggingface.co/pfnet/plamo-13b )
254- - [ pfnet/plamo-13b-instruct] ( https://huggingface.co/pfnet/plamo-13b-instruct )
255- - [ stabilityai/stablelm-2-zephyr-1_6b] ( https://huggingface.co/stabilityai/stablelm-2-zephyr-1_6b )
256- - [ internlm/internlm2-7b] ( https://huggingface.co/internlm/internlm2-7b )
257- - [ tiiuae/falcon-mamba-7b-instruct] ( https://huggingface.co/tiiuae/falcon-mamba-7b-instruct )
258-
259- Most
260- [ Mistral] ( https://huggingface.co/models?library=transformers,safetensors&other=mistral&sort=trending ) ,
261- [ Llama] ( https://huggingface.co/models?library=transformers,safetensors&other=llama&sort=trending ) ,
262- [ Phi-2] ( https://huggingface.co/models?library=transformers,safetensors&other=phi&sort=trending ) ,
263- and
264- [ Mixtral] ( https://huggingface.co/models?library=transformers,safetensors&other=mixtral&sort=trending )
265- style models should work out of the box.
266-
267- For some models (such as ` Qwen ` and ` plamo ` ) the tokenizer requires you to
268- enable the ` trust_remote_code ` option. You can do this by passing
269- ` --trust-remote-code ` in the command line. If you don't specify the flag
270- explicitly, you will be prompted to trust remote code in the terminal when
271- running the model.
272-
273- For ` Qwen ` models you must also specify the ` eos_token ` . You can do this by
274- passing ` --eos-token "<|endoftext|>" ` in the command
275- line.
276-
277- These options can also be set in the Python API. For example:
239+ ` mlx-lm ` supports thousands of LLMs available on the Hugging Face Hub. If the
240+ model you want to run is not supported, file an
241+ [ issue] ( https://github.com/ml-explore/mlx-lm/issues/new ) or better yet, submit
242+ a pull request. Many supported models are available in various quantization
243+ formats in the [ MLX Community] ( https://huggingface.co/mlx-community ) Hugging
244+ Face organization.
245+
246+ For some models the tokenizer may require you to enable the ` trust_remote_code `
247+ option. You can do this by passing ` --trust-remote-code ` in the command line.
248+ If you don't specify the flag explicitly, you will be prompted to trust remote
249+ code in the terminal when running the model.
250+
251+ Tokenizer options can also be set in the Python API. For example:
278252
279253``` python
280254model, tokenizer = load(
0 commit comments