Skip to content

Commit 84cc78e

Browse files
committed
add a page on local apps
1 parent 931924c commit 84cc78e

File tree

1 file changed

+93
-0
lines changed

1 file changed

+93
-0
lines changed

docs/hub/local-apps.md

Lines changed: 93 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,93 @@
1+
# Use AI Models Locally
2+
3+
You can run AI models from the Hub locally on your machine. This means that you can benefit from these advantages:
4+
5+
- **Privacy**: You won't be sending your data to a remote server.
6+
- **Speed**: Your hardware is the limiting factor, not the server or connection speed.
7+
- **Control**: You can configure models to your liking.
8+
- **Cost**: You can run models locally without paying for an API provider.
9+
10+
## How to Use Local Apps
11+
12+
Local apps are applications that can run Hugging Face models directly on your machine. To get started:
13+
14+
1. **Enable local apps** in your [Local Apps settings](https://huggingface.co/settings/local-apps).
15+
16+
![Local Apps](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/local-apps/settings.png)
17+
18+
1. **Choose a supported model** from the Hub by searching for it.
19+
20+
![Local Apps](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/local-apps/search_llamacpp.png)
21+
22+
3. **Select the local app** from the "Use this model" dropdown on the model page.
23+
24+
![Local Apps](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/local-apps/button.png)
25+
26+
4. **Copy and run** the provided command in your terminal.
27+
28+
![Local Apps](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/local-apps/command.png)
29+
30+
## Supported Local Apps
31+
32+
The best way to check if a local app is supported is to go to the Local Apps settings and see if the app is listed. Here is a quick overview of some of the most popular local apps:
33+
34+
<Tip>
35+
36+
To use these local apps, copy the snippets from the model card as above.
37+
38+
</Tip>
39+
40+
### Llama.cpp
41+
42+
Llama.cpp is a high-performance C/C++ library for running LLMs locally with optimized inference across different hardware. If you are running a CPU, this is the best option.
43+
44+
**Advantages:**
45+
- Extremely fast performance for CPU-based models
46+
- Low resource usage
47+
- Multiple interface options (CLI, server, Python library)
48+
- Hardware-optimized for CPU and GPU
49+
50+
To use Llama.cpp, navigate to the model card and click "Use this model" and copy the command.
51+
52+
```sh
53+
# Load and run the model:
54+
./llama-server -hf unsloth/gpt-oss-20b-GGUF:Q4_K_M
55+
```
56+
57+
### LM Studio
58+
59+
LM Studio is a desktop application that provides an easy way to download, run, and experiment with local LLMs.
60+
61+
**Advantages:**
62+
- Intuitive graphical interface
63+
- Built-in model browser
64+
- Developer tools and APIs
65+
- Free for personal and commercial use
66+
67+
Navigate to the model card and click "Use this model". LM Studio will open and you can start chatting through the interface.
68+
69+
### Jan
70+
71+
Jan is an open-source ChatGPT alternative that runs entirely offline with a user-friendly interface.
72+
73+
**Advantages:**
74+
- Complete privacy (all data stays local)
75+
- User-friendly GUI
76+
- Chat with documents and files
77+
- OpenAI-compatible API server
78+
79+
To use Jan, navigate to the model card and click "Use this model". Jan will open and you can start chatting through the interface.
80+
81+
### Ollama
82+
83+
Ollama is an application that lets you run large language models locally on your computer with a simple command-line interface.
84+
85+
**Advantages:**
86+
- Easy installation and setup
87+
- Direct integration with Hugging Face Hub
88+
89+
To use Ollama, navigate to the model card and click "Use this model" and copy the command.
90+
91+
```sh
92+
ollama run hf.co/unsloth/gpt-oss-20b-GGUF:Q4_K_M
93+
```

0 commit comments

Comments
 (0)