Skip to content

Commit 5338385

Browse files
committed
Add Moore Threads GPU Support in README.md and replace ./main with whisper-cli
Signed-off-by: Xiaodong Ye <[email protected]>
1 parent eb6f702 commit 5338385

File tree

1 file changed

+23
-2
lines changed

1 file changed

+23
-2
lines changed

README.md

Lines changed: 23 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ High-performance inference of [OpenAI's Whisper](https://github.com/openai/whisp
2323
- [Efficient GPU support for NVIDIA](#nvidia-gpu-support)
2424
- [OpenVINO Support](#openvino-support)
2525
- [Ascend NPU Support](#ascend-npu-support)
26+
- [Moore Threads GPU Support](#moore-threads-gpu-support)
2627
- [C-style API](https://github.com/ggml-org/whisper.cpp/blob/master/include/whisper.h)
2728

2829
Supported platforms:
@@ -381,6 +382,25 @@ Run the inference examples as usual, for example:
381382
- If you have trouble with Ascend NPU device, please create a issue with **[CANN]** prefix/tag.
382383
- If you run successfully with your Ascend NPU device, please help update the table `Verified devices`.
383384

385+
## Moore Threads GPU support
386+
387+
With Moore Threads cards the processing of the models is done efficiently on the GPU via muBLAS and custom MUSA kernels.
388+
First, make sure you have installed `MUSA SDK rc3.1.1`: https://developer.mthreads.com/sdk/download/musa?equipment=&os=&driverVersion=&version=rc3.1.1
389+
390+
Now build `whisper.cpp` with MUSA support:
391+
392+
```
393+
cmake -B build -DGGML_MUSA=1
394+
cmake --build build -j --config Release
395+
```
396+
397+
or specify the architecture for your Moore Threads GPU. For example, if you have a MTT S80 GPU, you can specify the architecture as follows:
398+
399+
```
400+
cmake -B build -DGGML_MUSA=1 -DMUSA_ARCHITECTURES="21"
401+
cmake --build build -j --config Release
402+
```
403+
384404
## FFmpeg support (Linux only)
385405

386406
If you want to support more audio formats (such as Opus and AAC), you can turn on the `WHISPER_FFMPEG` build flag to enable FFmpeg integration.
@@ -425,6 +445,7 @@ We have two Docker images available for this project:
425445

426446
1. `ghcr.io/ggml-org/whisper.cpp:main`: This image includes the main executable file as well as `curl` and `ffmpeg`. (platforms: `linux/amd64`, `linux/arm64`)
427447
2. `ghcr.io/ggml-org/whisper.cpp:main-cuda`: Same as `main` but compiled with CUDA support. (platforms: `linux/amd64`)
448+
3. `ghcr.io/ggml-org/whisper.cpp:main-musa`: Same as `main` but compiled with MUSA support. (platforms: `linux/amd64`)
428449

429450
### Usage
430451

@@ -437,11 +458,11 @@ docker run -it --rm \
437458
docker run -it --rm \
438459
-v path/to/models:/models \
439460
-v path/to/audios:/audios \
440-
whisper.cpp:main "./main -m /models/ggml-base.bin -f /audios/jfk.wav"
461+
whisper.cpp:main "whisper-cli -m /models/ggml-base.bin -f /audios/jfk.wav"
441462
# transcribe an audio file in samples folder
442463
docker run -it --rm \
443464
-v path/to/models:/models \
444-
whisper.cpp:main "./main -m /models/ggml-base.bin -f ./samples/jfk.wav"
465+
whisper.cpp:main "whisper-cli -m /models/ggml-base.bin -f ./samples/jfk.wav"
445466
```
446467

447468
## Installing with Conan

0 commit comments

Comments
 (0)