Skip to content

Commit 0dbf2e2

Browse files
authored
Merge pull request #26 from apicalshark/Temp
Temp
2 parents 6744d5c + cf80952 commit 0dbf2e2

File tree

204 files changed

+6504
-5182
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

204 files changed

+6504
-5182
lines changed

.clang-tidy

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,8 +17,10 @@ Checks: >
1717
-clang-analyzer-security.insecureAPI.DeprecatedOrUnsafeBufferHandling,
1818
performance-*,
1919
portability-*,
20+
-portability-simd-intrinsics,
2021
misc-*,
2122
-misc-const-correctness,
2223
-misc-non-private-member-variables-in-classes,
2324
-misc-no-recursion,
25+
-misc-use-anonymous-namespace,
2426
FormatStyle: none

.devops/full-musa.Dockerfile

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,9 @@ ARG BASE_MUSA_DEV_CONTAINER=mthreads/musa:${MUSA_VERSION}-devel-ubuntu${UBUNTU_V
66

77
FROM ${BASE_MUSA_DEV_CONTAINER} AS build
88

9+
# MUSA architecture to build for (defaults to all supported archs)
10+
ARG MUSA_DOCKER_ARCH=default
11+
912
RUN apt-get update && \
1013
apt-get install -y build-essential cmake python3 python3-pip git libcurl4-openssl-dev libgomp1
1114

@@ -19,7 +22,11 @@ WORKDIR /app
1922

2023
COPY . .
2124

22-
RUN cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON -DLLAMA_CURL=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
25+
# Use the default MUSA archs if not specified
26+
RUN if [ "${MUSA_DOCKER_ARCH}" != "default" ]; then \
27+
export CMAKE_ARGS="-DMUSA_ARCHITECTURES=${MUSA_DOCKER_ARCH}"; \
28+
fi && \
29+
cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON -DLLAMA_CURL=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
2330
cmake --build build --config Release -j$(nproc) && \
2431
cp build/bin/* .
2532

.devops/llama-cli-musa.Dockerfile

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,14 +8,21 @@ ARG BASE_MUSA_RUN_CONTAINER=mthreads/musa:${MUSA_VERSION}-runtime-ubuntu${UBUNTU
88

99
FROM ${BASE_MUSA_DEV_CONTAINER} AS build
1010

11+
# MUSA architecture to build for (defaults to all supported archs)
12+
ARG MUSA_DOCKER_ARCH=default
13+
1114
RUN apt-get update && \
1215
apt-get install -y build-essential git cmake
1316

1417
WORKDIR /app
1518

1619
COPY . .
1720

18-
RUN cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
21+
# Use the default MUSA archs if not specified
22+
RUN if [ "${MUSA_DOCKER_ARCH}" != "default" ]; then \
23+
export CMAKE_ARGS="-DMUSA_ARCHITECTURES=${MUSA_DOCKER_ARCH}"; \
24+
fi && \
25+
cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
1926
cmake --build build --config Release --target llama-cli -j$(nproc) && \
2027
mkdir -p /app/lib && \
2128
find build -name "*.so" -exec cp {} /app/lib \;

.devops/llama-server-musa.Dockerfile

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,14 +8,21 @@ ARG BASE_MUSA_RUN_CONTAINER=mthreads/musa:${MUSA_VERSION}-runtime-ubuntu${UBUNTU
88

99
FROM ${BASE_MUSA_DEV_CONTAINER} AS build
1010

11+
# MUSA architecture to build for (defaults to all supported archs)
12+
ARG MUSA_DOCKER_ARCH=default
13+
1114
RUN apt-get update && \
1215
apt-get install -y build-essential git cmake libcurl4-openssl-dev
1316

1417
WORKDIR /app
1518

1619
COPY . .
1720

18-
RUN cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON -DLLAMA_CURL=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
21+
# Use the default MUSA archs if not specified
22+
RUN if [ "${MUSA_DOCKER_ARCH}" != "default" ]; then \
23+
export CMAKE_ARGS="-DMUSA_ARCHITECTURES=${MUSA_DOCKER_ARCH}"; \
24+
fi && \
25+
cmake -B build -DGGML_NATIVE=OFF -DGGML_MUSA=ON -DLLAMA_CURL=ON ${CMAKE_ARGS} -DCMAKE_EXE_LINKER_FLAGS=-Wl,--allow-shlib-undefined . && \
1926
cmake --build build --config Release --target llama-server -j$(nproc) && \
2027
mkdir -p /app/lib && \
2128
find build -name "*.so" -exec cp {} /app/lib \;

.devops/nix/python-scripts.nix

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ let
3434

3535
# server tests
3636
openai
37-
behave
37+
pytest
3838
prometheus-client
3939
];
4040
in

.github/ISSUE_TEMPLATE/010-bug-compilation.yml

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,8 @@ body:
2424
- type: dropdown
2525
id: operating-system
2626
attributes:
27-
label: Which operating systems do you know to be affected?
27+
label: Operating systems
28+
description: Which operating systems do you know to be affected?
2829
multiple: true
2930
options:
3031
- Linux
@@ -41,14 +42,17 @@ body:
4142
description: Which GGML backends do you know to be affected?
4243
options: [AMX, BLAS, CPU, CUDA, HIP, Kompute, Metal, Musa, RPC, SYCL, Vulkan]
4344
multiple: true
45+
validations:
46+
required: true
4447
- type: textarea
45-
id: steps_to_reproduce
48+
id: info
4649
attributes:
47-
label: Steps to Reproduce
50+
label: Problem description & steps to reproduce
4851
description: >
49-
Please tell us how to reproduce the bug and any additional information that you think could be useful for fixing it.
52+
Please give us a summary of the problem and tell us how to reproduce it.
5053
If you can narrow down the bug to specific compile flags, that information would be very much appreciated by us.
5154
placeholder: >
55+
I'm trying to compile llama.cpp with CUDA support on a fresh install of Ubuntu and get error XY.
5256
Here are the exact commands that I used: ...
5357
validations:
5458
required: true

.github/ISSUE_TEMPLATE/011-bug-results.yml

Lines changed: 9 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,8 @@ body:
2626
- type: dropdown
2727
id: operating-system
2828
attributes:
29-
label: Which operating systems do you know to be affected?
29+
label: Operating systems
30+
description: Which operating systems do you know to be affected?
3031
multiple: true
3132
options:
3233
- Linux
@@ -43,6 +44,8 @@ body:
4344
description: Which GGML backends do you know to be affected?
4445
options: [AMX, BLAS, CPU, CUDA, HIP, Kompute, Metal, Musa, RPC, SYCL, Vulkan]
4546
multiple: true
47+
validations:
48+
required: true
4649
- type: textarea
4750
id: hardware
4851
attributes:
@@ -55,20 +58,20 @@ body:
5558
- type: textarea
5659
id: model
5760
attributes:
58-
label: Model
61+
label: Models
5962
description: >
60-
Which model at which quantization were you using when encountering the bug?
63+
Which model(s) at which quantization were you using when encountering the bug?
6164
If you downloaded a GGUF file off of Huggingface, please provide a link.
6265
placeholder: >
6366
e.g. Meta LLaMA 3.1 Instruct 8b q4_K_M
6467
validations:
6568
required: false
6669
- type: textarea
67-
id: steps_to_reproduce
70+
id: info
6871
attributes:
69-
label: Steps to Reproduce
72+
label: Problem description & steps to reproduce
7073
description: >
71-
Please tell us how to reproduce the bug and any additional information that you think could be useful for fixing it.
74+
Please give us a summary of the problem and tell us how to reproduce it.
7275
If you can narrow down the bug to specific hardware, compile flags, or command line arguments,
7376
that information would be very much appreciated by us.
7477
placeholder: >

.github/ISSUE_TEMPLATE/019-bug-misc.yml

Lines changed: 13 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ body:
1414
id: version
1515
attributes:
1616
label: Name and Version
17-
description: Which version of our software are you running? (use `--version` to get a version string)
17+
description: Which version of our software is affected? (You can use `--version` to get a version string.)
1818
placeholder: |
1919
$./llama-cli --version
2020
version: 2999 (42b4109e)
@@ -24,7 +24,8 @@ body:
2424
- type: dropdown
2525
id: operating-system
2626
attributes:
27-
label: Which operating systems do you know to be affected?
27+
label: Operating systems
28+
description: Which operating systems do you know to be affected?
2829
multiple: true
2930
options:
3031
- Linux
@@ -33,36 +34,38 @@ body:
3334
- BSD
3435
- Other? (Please let us know in description)
3536
validations:
36-
required: true
37+
required: false
3738
- type: dropdown
3839
id: module
3940
attributes:
4041
label: Which llama.cpp modules do you know to be affected?
4142
multiple: true
4243
options:
44+
- Documentation/Github
4345
- libllama (core library)
4446
- llama-cli
4547
- llama-server
4648
- llama-bench
4749
- llama-quantize
4850
- Python/Bash scripts
51+
- Test code
4952
- Other (Please specify in the next section)
5053
validations:
51-
required: true
54+
required: false
5255
- type: textarea
53-
id: steps_to_reproduce
56+
id: info
5457
attributes:
55-
label: Steps to Reproduce
58+
label: Problem description & steps to reproduce
5659
description: >
57-
Please tell us how to reproduce the bug and any additional information that you think could be useful for fixing it.
60+
Please give us a summary of the problem and tell us how to reproduce it (if applicable).
5861
validations:
5962
required: true
6063
- type: textarea
6164
id: first_bad_commit
6265
attributes:
6366
label: First Bad Commit
6467
description: >
65-
If the bug was not present on an earlier version: when did it start appearing?
68+
If the bug was not present on an earlier version and it's not trivial to track down: when did it start appearing?
6669
If possible, please do a git bisect and identify the exact commit that introduced the bug.
6770
validations:
6871
required: false
@@ -71,8 +74,8 @@ body:
7174
attributes:
7275
label: Relevant log output
7376
description: >
74-
Please copy and paste any relevant log output, including the command that you entered and any generated text.
77+
If applicable, please copy and paste any relevant log output, including the command that you entered and any generated text.
7578
This will be automatically formatted into code, so no need for backticks.
7679
render: shell
7780
validations:
78-
required: true
81+
required: false

.github/labeler.yml

Lines changed: 5 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -3,19 +3,18 @@ Kompute:
33
- changed-files:
44
- any-glob-to-any-file:
55
- ggml/include/ggml-kompute.h
6-
- ggml/src/ggml-kompute.cpp
6+
- ggml/src/ggml-kompute/**
77
- README-kompute.md
88
Apple Metal:
99
- changed-files:
1010
- any-glob-to-any-file:
1111
- ggml/include/ggml-metal.h
12-
- ggml/src/ggml-metal.cpp
12+
- ggml/src/ggml-metal/**
1313
- README-metal.md
1414
SYCL:
1515
- changed-files:
1616
- any-glob-to-any-file:
1717
- ggml/include/ggml-sycl.h
18-
- ggml/src/ggml-sycl.cpp
1918
- ggml/src/ggml-sycl/**
2019
- docs/backend/SYCL.md
2120
- examples/sycl/**
@@ -27,8 +26,8 @@ Nvidia GPU:
2726
Vulkan:
2827
- changed-files:
2928
- any-glob-to-any-file:
30-
- ggml/ggml_vk_generate_shaders.py
31-
- ggml/src/ggml-vulkan*
29+
- ggml/include/ggml-vulkan.h
30+
- ggml/src/ggml-vulkan/**
3231
documentation:
3332
- changed-files:
3433
- any-glob-to-any-file:
@@ -75,11 +74,7 @@ server:
7574
ggml:
7675
- changed-files:
7776
- any-glob-to-any-file:
78-
- ggml/include/ggml*.h
79-
- ggml/src/ggml*.c
80-
- ggml/src/ggml*.cpp
81-
- ggml/src/ggml*.h
82-
- ggml-cuda/**
77+
- ggml/**
8378
nix:
8479
- changed-files:
8580
- any-glob-to-any-file:

0 commit comments

Comments
 (0)