We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 9121ea2 commit f84a37bCopy full SHA for f84a37b
ggml/src/ggml-cpu/llamafile/sgemm.cpp
@@ -250,7 +250,13 @@ template <> inline float32x4_t load(const ggml_fp16_t *p) {
250
251
#if defined(__VXE__) || defined(__VXE2__)
252
template <> inline float32x4_t load(const ggml_fp16_t * p) {
253
- return vec_xl(0, p);
+ float tmp[4];
254
+
255
+ for (int i = 0; i < 4; i++) {
256
+ tmp[i] = GGML_FP16_TO_FP32(x[i]);
257
+ }
258
259
+ return vec_xl(0, (const float *)(tmp));
260
}
261
template <> inline float32x4_t load(const float * p) {
262
return vec_xl(0, p);
0 commit comments