llama.cpp/ggml/src
Francis Couture-Harpin 8fbd59308b ggml-quants : attempt to fix Arm 32-bit support 2024-06-28 22:52:57 -04:00
..
ggml-cuda llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-sycl llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
kompute@4565194ed7 llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
kompute-shaders llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
vulkan-shaders llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
CMakeLists.txt ggml : add GGML_CUDA_USE_GRAPHS option, restore GGML_CUDA_FORCE_CUBLAS (cmake) (#8140) 2024-06-26 21:34:14 +02:00
ggml-alloc.c llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-backend-impl.h llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-backend.c llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-blas.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-common.h bitnet : replace 1.58b with b1.58, as in the paper 2024-06-28 20:38:12 -04:00
ggml-cuda.cu llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-impl.h ggml-quants : attempt to fix Arm 32-bit support 2024-06-28 22:52:57 -04:00
ggml-kompute.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-metal.m llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-metal.metal llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-quants.c ggml-quants : attempt to fix Arm 32-bit support 2024-06-28 22:52:57 -04:00
ggml-quants.h ggml-quants : 1.625 bpw ternary packing for BitNet 1.58b 2024-06-27 02:06:22 -04:00
ggml-rpc.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-sycl.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-vulkan-shaders.hpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml-vulkan.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
ggml.c ggml-quants : ARM NEON vec_dot for q2_2 and q1_3 2024-06-27 02:06:28 -04:00
sgemm.cpp llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
sgemm.h llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00