This website requires JavaScript.
Explore
Help
Sign In
happyz
/
llama.cpp
mirror of
https://github.com/ggerganov/llama.cpp.git
Watch
1
Star
0
Fork
You've already forked llama.cpp
0
Code
Issues
Packages
Projects
Releases
Wiki
Activity
8,492
Commits
550
Branches
5,823
Tags
1.5
GiB
cf4d2bd2f5
Commit Graph
1 Commits
Author
SHA1
Message
Date
Neo Zhang
213c4a0b81
[SYCL] supprt Flash Attention for fp32/fp16/Q4/Q5/Q8 (
#20190
)
...
* support flash-attention for fp32/fp16/Q4/Q5/Q8 * rm warining * update for JIT
2026-03-08 12:00:07 +08:00