llama.cpp/docs/backend
Neo Zhang 213c4a0b81
[SYCL] supprt Flash Attention for fp32/fp16/Q4/Q5/Q8 (#20190)
* support flash-attention for fp32/fp16/Q4/Q5/Q8

* rm warining

* update for JIT
2026-03-08 12:00:07 +08:00
..
snapdragon
VirtGPU
BLIS.md
CANN.md
CUDA-FEDORA.md
OPENCL.md docs: add linux to index (#18907) 2026-01-18 18:03:35 +08:00
SYCL.md [SYCL] supprt Flash Attention for fp32/fp16/Q4/Q5/Q8 (#20190) 2026-03-08 12:00:07 +08:00
VirtGPU.md ggml-virtgpu: improve the reliability of the code (#19846) 2026-02-26 20:00:57 +08:00
zDNN.md ggml-zendnn : add ZenDNN backend for AMD CPUs (#17690) 2025-12-07 00:13:33 +08:00
ZenDNN.md