ik_llama.cpp/ggml/include
2026-01-22 13:20:23 +02:00
..
ggml-alloc.h Merge mainline llama.cpp (#3) 2024-07-27 07:55:01 +02:00
ggml-backend.h Async compute graph evaluation (2 or more GPUs) (#1089) 2025-12-27 08:18:06 +01:00
ggml-blas.h Merge mainline llama.cpp (#3) 2024-07-27 07:55:01 +02:00
ggml-cann.h Merge mainline llama.cpp (#3) 2024-07-27 07:55:01 +02:00
ggml-cpp.h Port mdmd from mainline + Qwen2/2.5-VL support (#798) 2025-09-27 08:45:29 +02:00
ggml-cuda.h CUDA: set compute parameters via command line arguments (#910) 2025-11-07 07:11:23 +02:00
ggml-kompute.h Merge mainline llama.cpp (#3) 2024-07-27 07:55:01 +02:00
ggml-metal.h Merge mainline - Aug 12 2024 (#17) 2024-08-12 15:14:32 +02:00
ggml-rpc.h server: improve speed of speculative decoding (#1119) 2026-01-10 08:01:22 +02:00
ggml-sycl.h Merge mainline llama.cpp (#3) 2024-07-27 07:55:01 +02:00
ggml-vulkan.h Vulkan: a fresh start (#608) 2025-07-15 08:03:13 +02:00
ggml.h Remove llamafile remnants (#1179) 2026-01-22 13:20:23 +02:00