Go to file
2026-03-16 22:40:32 +02:00
.github ci : disable AMX jobs 2026-03-16 22:40:32 +02:00
ci ggml : fix conv2d_dw SVE path (#1380) 2025-11-04 20:40:52 +02:00
cmake cmake : remove unused file (#1419) 2026-01-30 16:29:51 +02:00
docs Update gguf specification to synchronize the ggml_types declaration shown in the doc with the actual one. (#1342) 2025-09-16 13:42:24 +02:00
examples common : add nvfp4 (#0) 2026-03-15 21:50:13 +02:00
include ggml : restore ggml_type_sizef() to aboid major version bump (#1441) 2026-03-16 20:09:25 +02:00
scripts sync : whisper.cpp 2026-03-16 14:53:37 +02:00
src ggml : restore ggml_type_sizef() to aboid major version bump (#1441) 2026-03-16 20:09:25 +02:00
tests metal : add FA specialization for HSK = 320, HSV = 256 (llama/20549) 2026-03-15 21:50:13 +02:00
.editorconfig gguf : add file format specification (#302) 2023-11-01 19:01:49 +02:00
.gitignore gitignore : ignore idea files (#1339) 2025-09-09 13:17:07 +02:00
.gitmodules git : remove kompute submodule (#1300) 2025-07-12 16:12:49 +03:00
AUTHORS authors : update 2025-02-04 13:03:55 +02:00
CMakeLists.txt ggml : bump version to 0.9.8 (#1442) 2026-03-16 20:15:14 +02:00
CONTRIBUTING.md contrib : recommend PRs to llama.cpp (#1312) 2025-07-25 07:05:38 +03:00
ggml.pc.in pkg-config: include the new GGML_VERSION as a version (#1348) 2025-09-25 18:59:38 +02:00
LICENSE docs : Minor cleanups (llama/19252) 2026-02-07 10:37:38 +02:00
README.md readme : simplify 2026-03-16 14:54:53 +02:00
requirements.txt ci : update requirements.txt 2024-12-03 21:05:37 +02:00

ggml

Manifesto

Tensor library for machine learning

Note that this project is under active development.
Some of the development is currently happening in the llama.cpp and whisper.cpp repos

Features

  • Low-level cross-platform implementation
  • Integer quantization support
  • Broad hardware support
  • Automatic differentiation
  • ADAM and L-BFGS optimizers
  • No third-party dependencies
  • Zero memory allocations during runtime

Build

git clone https://github.com/ggml-org/ggml
cd ggml

# install python dependencies in a virtual environment
python3.10 -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt

# build the examples
mkdir build && cd build
cmake ..
cmake --build . --config Release -j 8

GPT inference (example)

# run the GPT-2 small 117M model
../examples/gpt-2/download-ggml-model.sh 117M
./bin/gpt-2-backend -m models/gpt-2-117M/ggml-model.bin -p "This is an example"

For more information, checkout the corresponding programs in the examples folder.

Resources