Go to file
2026-05-02 08:52:03 +03:00
.github ci : re-enable mac workflows (#0) 2026-04-14 15:27:40 +03:00
ci
cmake cmake : add FindNCCL.cmake (#0) 2026-04-21 10:59:04 +03:00
docs
examples examples : update to Q1_0 2026-05-01 11:53:27 +03:00
include CUDA: manage NCCL communicators in context (llama/21891) 2026-04-21 10:59:04 +03:00
scripts sync : llama.cpp 2026-05-02 08:41:22 +03:00
src ggml : remove obsolete rms_norm.wgsl (#0) 2026-05-02 08:52:03 +03:00
tests vulkan: Support asymmetric FA in coopmat2 path (llama/21753) 2026-05-02 08:41:22 +03:00
.editorconfig
.gitignore
.gitmodules
AUTHORS
CMakeLists.txt ggml : bump version to 0.10.2 (#1474) 2026-05-02 08:45:46 +03:00
CONTRIBUTING.md
ggml.pc.in
LICENSE
README.md
requirements.txt

ggml

Manifesto

Tensor library for machine learning

Note that this project is under active development.
Some of the development is currently happening in the llama.cpp and whisper.cpp repos

Features

  • Low-level cross-platform implementation
  • Integer quantization support
  • Broad hardware support
  • Automatic differentiation
  • ADAM and L-BFGS optimizers
  • No third-party dependencies
  • Zero memory allocations during runtime

Build

git clone https://github.com/ggml-org/ggml
cd ggml

# install python dependencies in a virtual environment
python3.10 -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt

# build the examples
mkdir build && cd build
cmake ..
cmake --build . --config Release -j 8

GPT inference (example)

# run the GPT-2 small 117M model
../examples/gpt-2/download-ggml-model.sh 117M
./bin/gpt-2-backend -m models/gpt-2-117M/ggml-model.bin -p "This is an example"

For more information, checkout the corresponding programs in the examples folder.

Resources