| .. |
|
baby-llama
|
|
|
|
batched
|
|
|
|
batched-bench
|
|
|
|
batched.swift
|
|
|
|
beam-search
|
|
|
|
benchmark
|
|
|
|
convert-llama2c-to-ggml
|
|
|
|
embedding
|
|
|
|
export-lora
|
|
|
|
finetune
|
|
|
|
gguf
|
|
|
|
gritlm
|
|
|
|
imatrix
|
|
|
|
infill
|
|
|
|
jeopardy
|
|
|
|
llama-bench
|
|
|
|
llama.android
|
|
|
|
llama.swiftui
|
|
|
|
llava
|
|
|
|
lookahead
|
|
|
|
lookup
|
|
|
|
main
|
|
|
|
main-cmake-pkg
|
|
|
|
parallel
|
|
|
|
passkey
|
|
|
|
perplexity
|
|
|
|
quantize
|
|
|
|
quantize-stats
|
|
|
|
save-load-state
|
|
|
|
server
|
|
|
|
simple
|
|
|
|
speculative
|
|
|
|
sycl
|
|
|
|
tokenize
|
|
|
|
train-text-from-scratch
|
|
|
|
alpaca.sh
|
|
|
|
base-translate.sh
|
|
|
|
chat-13B.bat
|
|
|
|
chat-13B.sh
|
|
|
|
chat-persistent.sh
|
|
|
|
chat-vicuna.sh
|
|
|
|
chat.sh
|
|
|
|
CMakeLists.txt
|
|
|
|
gpt4all.sh
|
|
|
|
json-schema-to-grammar.py
|
|
|
|
llama2-13b.sh
|
|
|
|
llama2.sh
|
|
|
|
llama.vim
|
|
|
|
llm.vim
|
|
|
|
make-ggml.py
|
|
|
|
Miku.sh
|
|
|
|
pydantic_models_to_grammar.py
|
|
|
|
pydantic-models-to-grammar-examples.py
|
|
|
|
reason-act.sh
|
|
|
|
server-embd.py
|
|
|
|
server-llama2-13B.sh
|
|
|