mirror of
https://github.com/ggerganov/llama.cpp
synced 2026-03-12 02:00:41 +01:00
* use hipblas based on cublas * Update Makefile for the Cuda kernels * Expand arch list and make it overrideable * Fix multi GPU on multiple amd architectures with rocblas_initialize() (#5) * add hipBLAS to README * new build arg LLAMA_CUDA_MMQ_Y * fix half2 decomposition * Add intrinsics polyfills for AMD * AMD assembly optimized __dp4a * Allow overriding CC_TURING * use "ROCm" instead of "CUDA" * ignore all build dirs * Add Dockerfiles * fix llama-bench * fix -nommq help for non CUDA/HIP --------- Co-authored-by: YellowRoseCx <80486540+YellowRoseCx@users.noreply.github.com> Co-authored-by: ardfork <134447697+ardfork@users.noreply.github.com> Co-authored-by: funnbot <22226942+funnbot@users.noreply.github.com> Co-authored-by: Engininja2 <139037756+Engininja2@users.noreply.github.com> Co-authored-by: Kerfuffle <44031344+KerfuffleV2@users.noreply.github.com> Co-authored-by: jammm <2500920+jammm@users.noreply.github.com> Co-authored-by: jdecourval <7315817+jdecourval@users.noreply.github.com> |
||
|---|---|---|
| .. | ||
| baby-llama | ||
| benchmark | ||
| convert-llama2c-to-ggml | ||
| embd-input | ||
| embedding | ||
| gguf | ||
| gptneox-wip | ||
| jeopardy | ||
| llama-bench | ||
| main | ||
| metal | ||
| perplexity | ||
| quantize | ||
| quantize-stats | ||
| save-load-state | ||
| server | ||
| simple | ||
| train-text-from-scratch | ||
| alpaca.sh | ||
| chat-13B.bat | ||
| chat-13B.sh | ||
| chat-persistent.sh | ||
| chat-vicuna.sh | ||
| chat.sh | ||
| CMakeLists.txt | ||
| gpt4all.sh | ||
| json-schema-to-grammar.py | ||
| llama2-13b.sh | ||
| llama2.sh | ||
| llama.vim | ||
| llm.vim | ||
| make-ggml.py | ||
| Miku.sh | ||
| reason-act.sh | ||
| server-llama2-13B.sh | ||