mirror of
https://github.com/ggerganov/llama.cpp
synced 2026-03-02 21:29:35 +01:00
* llama/ggml: add LLM training support more compact progress bar llama_save_model_to_file llama_opt_param_filter ggml_graph_dup force_grads refactor ggml_opt, fix test-opt * remove logits_all * refactor CUDA implementation for ACC * reset graph at beginning of opt period
6 lines
235 B
CMake
6 lines
235 B
CMake
set(TARGET llama-finetune)
|
|
add_executable(${TARGET} finetune.cpp)
|
|
install(TARGETS ${TARGET} RUNTIME)
|
|
target_link_libraries(${TARGET} PRIVATE common llama ${CMAKE_THREAD_LIBS_INIT})
|
|
target_compile_features(${TARGET} PRIVATE cxx_std_11)
|