mirror of
https://github.com/mudler/LocalAI.git
synced 2026-04-18 13:58:07 -04:00
deterministic builds
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
@@ -2,6 +2,8 @@
|
||||
LLAMA_VERSION?=0fcb3760b2b9a3a496ef14621a7e4dad7a8df90f
|
||||
LLAMA_REPO?=https://github.com/ggerganov/llama.cpp
|
||||
|
||||
TURBOQUANT_VERSION?=8ad0f00e9a38df6c29fc10363341dde300f92ae4
|
||||
|
||||
CMAKE_ARGS?=
|
||||
BUILD_TYPE?=
|
||||
NATIVE?=false
|
||||
|
||||
@@ -2,8 +2,13 @@
|
||||
# Each source declares a fork whose commits are extracted as patches
|
||||
# and applied on top of upstream llama.cpp during the build.
|
||||
# See scripts/patch_utils/apply_patches.sh for the generic patch engine.
|
||||
#
|
||||
# version_var: Makefile variable with the pinned fork commit SHA
|
||||
# base_var: Makefile variable with the upstream base commit SHA
|
||||
# Both are read from version_file (relative to backend dir) to compute the diff.
|
||||
sources:
|
||||
- name: turboquant
|
||||
repo: https://github.com/TheTom/llama-cpp-turboquant.git
|
||||
branch: feature/turboquant-kv-cache
|
||||
upstream_repo: https://github.com/ggml-org/llama.cpp.git
|
||||
version_var: TURBOQUANT_VERSION
|
||||
base_var: LLAMA_VERSION
|
||||
version_file: Makefile
|
||||
|
||||
Reference in New Issue
Block a user