mirror of
https://github.com/mudler/LocalAI.git
synced 2026-04-19 06:20:42 -04:00
This PR adds patchset from the great work of @TheTom in https://github.com/TheTom/llama-cpp-turboquant and creates a pipeline that updates the patches against upstream automatically. It also creates necessary scaffolding for doing this with other patches sources. Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
10 lines
438 B
YAML
10 lines
438 B
YAML
# Patch sources for the llama-cpp backend.
|
|
# Each source declares a fork whose commits are extracted as patches
|
|
# and applied on top of upstream llama.cpp during the build.
|
|
# See scripts/patch_utils/apply_patches.sh for the generic patch engine.
|
|
sources:
|
|
- name: turboquant
|
|
repo: https://github.com/TheTom/llama-cpp-turboquant.git
|
|
branch: feature/turboquant-kv-cache
|
|
upstream_repo: https://github.com/ggml-org/llama.cpp.git
|