Files
LocalAI/backend/cpp/llama-cpp/patches/sources.yaml
Ettore Di Giacinto e502e51d78 feat(llama.cpp): add turboquant support
This PR adds patchset from the great work of @TheTom in
https://github.com/TheTom/llama-cpp-turboquant and creates a pipeline
that updates the patches against upstream automatically.

It also creates necessary scaffolding for doing this with other patches
sources.

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2026-04-01 17:57:03 +00:00

10 lines
438 B
YAML

# Patch sources for the llama-cpp backend.
# Each source declares a fork whose commits are extracted as patches
# and applied on top of upstream llama.cpp during the build.
# See scripts/patch_utils/apply_patches.sh for the generic patch engine.
sources:
- name: turboquant
repo: https://github.com/TheTom/llama-cpp-turboquant.git
branch: feature/turboquant-kv-cache
upstream_repo: https://github.com/ggml-org/llama.cpp.git