* chore(llama.cpp): bump to 1ec7ba0c14f33f17e980daeeda5f35b225d41994
Picks up the upstream `spec : parallel drafting support` change
(ggml-org/llama.cpp#22838) which reshapes the speculative-decoding API
and `server_context_impl`.
Adapt the grpc-server wrapper accordingly:
* `common_params_speculative::type` (single enum) became `types`
(`std::vector<common_speculative_type>`). Update both the
"default to draft when a draft model is set" branch and the
`spec_type`/`speculative_type` option parser. The parser now also
tolerates comma-separated lists, mirroring the upstream
`common_speculative_types_from_names` semantics.
* `common_params_speculative_draft::n_ctx` is gone (draft now shares
the target context size). Keep the `draft_ctx_size` option name for
backward compatibility and ignore the value rather than failing.
* `server_context_impl::model` was renamed to `model_tgt`; update the
two reranker / model-metadata call sites.
Replaces #9763. Builds cleanly under the linux/amd64 cpu-llama-cpp
target locally.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* feat(llama-cpp): expose new speculative-decoding option keys
Upstream `spec : parallel drafting support` (ggml-org/llama.cpp#22838)
adds the `ngram_mod`, `ngram_map_k`, and `ngram_map_k4v` speculative
families and beefs up the draft-model knobs. The previous bump only
adapted the API; this exposes the new fields through the grpc-server
options dictionary so model configs can drive them.
New `options:` keys (all under `backend: llama-cpp`):
ngram_mod (`ngram_mod` type):
spec_ngram_mod_n_min / spec_ngram_mod_n_max / spec_ngram_mod_n_match
ngram_map_k (`ngram_map_k` type):
spec_ngram_map_k_size_n / spec_ngram_map_k_size_m / spec_ngram_map_k_min_hits
ngram_map_k4v (`ngram_map_k4v` type):
spec_ngram_map_k4v_size_n / spec_ngram_map_k4v_size_m /
spec_ngram_map_k4v_min_hits
ngram lookup caches (`ngram_cache` type):
spec_lookup_cache_static / lookup_cache_static
spec_lookup_cache_dynamic / lookup_cache_dynamic
Draft-model tuning (active when `spec_type` is `draft`):
draft_cache_type_k / spec_draft_cache_type_k
draft_cache_type_v / spec_draft_cache_type_v
draft_threads / spec_draft_threads
draft_threads_batch / spec_draft_threads_batch
draft_cpu_moe / spec_draft_cpu_moe (bool flag)
draft_n_cpu_moe / spec_draft_n_cpu_moe (first N MoE layers on CPU)
draft_override_tensor / spec_draft_override_tensor
(comma-separated <tensor regex>=<buffer type>; re-implements upstream's
static parse_tensor_buffer_overrides since it isn't exported)
`spec_type` already accepted comma-separated lists after the previous
commit, matching upstream's `common_speculative_types_from_names`.
Docs: refresh `docs/content/advanced/model-configuration.md` with
per-family tables and a note about multi-type chaining.
Builds locally with `make docker-build-llama-cpp` (linux/amd64
cpu-llama-cpp AVX variant).
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* fix(turboquant): bridge new llama.cpp spec API to the legacy fork layout
The previous commits in this series adapted backend/cpp/llama-cpp/grpc-server.cpp
to the post-#22838 (parallel drafting) llama.cpp API. The turboquant build
reuses the same grpc-server.cpp through backend/cpp/turboquant/Makefile,
which copies it into turboquant-<flavor>-build/ and runs patch-grpc-server.sh
on the copy. The fork branched before the API refactor, so it errors out on:
* `ctx_server.impl->model_tgt` (fork still has `model`)
* `params.speculative.{ngram_mod,ngram_map_k,ngram_map_k4v,ngram_cache}.*`
(none of these sub-structs exist in the fork)
* `params.speculative.draft.{cache_type_k/v, cpuparams[, _batch].n_threads,
tensor_buft_overrides}` (fork uses the pre-#22397 flat layout)
* `params.speculative.types` vector / `common_speculative_types_from_names`
(fork has a scalar `type` and only the singular helper)
Approach:
1. backend/cpp/llama-cpp/grpc-server.cpp: introduce a single feature switch
`LOCALAI_LEGACY_LLAMA_CPP_SPEC`. When defined, the two `speculative.type[s]`
discriminations (the "default to draft when a draft model is set" branch
and the `spec_type` / `speculative_type` option parser) fall back to the
singular scalar form, and the entire new-option block (ngram_mod / map_k
/ map_k4v / ngram_cache / draft.{cache_type_*, cpuparams*,
tensor_buft_overrides}) is preprocessed out. The macro is *not* defined
in the source tree — stock llama-cpp builds get the full new API.
2. backend/cpp/turboquant/patch-grpc-server.sh: two new patch steps applied
to the per-flavor build copy at turboquant-<flavor>-build/grpc-server.cpp:
- substitute `ctx_server.impl->model_tgt` -> `ctx_server.impl->model`
- inject `#define LOCALAI_LEGACY_LLAMA_CPP_SPEC 1` before the first
`#include`, so the guarded blocks above drop out for the fork build.
Both patches are idempotent and follow the existing sed/awk pattern in
this script (KV cache types, `get_media_marker`, flat speculative
renames). Stock llama-cpp's `grpc-server.cpp` is never touched.
Drop both legacy patches once the turboquant fork rebases past
ggml-org/llama.cpp#22397 / #22838.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* fix(turboquant): close draft_ctx_size brace inside legacy guard
The previous turboquant fix wrapped the new option-handler blocks in
`#ifndef LOCALAI_LEGACY_LLAMA_CPP_SPEC ... #endif` but placed the guard
in the middle of an `else if` chain — the `} else if` openings of the
new blocks were responsible for closing the previous block's brace.
With the macro defined the new blocks vanish, draft_ctx_size's `{`
loses its closer, the for-loop's `}` is consumed instead, and the
file ends with a stray opening brace — clang reports it as
`function-definition is not allowed here before '{'` on the next
top-level `int main(...)` and `expected '}' at end of input`.
Move the chain split inside the draft_ctx_size branch:
} else if (... "draft_ctx_size") {
// ...
#ifdef LOCALAI_LEGACY_LLAMA_CPP_SPEC
} // legacy: chain ends here
#else
} else if (... "spec_ngram_mod_n_min") { // modern: chain continues
...
} else if (... "draft_override_tensor") {
...
} // closes last branch
#endif
} // closes for-loop
Brace count is now balanced under both preprocessor branches (verified
with `tr -cd '{' | wc -c` against the patched and unpatched outputs).
Local `make docker-build-turboquant` builds the linux/amd64 cpu-llama-cpp
`turboquant-avx` variant cleanly.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
* fix(ci): forward AMDGPU_TARGETS into Dockerfile.turboquant builder-prebuilt
Dockerfile.turboquant's `builder-prebuilt` stage was missing the
`ARG AMDGPU_TARGETS` / `ENV AMDGPU_TARGETS=${AMDGPU_TARGETS}` pair that
`builder-fromsource` already has (and that `Dockerfile.llama-cpp`
mirrors across both stages). When CI uses the prebuilt base image
(quay.io/go-skynet/ci-cache:base-grpc-*, the common path) the build-arg
passed by the workflow never reaches the env inside the compile stage.
backend/cpp/llama-cpp/Makefile:38 (introduced by #9626) errors out on
hipblas builds when AMDGPU_TARGETS is empty, and the turboquant
Makefile reuses backend/cpp/llama-cpp via a sibling build dir, so the
same check fires from turboquant-fallback under BUILD_TYPE=hipblas:
Makefile:38: *** AMDGPU_TARGETS is empty — set it to a comma-separated
list of gfx targets e.g. gfx1100,gfx1101. Stop.
make: *** [Makefile:66: turboquant-fallback] Error 2
The bug is latent on master because the docker layer cache stays warm
across builds — the compile step rarely re-runs from scratch. The
llama.cpp bump in this PR invalidates the cache, so the missing env var
becomes load-bearing and the hipblas turboquant CI job fails.
Mirror the existing pattern from Dockerfile.llama-cpp.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Co-authored-by: Ettore Di Giacinto <mudler@localai.io>
29 KiB
+++ disableToc = false title = "Model Configuration" weight = 23 url = '/advanced/model-configuration' +++
LocalAI uses YAML configuration files to define model parameters, templates, and behavior. This page provides a complete reference for all available configuration options.
Overview
Model configuration files allow you to:
- Define default parameters (temperature, top_p, etc.)
- Configure prompt templates
- Specify backend settings
- Set up function calling
- Configure GPU and memory options
- And much more
Configuration File Locations
You can create model configuration files in several ways:
- Individual YAML files in the models directory (e.g.,
models/gpt-3.5-turbo.yaml) - Single config file with multiple models using
--models-config-fileorLOCALAI_MODELS_CONFIG_FILE - Remote URLs - specify a URL to a YAML configuration file at startup
Example: Basic Configuration
name: gpt-3.5-turbo
parameters:
model: luna-ai-llama2-uncensored.ggmlv3.q5_K_M.bin
temperature: 0.3
context_size: 512
threads: 10
backend: llama-stable
template:
completion: completion
chat: chat
Example: Multiple Models in One File
When using --models-config-file, you can define multiple models as a list:
- name: model1
parameters:
model: model1.bin
context_size: 512
backend: llama-stable
- name: model2
parameters:
model: model2.bin
context_size: 1024
backend: llama-stable
Core Configuration Fields
Basic Model Settings
| Field | Type | Description | Example |
|---|---|---|---|
name |
string | Model name, used to identify the model in API calls | gpt-3.5-turbo |
backend |
string | Backend to use (e.g. llama-cpp, vllm, diffusers, whisper) |
llama-cpp |
description |
string | Human-readable description of the model | A conversational AI model |
usage |
string | Usage instructions or notes | Best for general conversation |
Model File and Downloads
| Field | Type | Description |
|---|---|---|
parameters.model |
string | Path to the model file (relative to models directory) or URL |
download_files |
array | List of files to download. Each entry has filename, uri, and optional sha256 |
Example:
parameters:
model: my-model.gguf
download_files:
- filename: my-model.gguf
uri: https://example.com/model.gguf
sha256: abc123...
Parameters Section
The parameters section contains all OpenAI-compatible request parameters and model-specific options.
OpenAI-Compatible Parameters
These settings will be used as defaults for all the API calls to the model.
| Field | Type | Default | Description |
|---|---|---|---|
temperature |
float | 0.9 |
Sampling temperature (0.0-2.0). Higher values make output more random |
top_p |
float | 0.95 |
Nucleus sampling: consider tokens with top_p probability mass |
top_k |
int | 40 |
Consider only the top K most likely tokens |
max_tokens |
int | 0 |
Maximum number of tokens to generate (0 = unlimited) |
frequency_penalty |
float | 0.0 |
Penalty for token frequency (-2.0 to 2.0) |
presence_penalty |
float | 0.0 |
Penalty for token presence (-2.0 to 2.0) |
repeat_penalty |
float | 1.1 |
Penalty for repeating tokens |
repeat_last_n |
int | 64 |
Number of previous tokens to consider for repeat penalty |
seed |
int | -1 |
Random seed (omit for random) |
echo |
bool | false |
Echo back the prompt in the response |
n |
int | 1 |
Number of completions to generate |
logprobs |
bool/int | false |
Return log probabilities of tokens |
top_logprobs |
int | 0 |
Number of top logprobs to return per token (0-20) |
logit_bias |
map | {} |
Map of token IDs to bias values (-100 to 100) |
typical_p |
float | 1.0 |
Typical sampling parameter |
tfz |
float | 1.0 |
Tail free z parameter |
keep |
int | 0 |
Number of tokens to keep from the prompt |
Language and Translation
| Field | Type | Description |
|---|---|---|
language |
string | Language code for transcription/translation |
translate |
bool | Whether to translate audio transcription |
Custom Parameters
| Field | Type | Description |
|---|---|---|
batch |
int | Batch size for processing |
ignore_eos |
bool | Ignore end-of-sequence tokens |
negative_prompt |
string | Negative prompt for image generation |
rope_freq_base |
float32 | RoPE frequency base |
rope_freq_scale |
float32 | RoPE frequency scale |
negative_prompt_scale |
float32 | Scale for negative prompt |
tokenizer |
string | Tokenizer to use (RWKV) |
LLM Configuration
These settings apply to most LLM backends (llama.cpp, vLLM, etc.):
Performance Settings
| Field | Type | Default | Description |
|---|---|---|---|
threads |
int | processor count |
Number of threads for parallel computation |
context_size |
int | 512 |
Maximum context size (number of tokens) |
f16 |
bool | false |
Enable 16-bit floating point precision (GPU acceleration) |
gpu_layers |
int | 0 |
Number of layers to offload to GPU (0 = CPU only) |
Memory Management
| Field | Type | Default | Description |
|---|---|---|---|
mmap |
bool | true |
Use memory mapping for model loading (faster, less RAM) |
mmlock |
bool | false |
Lock model in memory (prevents swapping) |
low_vram |
bool | false |
Use minimal VRAM mode |
no_kv_offloading |
bool | false |
Disable KV cache offloading |
GPU Configuration
| Field | Type | Description |
|---|---|---|
tensor_split |
string | Comma-separated GPU memory allocation (e.g., "0.8,0.2" for 80%/20%) |
main_gpu |
string | Main GPU identifier for multi-GPU setups |
cuda |
bool | Explicitly enable/disable CUDA |
Sampling and Generation
| Field | Type | Default | Description |
|---|---|---|---|
mirostat |
int | 0 |
Mirostat sampling mode (0=disabled, 1=Mirostat, 2=Mirostat 2.0) |
mirostat_tau |
float | 5.0 |
Mirostat target entropy |
mirostat_eta |
float | 0.1 |
Mirostat learning rate |
LoRA Configuration
| Field | Type | Description |
|---|---|---|
lora_adapter |
string | Path to LoRA adapter file |
lora_base |
string | Base model for LoRA |
lora_scale |
float32 | LoRA scale factor |
lora_adapters |
array | Multiple LoRA adapters |
lora_scales |
array | Scales for multiple LoRA adapters |
Advanced Options
| Field | Type | Description |
|---|---|---|
no_mulmatq |
bool | Disable matrix multiplication queuing |
draft_model |
string | Draft model GGUF file for speculative decoding (see Speculative Decoding) |
n_draft |
int32 | Maximum number of draft tokens per speculative step (default: 16) |
quantization |
string | Quantization format |
load_format |
string | Model load format |
numa |
bool | Enable NUMA (Non-Uniform Memory Access) |
rms_norm_eps |
float32 | RMS normalization epsilon |
ngqa |
int32 | Natural question generation parameter |
rope_scaling |
string | RoPE scaling configuration |
type |
string | Model type/architecture |
grammar |
string | Grammar file path for constrained generation |
YARN Configuration
YARN (Yet Another RoPE extensioN) settings for context extension:
| Field | Type | Description |
|---|---|---|
yarn_ext_factor |
float32 | YARN extension factor |
yarn_attn_factor |
float32 | YARN attention factor |
yarn_beta_fast |
float32 | YARN beta fast parameter |
yarn_beta_slow |
float32 | YARN beta slow parameter |
Speculative Decoding
Speculative decoding speeds up text generation by predicting multiple tokens ahead and verifying them in a single forward pass. The output is identical to normal decoding — only faster. This feature is only available with the llama-cpp backend.
There are two approaches:
Draft Model Speculative Decoding
Uses a smaller, faster model from the same model family to draft candidate tokens, which the main model then verifies. Requires a separate GGUF file for the draft model.
name: my-model
backend: llama-cpp
parameters:
model: large-model.gguf
draft_model: small-draft-model.gguf
n_draft: 8
options:
- spec_p_min:0.8
- draft_gpu_layers:99
N-gram Self-Speculative Decoding
Uses patterns from the token history to predict future tokens — no extra model required. Works well for repetitive or structured output (code, JSON, lists).
name: my-model
backend: llama-cpp
parameters:
model: my-model.gguf
options:
- spec_type:ngram_simple
- spec_n_max:16
Speculative Decoding Options
These are set via the options: array in the model configuration (format: key:value):
Common options
| Option | Type | Default | Description |
|---|---|---|---|
spec_type / speculative_type |
string | none |
Speculative decoding type, or comma-separated list to chain multiple (see table below) |
spec_n_max / draft_max |
int | 16 | Maximum number of tokens to draft per step |
spec_n_min / draft_min |
int | 0 | Minimum draft tokens required to use speculation |
spec_p_min / draft_p_min |
float | 0.75 | Minimum probability threshold for greedy acceptance |
spec_p_split |
float | 0.1 | Split probability for tree-based branching |
Draft-model options (apply when spec_type=draft, i.e. a draft_model is configured)
| Option | Type | Default | Description |
|---|---|---|---|
draft_gpu_layers |
int | -1 | GPU layers for the draft model (-1 = use default) |
draft_threads / spec_draft_threads |
int | same as main | Threads used by the draft model (<= 0 = hardware concurrency) |
draft_threads_batch / spec_draft_threads_batch |
int | same as draft_threads |
Threads used by the draft model during batch / prompt processing |
draft_cache_type_k / spec_draft_cache_type_k |
string | f16 |
KV cache K data type for the draft model (same values as cache_type_k) |
draft_cache_type_v / spec_draft_cache_type_v |
string | f16 |
KV cache V data type for the draft model |
draft_cpu_moe / spec_draft_cpu_moe |
bool | false | Keep all MoE expert weights of the draft model on CPU |
draft_n_cpu_moe / spec_draft_n_cpu_moe |
int | 0 | Keep MoE expert weights of the first N draft-model layers on CPU |
draft_override_tensor / spec_draft_override_tensor |
string | "" | Comma-separated <tensor regex>=<buffer type> overrides for the draft model |
draft_ctx_size |
int | (ignored) | Deprecated upstream: the draft now shares the target context size. Accepted for backward compatibility but has no effect. |
ngram_simple options (used when spec_type includes ngram_simple)
| Option | Type | Default | Description |
|---|---|---|---|
spec_ngram_size_n / ngram_size_n |
int | 12 | N-gram lookup size |
spec_ngram_size_m / ngram_size_m |
int | 48 | M-gram proposal size |
spec_ngram_min_hits / ngram_min_hits |
int | 1 | Minimum hits for accepting n-gram proposals |
ngram_mod options (used when spec_type includes ngram_mod)
| Option | Type | Default | Description |
|---|---|---|---|
spec_ngram_mod_n_min |
int | 48 | Minimum number of ngram tokens to use |
spec_ngram_mod_n_max |
int | 64 | Maximum number of ngram tokens to use |
spec_ngram_mod_n_match |
int | 24 | Ngram lookup length |
ngram_map_k options (used when spec_type includes ngram_map_k)
| Option | Type | Default | Description |
|---|---|---|---|
spec_ngram_map_k_size_n |
int | 12 | N-gram lookup size |
spec_ngram_map_k_size_m |
int | 48 | M-gram proposal size |
spec_ngram_map_k_min_hits |
int | 1 | Minimum hits for accepting proposals |
ngram_map_k4v options (used when spec_type includes ngram_map_k4v)
| Option | Type | Default | Description |
|---|---|---|---|
spec_ngram_map_k4v_size_n |
int | 12 | N-gram lookup size |
spec_ngram_map_k4v_size_m |
int | 48 | M-gram proposal size |
spec_ngram_map_k4v_min_hits |
int | 1 | Minimum hits for accepting proposals |
ngram_cache lookup files
| Option | Type | Default | Description |
|---|---|---|---|
spec_lookup_cache_static / lookup_cache_static |
string | "" | Path to a static ngram lookup cache file |
spec_lookup_cache_dynamic / lookup_cache_dynamic |
string | "" | Path to a dynamic ngram lookup cache file (updated by generation) |
Speculative Type Values
| Type | Description |
|---|---|
none |
No speculative decoding (default) |
draft |
Draft model-based speculation (auto-set when draft_model is configured) |
eagle3 |
EAGLE3 draft model architecture |
ngram_simple |
Simple self-speculative using token history |
ngram_map_k |
N-gram with key-only map |
ngram_map_k4v |
N-gram with keys and 4 m-gram values |
ngram_mod |
Modified n-gram speculation |
ngram_cache |
3-level n-gram cache |
Multiple types can be chained by passing a comma-separated list to spec_type (e.g. spec_type:ngram_simple,ngram_mod). The runtime tries them in order and accepts the first proposal that meets the acceptance criteria.
{{% notice note %}}
Speculative decoding is automatically disabled when multimodal models (with mmproj) are active. The n_draft parameter can also be overridden per-request.
{{% /notice %}}
Prompt Caching
| Field | Type | Description |
|---|---|---|
prompt_cache_path |
string | Path to store prompt cache (relative to models directory) |
prompt_cache_all |
bool | Cache all prompts automatically |
prompt_cache_ro |
bool | Read-only prompt cache |
Text Processing
| Field | Type | Description |
|---|---|---|
stopwords |
array | Words or phrases that stop generation |
cutstrings |
array | Strings to cut from responses |
trimspace |
array | Strings to trim whitespace from |
trimsuffix |
array | Suffixes to trim from responses |
extract_regex |
array | Regular expressions to extract content |
System Prompt
| Field | Type | Description |
|---|---|---|
system_prompt |
string | Default system prompt for the model |
vLLM-Specific Configuration
These options apply when using the vllm backend:
| Field | Type | Description |
|---|---|---|
gpu_memory_utilization |
float32 | GPU memory utilization (0.0-1.0, default 0.9) |
trust_remote_code |
bool | Trust and execute remote code |
enforce_eager |
bool | Force eager execution mode |
swap_space |
int | Swap space in GB |
max_model_len |
int | Maximum model length |
tensor_parallel_size |
int | Tensor parallelism size |
disable_log_stats |
bool | Disable logging statistics |
dtype |
string | Data type (e.g., float16, bfloat16) |
flash_attention |
string | Flash attention configuration |
cache_type_k |
string | Key cache quantization type. Maps to llama.cpp's -ctk. Accepted values for llama.cpp-family backends (llama-cpp, ik-llama-cpp, turboquant): f16, f32, q8_0, q4_0, q4_1, q5_0, q5_1. The turboquant backend additionally accepts turbo2, turbo3, turbo4 — the fork's TurboQuant KV-cache schemes. turbo3/turbo4 auto-enable flash_attention. |
cache_type_v |
string | Value cache quantization type. Maps to llama.cpp's -ctv. Same accepted values as cache_type_k. Note: any quantized V cache requires flash_attention to be enabled. |
limit_mm_per_prompt |
object | Limit multimodal content per prompt: {image: int, video: int, audio: int} |
Template Configuration
Templates use Go templates with Sprig functions.
| Field | Type | Description |
|---|---|---|
template.chat |
string | Template for chat completion endpoint |
template.chat_message |
string | Template for individual chat messages |
template.completion |
string | Template for text completion |
template.edit |
string | Template for edit operations |
template.function |
string | Template for function/tool calls |
template.multimodal |
string | Template for multimodal interactions |
template.reply_prefix |
string | Prefix to add to model replies |
template.use_tokenizer_template |
bool | Use tokenizer's built-in template (vLLM/transformers) |
template.join_chat_messages_by_character |
string | Character to join chat messages (default: \n) |
Template Variables
Templating supports sprig functions.
Following are common variables available in templates:
{{.Input}}- User input{{.Instruction}}- Instruction for edit operations{{.System}}- System message{{.Prompt}}- Full prompt{{.Functions}}- Function definitions (for function calling){{.FunctionCall}}- Function call result
Example Template
template:
chat: |
{{.System}}
{{range .Messages}}
{{if eq .Role "user"}}User: {{.Content}}{{end}}
{{if eq .Role "assistant"}}Assistant: {{.Content}}{{end}}
{{end}}
Assistant:
Function Calling Configuration
Configure how the model handles function/tool calls:
| Field | Type | Default | Description |
|---|---|---|---|
function.disable_no_action |
bool | false |
Disable the no-action behavior |
function.no_action_function_name |
string | answer |
Name of the no-action function |
function.no_action_description_name |
string | Description for no-action function | |
function.function_name_key |
string | name |
JSON key for function name |
function.function_arguments_key |
string | arguments |
JSON key for function arguments |
function.response_regex |
array | Named regex patterns to extract function calls | |
function.argument_regex |
array | Named regex to extract function arguments | |
function.argument_regex_key_name |
string | key |
Named regex capture for argument key |
function.argument_regex_value_name |
string | value |
Named regex capture for argument value |
function.json_regex_match |
array | Regex patterns to match JSON in tool mode | |
function.replace_function_results |
array | Replace function call results with patterns | |
function.replace_llm_results |
array | Replace LLM results with patterns | |
function.capture_llm_results |
array | Capture LLM results as text (e.g., for "thinking" blocks) |
Grammar Configuration
| Field | Type | Default | Description |
|---|---|---|---|
function.grammar.disable |
bool | false |
Completely disable grammar enforcement |
function.grammar.parallel_calls |
bool | false |
Allow parallel function calls |
function.grammar.mixed_mode |
bool | false |
Allow mixed-mode grammar enforcing |
function.grammar.no_mixed_free_string |
bool | false |
Disallow free strings in mixed mode |
function.grammar.disable_parallel_new_lines |
bool | false |
Disable parallel processing for new lines |
function.grammar.prefix |
string | Prefix to add before grammar rules | |
function.grammar.expect_strings_after_json |
bool | false |
Expect strings after JSON data |
Diffusers Configuration
For image generation models using the diffusers backend:
| Field | Type | Description |
|---|---|---|
diffusers.cuda |
bool | Enable CUDA for diffusers |
diffusers.pipeline_type |
string | Pipeline type (e.g., stable-diffusion, stable-diffusion-xl) |
diffusers.scheduler_type |
string | Scheduler type (e.g., euler, ddpm) |
diffusers.enable_parameters |
string | Comma-separated parameters to enable |
diffusers.cfg_scale |
float32 | Classifier-free guidance scale |
diffusers.img2img |
bool | Enable image-to-image transformation |
diffusers.clip_skip |
int | Number of CLIP layers to skip |
diffusers.clip_model |
string | CLIP model to use |
diffusers.clip_subfolder |
string | CLIP model subfolder |
diffusers.control_net |
string | ControlNet model to use |
step |
int | Number of diffusion steps |
TTS Configuration
For text-to-speech models:
| Field | Type | Description |
|---|---|---|
tts.voice |
string | Voice file path or voice ID |
tts.audio_path |
string | Path to audio files (for Vall-E) |
Roles Configuration
Map conversation roles to specific strings:
roles:
user: "### Instruction:"
assistant: "### Response:"
system: "### System Instruction:"
Feature Flags
Enable or disable experimental features:
feature_flags:
feature_name: true
another_feature: false
MCP Configuration
Model Context Protocol (MCP) configuration:
| Field | Type | Description |
|---|---|---|
mcp.remote |
string | YAML string defining remote MCP servers |
mcp.stdio |
string | YAML string defining STDIO MCP servers |
Agent Configuration
Agent/autonomous agent configuration:
| Field | Type | Description |
|---|---|---|
agent.max_attempts |
int | Maximum number of attempts |
agent.max_iterations |
int | Maximum number of iterations |
agent.enable_reasoning |
bool | Enable reasoning capabilities |
agent.enable_planning |
bool | Enable planning capabilities |
agent.enable_mcp_prompts |
bool | Enable MCP prompts |
agent.enable_plan_re_evaluator |
bool | Enable plan re-evaluation |
Reasoning Configuration
Configure how reasoning tags are extracted and processed from model output. Reasoning tags are used by models like DeepSeek, Command-R, and others to include internal reasoning steps in their responses.
| Field | Type | Default | Description |
|---|---|---|---|
reasoning.disable |
bool | false |
When true, disables reasoning extraction entirely. The original content is returned without any processing. |
reasoning.disable_reasoning_tag_prefill |
bool | false |
When true, disables automatic prepending of thinking start tokens. Use this when your model already includes reasoning tags in its output format. |
reasoning.strip_reasoning_only |
bool | false |
When true, extracts and removes reasoning tags from content but discards the reasoning text. Useful when you want to clean reasoning tags from output without storing the reasoning content. |
reasoning.thinking_start_tokens |
array | [] |
List of custom thinking start tokens to detect in prompts. Custom tokens are checked before default tokens. |
reasoning.tag_pairs |
array | [] |
List of custom tag pairs for reasoning extraction. Each entry has start and end fields. Custom pairs are checked before default pairs. |
Reasoning Tag Formats
The reasoning extraction supports multiple tag formats used by different models:
<thinking>...</thinking>- General thinking tag<think>...</think>- DeepSeek, Granite, ExaOne, GLM models<|START_THINKING|>...<|END_THINKING|>- Command-R models<|inner_prefix|>...<|inner_suffix|>- Apertus models<seed:think>...</seed:think>- Seed models<|think|>...<|end|><|begin|>assistant<|content|>- Solar Open models[THINK]...[/THINK]- Magistral models
Examples
Disable reasoning extraction:
reasoning:
disable: true
Extract reasoning but don't prepend tags:
reasoning:
disable_reasoning_tag_prefill: true
Strip reasoning tags without storing reasoning content:
reasoning:
strip_reasoning_only: true
Complete example with reasoning configuration:
name: deepseek-model
backend: llama-cpp
parameters:
model: deepseek.gguf
reasoning:
disable: false
disable_reasoning_tag_prefill: false
strip_reasoning_only: false
Example with custom tokens and tag pairs:
name: custom-reasoning-model
backend: llama-cpp
parameters:
model: custom.gguf
reasoning:
thinking_start_tokens:
- "<custom:think>"
- "<my:reasoning>"
tag_pairs:
- start: "<custom:think>"
end: "</custom:think>"
- start: "<my:reasoning>"
end: "</my:reasoning>"
Note: Custom tokens and tag pairs are checked before the default ones, giving them priority. This allows you to override default behavior or add support for new reasoning tag formats.
Per-Request Override via Metadata
The reasoning.disable setting from model configuration can be overridden on a per-request basis using the metadata field in the OpenAI chat completion request. This allows you to enable or disable thinking for individual requests without changing the model configuration.
The metadata field accepts a map[string]string that is forwarded to the backend. The enable_thinking key controls thinking behavior:
# Enable thinking for a single request (overrides model config)
curl http://localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "qwen3",
"messages": [{"role": "user", "content": "Explain quantum computing"}],
"metadata": {"enable_thinking": "true"}
}'
# Disable thinking for a single request (overrides model config)
curl http://localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "qwen3",
"messages": [{"role": "user", "content": "Hello"}],
"metadata": {"enable_thinking": "false"}
}'
Priority order:
- Request-level
metadata.enable_thinking(highest priority) - Model config
reasoning.disable(fallback) - Auto-detected from model template (default)
Pipeline Configuration
Define pipelines for audio-to-audio processing and the [Realtime API]({{%relref "features/openai-realtime" %}}):
| Field | Type | Description |
|---|---|---|
pipeline.tts |
string | TTS model name |
pipeline.llm |
string | LLM model name |
pipeline.transcription |
string | Transcription model name |
pipeline.vad |
string | Voice activity detection model name |
gRPC Configuration
Backend gRPC communication settings:
| Field | Type | Description |
|---|---|---|
grpc.attempts |
int | Number of retry attempts |
grpc.attempts_sleep_time |
int | Sleep time between retries (seconds) |
Overrides
Override model configuration values at runtime (llama.cpp):
overrides:
- "qwen3moe.expert_used_count=int:10"
- "some_key=string:value"
Format: KEY=TYPE:VALUE where TYPE is int, float, string, or bool.
Known Use Cases
Specify which endpoints this model supports:
known_usecases:
- chat
- completion
- embeddings
Available flags: chat, completion, edit, embeddings, rerank, image, transcript, tts, sound_generation, tokenize, vad, video, detection, llm (combination of CHAT, COMPLETION, EDIT).
Complete Example
Here's a comprehensive example combining many options:
name: my-llm-model
description: A high-performance LLM model
backend: llama-stable
parameters:
model: my-model.gguf
temperature: 0.7
top_p: 0.9
top_k: 40
max_tokens: 2048
context_size: 4096
threads: 8
f16: true
gpu_layers: 35
system_prompt: "You are a helpful AI assistant."
template:
chat: |
{{.System}}
{{range .Messages}}
{{if eq .Role "user"}}User: {{.Content}}
{{else if eq .Role "assistant"}}Assistant: {{.Content}}
{{end}}
{{end}}
Assistant:
roles:
user: "User:"
assistant: "Assistant:"
system: "System:"
stopwords:
- "\n\nUser:"
- "\n\nHuman:"
prompt_cache_path: "cache/my-model"
prompt_cache_all: true
function:
grammar:
parallel_calls: true
mixed_mode: false
feature_flags:
experimental_feature: true
Related Documentation
- See [Advanced Usage]({{%relref "advanced/advanced-usage" %}}) for other configuration options
- See [Prompt Templates]({{%relref "advanced/advanced-usage#prompt-templates" %}}) for template examples
- See [CLI Reference]({{%relref "reference/cli-reference" %}}) for command-line options
GPU Auto-Fit Mode
Note: By default, LocalAI sets gpu_layers to a very large value (9999999), which effectively disables llama-cpp's auto-fit functionality. This is intentional to work with LocalAI's VRAM-based model unloading mechanism.
To enable llama-cpp's auto-fit mode, set gpu_layers: -1 in your model configuration. However, be aware of the following:
-
Trade-off: Enabling auto-fit conflicts with LocalAI's built-in VRAM threshold-based unloading. Auto-fit attempts to fit all tensors into GPU memory automatically, while LocalAI's unloading mechanism removes models when VRAM usage exceeds thresholds.
-
Known Issues: Setting
gpu_layers: -1may triggertensor_buft_overridebuffer errors in some configurations, particularly when the model exceeds available GPU memory. -
Recommendation:
- Use the default settings for most use cases (LocalAI manages VRAM automatically)
- Only enable
gpu_layers: -1if you understand the implications and have tested on your specific hardware - Monitor VRAM usage carefully when using auto-fit mode
This is a known limitation being tracked in issue #8562. A future implementation may provide a runtime toggle or custom logic to reconcile auto-fit with threshold-based unloading.