From 55c05211d31ec3e25cf649ebd2bb4a7aa2473bdc Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Fri, 17 Apr 2026 16:10:02 +0200 Subject: [PATCH] chore(model gallery): :robot: add 1 new models via gallery agent (#9399) chore(model gallery): :robot: add new models via gallery agent Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <2420543+mudler@users.noreply.github.com> --- gallery/index.yaml | 62 ++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 62 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 3f0a53c4a..a29cc382b 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -1,4 +1,66 @@ --- +- name: "qwen3.6-35b-a3b" + url: "github:mudler/LocalAI/gallery/virtual.yaml@master" + urls: + - https://huggingface.co/unsloth/Qwen3.6-35B-A3B-GGUF + description: | + # Qwen3.6-35B-A3B + + [](https://chat.qwen.ai) + + > [!Note] + > This repository contains model weights and configuration files for the post-trained model in the Hugging Face Transformers format. + > + > These artifacts are compatible with Hugging Face Transformers, vLLM, SGLang, KTransformers, etc. + + Following the February release of the Qwen3.5 series, we're pleased to share the first open-weight variant of Qwen3.6. Built on direct feedback from the community, Qwen3.6 prioritizes stability and real-world utility, offering developers a more intuitive, responsive, and genuinely productive coding experience. + + ## Qwen3.6 Highlights + + This release delivers substantial upgrades, particularly in + + - **Agentic Coding:** the model now handles frontend workflows and repository-level reasoning with greater fluency and precision. + - **Thinking Preservation:** we've introduced a new option to retain reasoning context from historical messages, streamlining iterative development and reducing overhead. + + For more details, please refer to our blog post Qwen3.6-35B-A3B. + + ## Model Overview + + ... + license: "apache-2.0" + tags: + - llm + - gguf + - qwen + icon: https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen3.6/Figures/qwen3.6_35b_a3b_score.png + overrides: + backend: llama-cpp + function: + automatic_tool_parsing_fallback: true + grammar: + disable: true + known_usecases: + - chat + mmproj: llama-cpp/mmproj/Qwen3.6-35B-A3B-GGUF/mmproj-F32.gguf + options: + - use_jinja:true + parameters: + min_p: 0 + model: llama-cpp/models/Qwen3.6-35B-A3B-GGUF/Qwen3.6-35B-A3B-UD-Q4_K_M.gguf + presence_penalty: 1.5 + repeat_penalty: 1 + temperature: 0.7 + top_k: 20 + top_p: 0.8 + template: + use_tokenizer_template: true + files: + - filename: llama-cpp/models/Qwen3.6-35B-A3B-GGUF/Qwen3.6-35B-A3B-UD-Q4_K_M.gguf + sha256: ac0e2c1189e055faa36eff361580e79c5bd6f8e76bffb4ce547f167d53e31a61 + uri: https://huggingface.co/unsloth/Qwen3.6-35B-A3B-GGUF/resolve/main/Qwen3.6-35B-A3B-UD-Q4_K_M.gguf + - filename: llama-cpp/mmproj/Qwen3.6-35B-A3B-GGUF/mmproj-F32.gguf + sha256: 0a1c1cd2772ae6de5e87e023cea454720924675f11fe2b0e7bb7648e48debdc0 + uri: https://huggingface.co/unsloth/Qwen3.6-35B-A3B-GGUF/resolve/main/mmproj-F32.gguf - name: "gemma-4-26b-a4b-it-apex" url: "github:mudler/LocalAI/gallery/virtual.yaml@master" urls: