From 2955061b425d1e9ff1c9a4b985d07aef86f5d577 Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Thu, 30 Oct 2025 09:39:31 +0100 Subject: [PATCH] chore(model gallery): :robot: add 1 new models via gallery agent (#6910) chore(model gallery): :robot: add new models via gallery agent Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <2420543+mudler@users.noreply.github.com> --- gallery/index.yaml | 37 +++++++++++++++++++++++++++++++++++++ 1 file changed, 37 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 0a49e7328..6150c809d 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -23010,3 +23010,40 @@ - filename: Qwen3-VLTO-32B-Instruct.i1-Q4_K_S.gguf sha256: 789d55249614cd1acee1a23278133cd56ca898472259fa2261f77d65ed7f8367 uri: huggingface://mradermacher/Qwen3-VLTO-32B-Instruct-i1-GGUF/Qwen3-VLTO-32B-Instruct.i1-Q4_K_S.gguf +- !!merge <<: *qwen3 + name: "qwen3-vlto-32b-thinking" + urls: + - https://huggingface.co/mradermacher/Qwen3-VLTO-32B-Thinking-GGUF + description: | + **Model Name:** Qwen3-VLTO-32B-Thinking + **Model Type:** Large Language Model (Text-Only) + **Base Model:** Qwen/Qwen3-VL-32B-Thinking (vanilla Qwen3-VL-32B with vision components removed) + **Architecture:** Transformer-based, 32-billion parameter model optimized for reasoning and complex text generation. + + ### Description: + Qwen3-VLTO-32B-Thinking is a pure text-only variant of the Qwen3-VL-32B-Thinking model, stripped of its vision capabilities while preserving the full reasoning and language understanding power. It is derived by transferring the weights from the vision-language model into a text-only transformer architecture, maintaining the same high-quality behavior for tasks such as logical reasoning, code generation, and dialogue. + + This model is ideal for applications requiring deep linguistic reasoning and long-context understanding without image input. It supports advanced multimodal reasoning capabilities *in text form*—perfect for research, chatbots, and content generation. + + ### Key Features: + - ✅ 32B parameters, high reasoning capability + - ✅ No vision components — fully text-only + - ✅ Trained for complex thinking and step-by-step reasoning + - ✅ Compatible with Hugging Face Transformers and GGUF inference tools + - ✅ Available in multiple quantization levels (Q2_K to Q8_0) for efficient deployment + + ### Use Case: + Ideal for advanced text generation, logical inference, coding, and conversational AI where vision is not needed. + + > 🔗 **Base Model**: [Qwen/Qwen3-VL-32B-Thinking](https://huggingface.co/Qwen/Qwen3-VL-32B-Thinking) + > 📦 **Quantized Versions**: Available via [mradermacher/Qwen3-VLTO-32B-Thinking-GGUF](https://huggingface.co/mradermacher/Qwen3-VLTO-32B-Thinking-GGUF) + + --- + *Note: The original model was created by Alibaba’s Qwen team. This variant was adapted by qingy2024 and quantized by mradermacher.* + overrides: + parameters: + model: Qwen3-VLTO-32B-Thinking.Q4_K_M.gguf + files: + - filename: Qwen3-VLTO-32B-Thinking.Q4_K_M.gguf + sha256: d88b75df7c40455dfa21ded23c8b25463a8d58418bb6296304052b7e70e96954 + uri: huggingface://mradermacher/Qwen3-VLTO-32B-Thinking-GGUF/Qwen3-VLTO-32B-Thinking.Q4_K_M.gguf