diff --git a/gallery/index.yaml b/gallery/index.yaml index 75fa605c1..70bd72418 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -36,8 +36,6 @@ url: "github:mudler/LocalAI/gallery/virtual.yaml@master" urls: - https://huggingface.co/mradermacher/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-heretic-i1-GGUF - description: | - Describe the model in a clear and concise way that can be shared in a model gallery. tags: - default - default @@ -63,8 +61,6 @@ url: "github:mudler/LocalAI/gallery/virtual.yaml@master" urls: - https://huggingface.co/bartowski/Qwen_Qwen3.5-0.8B-GGUF - description: | - Describe the model in a clear and concise way that can be shared in a model gallery. license: "unknown" tags: - llm @@ -99,11 +95,7 @@ url: "github:mudler/LocalAI/gallery/virtual.yaml@master" urls: - https://huggingface.co/bartowski/Qwen_Qwen3.5-2B-GGUF - description: | - Describe the model in a clear and concise way that can be shared in a model gallery. - license: "proprietary" tags: - - llama - qwen - qwen3.5 - quantized @@ -135,12 +127,8 @@ url: "github:mudler/LocalAI/gallery/virtual.yaml@master" urls: - https://huggingface.co/bartowski/Qwen_Qwen3.5-4B-GGUF - description: | - Describe the model in a clear and concise way that can be shared in a model gallery. tags: - qwen - - llama2 - - gemma - qwen3.5 - 4b overrides: @@ -171,7 +159,6 @@ - https://huggingface.co/mradermacher/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-i1-GGUF description: | Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-i1-GGUF - A GGUF quantized model optimized for local inference. Specialized for reasoning and chain-of-thought tasks. Based on Qwen 3.5 architecture with enhanced language understanding. Available in multiple quantization levels for various hardware requirements. Distilled from Claude-style reasoning models for enhanced logical reasoning capabilities. - license: "license" tags: - llm - qwen @@ -201,7 +188,6 @@ - https://huggingface.co/Jackrong/Qwen3.5-4B-Claude-4.6-Opus-Reasoning-Distilled-GGUF description: | Qwen3.5-4B-Claude-4.6-Opus-Reasoning-Distilled-GGUF - A GGUF quantized model optimized for local inference. Specialized for reasoning and chain-of-thought tasks. Based on Qwen 3.5 architecture with enhanced language understanding. Available in multiple quantization levels for various hardware requirements. Distilled from Claude-style reasoning models for enhanced logical reasoning capabilities. - license: "gpl-3.0" tags: - llm - guf