From 4730b52461ff6e4ff014584b71d9f706e80c9664 Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Sun, 9 Nov 2025 08:40:40 +0100 Subject: [PATCH] chore(model gallery): :robot: add 1 new models via gallery agent (#7205) chore(model gallery): :robot: add new models via gallery agent Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <2420543+mudler@users.noreply.github.com> --- gallery/index.yaml | 39 +++++++++++++++++++++++++++++++++++++++ 1 file changed, 39 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index ef53e48f3..ee35e5c3c 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -23203,3 +23203,42 @@ - filename: Spiral-Qwen3-4B-Multi-Env.Q4_K_M.gguf sha256: e91914c18cb91f2a3ef96d8e62a18b595dd6c24fad901dea639e714bc7443b09 uri: huggingface://mradermacher/Spiral-Qwen3-4B-Multi-Env-GGUF/Spiral-Qwen3-4B-Multi-Env.Q4_K_M.gguf +- !!merge <<: *gptoss + name: "metatune-gpt20b-r1.1-i1" + urls: + - https://huggingface.co/mradermacher/metatune-gpt20b-R1.1-i1-GGUF + description: | + **Model Name:** MetaTune-GPT20B-R1.1 + **Base Model:** unsloth/gpt-oss-20b-unsloth-bnb-4bit + **Repository:** [EpistemeAI/metatune-gpt20b-R1.1](https://huggingface.co/EpistemeAI/metatune-gpt20b-R1.1) + **License:** Apache 2.0 + + **Description:** + MetaTune-GPT20B-R1.1 is a large language model fine-tuned for recursive self-improvement, making it one of the first publicly released models capable of autonomously generating training data, evaluating its own performance, and adjusting its hyperparameters to improve over time. Built upon the open-weight GPT-OSS 20B architecture and trained with Unsloth's optimized 4-bit quantization, this model excels in complex reasoning, agentic tasks, and function calling. It supports tools like web browsing and structured output generation, and is particularly effective in high-reasoning use cases such as scientific problem-solving and math reasoning. + + **Performance Highlights (Zero-shot):** + - **GPQA Diamond:** 93.3% exact match + - **GSM8K (Chain-of-Thought):** 100% exact match + + **Recommended Use:** + - Advanced reasoning & planning + - Autonomous agent workflows + - Research, education, and technical problem-solving + + **Safety Note:** + Use with caution. For safety-critical applications, pair with a safety guardrail model such as [openai/gpt-oss-safeguard-20b](https://huggingface.co/openai/gpt-oss-safeguard-20b). + + **Fine-Tuned From:** unsloth/gpt-oss-20b-unsloth-bnb-4bit + **Training Method:** Recursive Self-Improvement on the [Recursive Self-Improvement Dataset](https://huggingface.co/datasets/EpistemeAI/recursive_self_improvement_dataset) + **Framework:** Hugging Face TRL + Unsloth for fast, efficient training + + **Inference Tip:** Set reasoning level to "high" for best results and to reduce prompt injection risks. + + 👉 [View on Hugging Face](https://huggingface.co/EpistemeAI/metatune-gpt20b-R1.1) | [GitHub: Recursive Self-Improvement](https://github.com/openai/harmony) + overrides: + parameters: + model: metatune-gpt20b-R1.1.i1-Q4_K_M.gguf + files: + - filename: metatune-gpt20b-R1.1.i1-Q4_K_M.gguf + sha256: 82a77f5681c917df6375bc0b6c28bf2800d1731e659fd9bbde7b5598cf5e9d0a + uri: huggingface://mradermacher/metatune-gpt20b-R1.1-i1-GGUF/metatune-gpt20b-R1.1.i1-Q4_K_M.gguf