From a48d9ce27c74fb9b99c8d36a0f1ecf01123cfa48 Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Wed, 29 Oct 2025 08:19:51 +0100 Subject: [PATCH] chore(model gallery): :robot: add 1 new models via gallery agent (#6879) chore(model gallery): :robot: add new models via gallery agent Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> Co-authored-by: mudler <2420543+mudler@users.noreply.github.com> --- gallery/index.yaml | 48 ++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 48 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index c8d221802..b1f1241b0 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -22859,3 +22859,51 @@ overrides: parameters: model: mradermacher/DeepKAT-32B-i1-GGUF +- !!merge <<: *granite4 + name: "ibm-granite.granite-4.0-1b" + urls: + - https://huggingface.co/DevQuasar/ibm-granite.granite-4.0-1b-GGUF + description: | + ### **Granite-4.0-1B** + *By IBM | Apache 2.0 License* + + **Overview:** + Granite-4.0-1B is a lightweight, instruction-tuned language model designed for efficient on-device and research use. Built on a decoder-only dense transformer architecture, it delivers strong performance in instruction following, code generation, tool calling, and multilingual tasks—making it ideal for applications requiring low latency and minimal resource usage. + + **Key Features:** + - **Size:** 1.6 billion parameters (1B Dense), optimized for efficiency. + - **Capabilities:** + - Text generation, summarization, question answering + - Code completion and function calling (e.g., API integration) + - Multilingual support (English, Spanish, French, German, Japanese, Chinese, Arabic, Korean, Portuguese, Italian, Dutch, Czech) + - Robust safety and alignment via instruction tuning and reinforcement learning + - **Architecture:** Uses GQA (Grouped Query Attention), SwiGLU activation, RMSNorm, shared input/output embeddings, and RoPE position embeddings. + - **Context Length:** Up to 128K tokens — suitable for long-form content and complex reasoning. + - **Training:** Finetuned from *Granite-4.0-1B-Base* using open-source datasets, synthetic data, and human-curated instruction pairs. + + **Performance Highlights (1B Dense):** + - **MMLU (5-shot):** 59.39 + - **HumanEval (pass@1):** 74 + - **IFEval (Alignment):** 80.82 + - **GSM8K (8-shot):** 76.35 + - **SALAD-Bench (Safety):** 93.44 + + **Use Cases:** + - On-device AI applications + - Research and prototyping + - Fine-tuning for domain-specific tasks + - Low-resource environments with high performance expectations + + **Resources:** + - [Hugging Face Model](https://huggingface.co/ibm-granite/granite-4.0-1b) + - [Granite Docs](https://www.ibm.com/granite/docs/) + - [GitHub Repository](https://github.com/ibm-granite/granite-4.0-nano-language-models) + + > *“Make knowledge free for everyone.” – IBM Granite Team* + overrides: + parameters: + model: ibm-granite.granite-4.0-1b.Q4_K_M.gguf + files: + - filename: ibm-granite.granite-4.0-1b.Q4_K_M.gguf + sha256: 0e0ef42486b7f1f95dfe33af2e696df1149253e500c48f3fb8db0125afa2922c + uri: huggingface://DevQuasar/ibm-granite.granite-4.0-1b-GGUF/ibm-granite.granite-4.0-1b.Q4_K_M.gguf