Ettore Di Giacinto
856d90400a
chore(uv): add --index-strategy=unsafe-first-match to l4t
...
This is because the main index might not contain all the dependencies
for torch
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-08 21:47:18 +00:00
Copilot
06323df457
Optimize GPU library copying to preserve symlinks and avoid duplicates ( #7931 )
...
* Initial plan
* Optimize library copying to preserve symlinks and avoid duplicates
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
* Address code review feedback: extract get_inode helper, use file type detection for sorting
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
* Simplify implementation by removing inode tracking
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
* Add clarifying comment about basename deduplication
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
---------
Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com >
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-08 22:26:48 +01:00
Richard Palethorpe
98f28bf583
chore(docs): Add Crush and VoxInput to the integrations ( #7924 )
...
* chore(docs): Add Crush and VoxInput to the integrations
Signed-off-by: Richard Palethorpe <io@richiejp.com >
* Apply suggestion from @mudler
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Richard Palethorpe <io@richiejp.com >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2026-01-08 21:39:25 +01:00
Ettore Di Giacinto
383312b50e
chore(l4t-12): do not use python 3.12 (wheels are only for 3.10) ( #7928 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-08 19:00:07 +01:00
Ettore Di Giacinto
b736db4bbe
chore(ci): use latest jetpack image for l4t ( #7926 )
...
This image is for HW prior Jetpack 7. Jetpack 7 broke compatibility with
older devices (which are still in use) such as AGX Orin or Jetsons.
While we do have l4t-cuda-13 images with sbsa support for new Nvidia
devices (Thor, DGX, etc). For older HW we are forced to keep old images
around as 24.04 does not seem to be supported.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-08 18:30:59 +01:00
LocalAI [bot]
09bc2e4a00
chore(model gallery): 🤖 add 1 new models via gallery agent ( #7922 )
...
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-08 11:06:21 +01:00
LocalAI [bot]
c03e532a18
chore: ⬆️ Update ggml-org/llama.cpp to ae9f8df77882716b1702df2bed8919499e64cc28 ( #7915 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-07 23:24:01 +01:00
Ettore Di Giacinto
fcb58ee243
fix(intel): Add ARG for Ubuntu codename in Dockerfile ( #7917 )
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2026-01-07 21:55:18 +01:00
Copilot
b2ff1cea2a
feat: enable Vulkan arm64 image builds ( #7912 )
...
* Initial plan
* Add arm64 support for Vulkan builds in Dockerfiles and workflows
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
---------
Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com >
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-07 21:49:50 +01:00
Ettore Di Giacinto
b964b3d53e
feat(backends): add moonshine backend for faster transcription ( #7833 )
...
* feat(backends): add moonshine backend for faster transcription
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add backend to CI, update AGENTS.md from this exercise
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-07 21:44:35 +01:00
LocalAI [bot]
0b26669d0b
chore(model gallery): 🤖 add 1 new models via gallery agent ( #7916 )
...
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-07 21:43:39 +01:00
Ettore Di Giacinto
5a9698bc69
chore(Dockerfile): restore GPU vendor specific sections ( #7911 )
...
Until we figure out https://github.com/mudler/LocalAI/issues/7909
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-07 16:34:23 +01:00
Ettore Di Giacinto
1fe0e9f74f
chore(ci): restore building of GPU vendor images ( #7910 )
...
Until we figure out https://github.com/mudler/LocalAI/issues/7909
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-07 16:32:22 +01:00
Ettore Di Giacinto
ffb2dc4666
chore(detection): detect GPU vendor from files present in the system ( #7908 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-07 16:18:27 +01:00
Ettore Di Giacinto
cfc2225fc7
chore(dockerfile): drop driver-requirements section ( #7907 )
...
* chore(dockerfile): drop driver-requirements section
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* chore(ci): drop other builds
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-07 16:18:14 +01:00
Copilot
fd53978a7b
feat: package GPU libraries inside backend containers for unified base image ( #7891 )
...
* Initial plan
* Add GPU library packaging for isolated backend environments
- Create scripts/build/package-gpu-libs.sh for packaging CUDA, ROCm, SYCL, and Vulkan libraries
- Update llama-cpp, whisper, stablediffusion-ggml package.sh to include GPU libraries
- Update Dockerfile.python to package GPU libraries into Python backends
- Update libbackend.sh to set LD_LIBRARY_PATH for GPU library loading
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
* Address code review feedback: fix variable consistency and quoting
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
* Fix code review issues: improve glob handling and remove redundant variable
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
* Simplify main Dockerfile and workflow to use unified base image
- Remove GPU-specific driver installation from Dockerfile (CUDA, ROCm, Vulkan, Intel)
- Simplify image.yml workflow to build single unified base image for linux/amd64 and linux/arm64
- GPU libraries are now packaged in individual backend containers
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
---------
Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com >
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2026-01-07 15:48:51 +01:00
LocalAI [bot]
7abc0242bb
chore(model gallery): 🤖 add 1 new models via gallery agent ( #7903 )
...
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-07 09:46:36 +01:00
LocalAI [bot]
23df29fbd3
chore: ⬆️ Update leejet/stable-diffusion.cpp to 9be0b91927dfa4007d053df72dea7302990226bb ( #7895 )
...
⬆️ Update leejet/stable-diffusion.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-06 22:18:53 +01:00
LocalAI [bot]
fb9879949c
chore: ⬆️ Update ggml-org/llama.cpp to ccbc84a5374bab7a01f68b129411772ddd8e7c79 ( #7894 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-06 22:18:35 +01:00
Manish Dewangan
1642b39cb8
[gallery] add JSON schema for gallery model specification ( #7890 )
...
Add JSON Schema for gallery model specification
Signed-off-by: devmanishofficial <devmanishofficial@gmail.com >
2026-01-06 22:10:43 +01:00
Richard Palethorpe
e6ba26c3e7
chore: Update to Ubuntu24.04 (cont #7423 ) ( #7769 )
...
* ci(workflows): bump GitHub Actions images to Ubuntu 24.04
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* ci(workflows): remove CUDA 11.x support from GitHub Actions (incompatible with ubuntu:24.04)
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* ci(workflows): bump GitHub Actions CUDA support to 12.9
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* build(docker): bump base image to ubuntu:24.04 and adjust Vulkan SDK/packages
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* fix(backend): correct context paths for Python backends in workflows, Makefile and Dockerfile
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore(make): disable parallel backend builds to avoid race conditions
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore(make): export CUDA_MAJOR_VERSION and CUDA_MINOR_VERSION for override
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* build(backend): update backend Dockerfiles to Ubuntu 24.04
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore(backend): add ROCm env vars and default AMDGPU_TARGETS for hipBLAS builds
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore(chatterbox): bump ROCm PyTorch to 2.9.1+rocm6.4 and update index URL; align hipblas requirements
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore: add local-ai-launcher to .gitignore
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* ci(workflows): fix backends GitHub Actions workflows after rebase
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* build(docker): use build-time UBUNTU_VERSION variable
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore(docker): remove libquadmath0 from requirements-stage base image
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore(make): add backends/vllm to .NOTPARALLEL to prevent parallel builds
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* fix(docker): correct CUDA installation steps in backend Dockerfiles
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* chore(backend): update ROCm to 6.4 and align Python hipblas requirements
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* ci(workflows): switch GitHub Actions runners to Ubuntu-24.04 for CUDA on arm64 builds
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* build(docker): update base image and backend Dockerfiles for Ubuntu 24.04 compatibility on arm64
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* build(backend): increase timeout for uv installs behind slow networks on backend/Dockerfile.python
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* ci(workflows): switch GitHub Actions runners to Ubuntu-24.04 for vibevoice backend
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* ci(workflows): fix failing GitHub Actions runners
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
* fix: Allow FROM_SOURCE to be unset, use upstream Intel images etc.
Signed-off-by: Richard Palethorpe <io@richiejp.com >
* chore(build): rm all traces of CUDA 11
Signed-off-by: Richard Palethorpe <io@richiejp.com >
* chore(build): Add Ubuntu codename as an argument
Signed-off-by: Richard Palethorpe <io@richiejp.com >
---------
Signed-off-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
Signed-off-by: Richard Palethorpe <io@richiejp.com >
Co-authored-by: Alessandro Sturniolo <alessandro.sturniolo@gmail.com >
2026-01-06 15:26:42 +01:00
Ettore Di Giacinto
26c4f80d1b
chore(llama.cpp/flags): simplify conditionals ( #7887 )
...
If ggml handle conditionals correctly we don't need to handle it here.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-06 15:02:20 +01:00
coffeerunhobby
5add7b47f5
fix: BMI2 crash on AVX-only CPUs (Intel Ivy Bridge/Sandy Bridge) ( #7864 )
...
* Fix BMI2 crash on AVX-only CPUs (Intel Ivy Bridge/Sandy Bridge)
Signed-off-by: coffeerunhobby <coffeerunhobby@users.noreply.github.com >
* Address feedback from review
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: coffeerunhobby <coffeerunhobby@users.noreply.github.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: coffeerunhobby <coffeerunhobby@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2026-01-06 00:13:48 +00:00
Ettore Di Giacinto
3244ccc224
chore(image-ui): simplify interface ( #7882 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-05 23:20:28 +01:00
LocalAI [bot]
4f7b6b0bff
chore: ⬆️ Update ggml-org/llama.cpp to e443fbcfa51a8a27b15f949397ab94b5e87b2450 ( #7881 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-05 22:55:40 +01:00
LocalAI [bot]
3a629cea2f
chore: ⬆️ Update ggml-org/whisper.cpp to 679bdb53dbcbfb3e42685f50c7ff367949fd4d48 ( #7879 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-05 22:55:16 +01:00
LocalAI [bot]
f917feda29
chore: ⬆️ Update leejet/stable-diffusion.cpp to c5602a676caff5fe5a9f3b76b2bc614faf5121a5 ( #7880 )
...
⬆️ Update leejet/stable-diffusion.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-05 22:54:56 +01:00
dependabot[bot]
e2018cdc8f
chore(deps): bump github.com/labstack/echo/v4 from 4.14.0 to 4.15.0 ( #7875 )
...
Bumps [github.com/labstack/echo/v4](https://github.com/labstack/echo ) from 4.14.0 to 4.15.0.
- [Release notes](https://github.com/labstack/echo/releases )
- [Changelog](https://github.com/labstack/echo/blob/master/CHANGELOG.md )
- [Commits](https://github.com/labstack/echo/compare/v4.14.0...v4.15.0 )
---
updated-dependencies:
- dependency-name: github.com/labstack/echo/v4
dependency-version: 4.15.0
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2026-01-05 22:54:30 +01:00
Manish Dewangan
a3b8a94187
fix(ui): fix 404 on API menu link by pointing to index.html ( #7878 )
...
Signed-off-by: devmanishofficial <devmanishofficial@gmail.com >
2026-01-05 22:54:14 +01:00
dependabot[bot]
41de7d32ad
chore(deps): bump dependabot/fetch-metadata from 2.4.0 to 2.5.0 ( #7876 )
...
Bumps [dependabot/fetch-metadata](https://github.com/dependabot/fetch-metadata ) from 2.4.0 to 2.5.0.
- [Release notes](https://github.com/dependabot/fetch-metadata/releases )
- [Commits](https://github.com/dependabot/fetch-metadata/compare/v2.4.0...v2.5.0 )
---
updated-dependencies:
- dependency-name: dependabot/fetch-metadata
dependency-version: 2.5.0
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2026-01-05 20:10:07 +00:00
Richard Palethorpe
93364df0a8
chore(AGENTS.md): Add section to help with building backends ( #7871 )
...
Signed-off-by: Richard Palethorpe <io@richiejp.com >
2026-01-05 18:25:52 +01:00
Ettore Di Giacinto
21c84f432f
feat(function): Add tool streaming, XML Tool Call Parsing Support ( #7865 )
...
* feat(function): Add XML Tool Call Parsing Support
Extend the function parsing system in LocalAI to support XML-style tool calls, similar to how JSON tool calls are currently parsed. This will allow models that return XML format (like <tool_call><function=name><parameter=key>value</parameter></function></tool_call>) to be properly parsed alongside text content.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* thinking before tool calls, more strict support for corner cases with no tools
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Support streaming tools
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Iterative JSON
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Iterative parsing
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Consume JSON marker
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fixup
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* add tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fix pending TODOs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Don't run other parsing with ParseRegex
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-05 18:25:40 +01:00
LocalAI [bot]
9d3da0bed5
chore: ⬆️ Update ggml-org/llama.cpp to 4974bf53cf14073c7b66e1151348156aabd42cb8 ( #7861 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-05 00:10:18 +01:00
LocalAI [bot]
1b063b5595
chore: ⬆️ Update leejet/stable-diffusion.cpp to b90b1ee9cf84ea48b478c674dd2ec6a33fd504d6 ( #7862 )
...
⬆️ Update leejet/stable-diffusion.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-04 23:52:01 +01:00
Ettore Di Giacinto
560bf50299
chore(Makefile): refactor common make targets ( #7858 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-04 21:12:50 +01:00
LocalAI [bot]
a7e155240b
chore: ⬆️ Update ggml-org/llama.cpp to e57f52334b2e8436a94f7e332462dfc63a08f995 ( #7848 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-04 10:27:45 +01:00
LocalAI [bot]
793e4907a2
feat(swagger): update swagger ( #7847 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-03 22:09:39 +01:00
Ettore Di Giacinto
d38811560c
chore(docs): add opencode, GHA, and realtime voice assistant examples
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-03 22:03:43 +01:00
Ettore Di Giacinto
33cc0b8e13
fix(chat/ui): record model name in history for consistency ( #7845 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-03 18:05:33 +01:00
lif
4cd95b8a9d
fix: Highly inconsistent agent response to cogito agent calling MCP server - Body "Invalid http method" ( #7790 )
...
* fix: resolve duplicate MCP route registration causing 50% failure rate
Fixes #7772
The issue was caused by duplicate registration of the MCP endpoint
/mcp/v1/chat/completions in both openai.go and localai.go, leading
to a race condition where requests would randomly hit different
handlers with incompatible behaviors.
Changes:
- Removed duplicate MCP route registration from openai.go
- Kept the localai.MCPStreamEndpoint as the canonical handler
- Added all three MCP route patterns for backward compatibility:
* /v1/mcp/chat/completions
* /mcp/v1/chat/completions
* /mcp/chat/completions
- Added comments to clarify route ownership and prevent future conflicts
- Fixed formatting in ui_api.go
The localai.MCPStreamEndpoint handler is more feature-complete as it
supports both streaming and non-streaming modes, while the removed
openai.MCPCompletionEndpoint only supported synchronous requests.
This eliminates the ~50% failure rate where the cogito library would
receive "Invalid http method" errors when internal HTTP requests were
routed to the wrong handler.
🤖 Generated with [Claude Code](https://claude.com/claude-code )
Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com >
Signed-off-by: majiayu000 <1835304752@qq.com >
* Address feedback from review
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: majiayu000 <1835304752@qq.com >
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: Claude Sonnet 4.5 <noreply@anthropic.com >
Co-authored-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-03 15:43:23 +01:00
LocalAI [bot]
8c504113a2
chore(model gallery): 🤖 add 1 new models via gallery agent ( #7840 )
...
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-03 08:42:05 +01:00
coffeerunhobby
666d110714
fix: Prevent BMI2 instruction crash on AVX-only CPUs ( #7817 )
...
* Fix: Prevent BMI2 instruction crash on AVX-only CPUs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* fix: apply no-bmi flags on non-darwin
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Co-authored-by: coffeerunhobby <coffeerunhobby@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-03 08:36:55 +01:00
LocalAI [bot]
641606ae93
chore: ⬆️ Update ggml-org/llama.cpp to 706e3f93a60109a40f1224eaf4af0d59caa7c3ae ( #7836 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-02 21:26:37 +00:00
Ettore Di Giacinto
5f6c941399
fix(llama.cpp/mmproj): fix loading mmproj in nested sub-dirs different from model path ( #7832 )
...
fix(mmproj): fix loading mmproj in nested sub-dirs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-02 20:17:30 +01:00
LocalAI [bot]
1639fc6309
chore(model gallery): 🤖 add 1 new models via gallery agent ( #7831 )
...
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-02 15:10:00 +01:00
Ettore Di Giacinto
841e8f6d47
fix(image-gen): fix scrolling issues ( #7829 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-02 09:05:49 +01:00
LocalAI [bot]
fd152c97c0
chore(model gallery): 🤖 add 1 new models via gallery agent ( #7826 )
...
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-02 08:45:43 +01:00
LocalAI [bot]
949de04052
chore: ⬆️ Update ggml-org/llama.cpp to ced765be44ce173c374f295b3c6f4175f8fd109b ( #7822 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2026-01-02 08:44:49 +01:00
Ettore Di Giacinto
76cfe1f367
feat(image-gen/UI): move controls to the left, make the page more compact ( #7823 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2026-01-01 22:07:42 +01:00
LocalAI [bot]
5ee6c1810b
feat(swagger): update swagger ( #7820 )
...
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2026-01-01 21:16:38 +01:00