Ettore Di Giacinto
d0fb23514f
Revert "fix(gallery): correctly show status for downloading OCI images"
...
This reverts commit 780d034ac9 .
2025-07-01 21:32:04 +02:00
Ettore Di Giacinto
780d034ac9
fix(gallery): correctly show status for downloading OCI images
...
We can't use the mutate.Extract written bytes as current status as that
will be bigger than the compressed image size. Image manifest don't have
any guarantee of the type of artifact (can be compressed or not) when
showing the layer size.
Split the extraction process in two parts: Downloading and extracting as
a flattened system, in this way we can display the status of downloading
and extracting accordingly.
This change also fixes a small nuance in detecting installed backends,
now it's more consistent and looks if a metadata.json and/or a path with
a `run.sh` file is present.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-07-01 19:56:28 +02:00
Ettore Di Giacinto
ec2a044c7e
chore(model gallery): add pinkpixel_crystal-think-v2 ( #5773 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-07-01 16:18:19 +02:00
Ettore Di Giacinto
ad6fdd21fd
chore(model gallery): add steelskull_l3.3-shakudo-70b ( #5772 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-07-01 16:15:22 +02:00
Ettore Di Giacinto
cd94e6b352
chore(model gallery): add thedrummer_anubis-70b-v1.1 ( #5771 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-07-01 13:54:29 +02:00
Richard Palethorpe
b37cef3718
fix: Diffusers and XPU fixes ( #5737 )
...
* fix(README): Add device flags for Intel/XPU
Signed-off-by: Richard Palethorpe <io@richiejp.com >
* fix(diffusers/xpu): Set device to XPU and ignore CUDA request when on Intel
Signed-off-by: Richard Palethorpe <io@richiejp.com >
---------
Signed-off-by: Richard Palethorpe <io@richiejp.com >
2025-07-01 12:36:17 +02:00
Dedy F. Setyawan
9f957d547d
fix(docs): Improve Header Responsiveness - Hide "Star us on GitHub!" on Mobile ( #5770 )
2025-07-01 12:15:16 +02:00
LocalAI [bot]
f0d9f0c5d8
chore: ⬆️ Update ggml-org/llama.cpp to 0a5a3b5cdfd887cf0f8e09d9ff89dee130cfcdde ( #5759 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-30 22:46:55 +00:00
LocalAI [bot]
d33e1c72a3
chore: ⬆️ Update ggml-org/llama.cpp to caf5681fcb47dfe9bafee94ef9aa8f669ac986c7 ( #5758 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-29 22:49:47 +00:00
Ettore Di Giacinto
33f9ee06c9
fix(gallery): automatically install model from name ( #5757 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-29 17:42:58 +02:00
Ettore Di Giacinto
c54677402d
chore(model gallery): add qwen3-33b-a3b-stranger-thoughts-abliterated-uncensored ( #5755 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-29 10:04:33 +02:00
LocalAI [bot]
3fe3a7b23d
chore: ⬆️ Update ggml-org/llama.cpp to 27208bf657cfe7262791df473927225e48efe482 ( #5753 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-29 09:06:37 +02:00
LocalAI [bot]
f8ff6fa1fd
docs: ⬆️ update docs version mudler/LocalAI ( #5752 )
...
⬆️ Update docs version mudler/LocalAI
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-28 22:17:49 +02:00
Ettore Di Giacinto
dfadc3696e
feat(llama.cpp): allow to set kv-overrides ( #5745 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-28 21:26:07 +02:00
Ettore Di Giacinto
dbcf5fb4fc
chore(model gallery): add gemma-3-4b-it-max-horror-uncensored-dbl-x-imatrix ( #5751 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-28 18:18:14 +02:00
Ettore Di Giacinto
2633137a17
chore(model gallery): add qwen3-22b-a3b-the-harley-quinn ( #5750 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-28 18:17:10 +02:00
Ettore Di Giacinto
d9c17dd23b
chore(model gallery): add mistral-small-3.2-46b-the-brilliant-raconteur-ii-instruct-2506 ( #5749 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-28 18:15:25 +02:00
Ettore Di Giacinto
d8b7bd4860
chore(model gallery): add qwen3-42b-a3b-stranger-thoughts-deep20x-abliterated-uncensored-i1 ( #5748 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-28 18:12:17 +02:00
Ettore Di Giacinto
a611cbc0f4
chore(model gallery): add qwen3-55b-a3b-total-recall-deep-40x ( #5747 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-28 17:54:32 +02:00
Ettore Di Giacinto
850b525159
chore(model gallery): add qwen3-55b-a3b-total-recall-v1.3-i1 ( #5746 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-28 17:47:46 +02:00
Ettore Di Giacinto
35b3426a2a
Update README.md
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-06-28 09:16:25 +02:00
LocalAI [bot]
cd2b0c0e7c
chore: ⬆️ Update ggml-org/llama.cpp to 72babea5dea56c8a8e8420ccf731b12a5cf37854 ( #5743 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
v3.1.1
2025-06-27 23:46:27 +02:00
LocalAI [bot]
73d80c43a8
chore: ⬆️ Update ggml-org/whisper.cpp to c88ffbf9baeaae8c2cc0a4f496618314bb2ee9e0 ( #5742 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-27 23:45:57 +02:00
LocalAI [bot]
665562b850
docs: ⬆️ update docs version mudler/LocalAI ( #5741 )
...
⬆️ Update docs version mudler/LocalAI
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-27 22:23:43 +02:00
Ettore Di Giacinto
7a78e4f482
fix(backends gallery): meta packages do not have URIs ( #5740 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-27 22:23:14 +02:00
Ettore Di Giacinto
6f41a6f934
fix(backends gallery): correctly identify gpu vendor ( #5739 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-27 22:22:58 +02:00
Ettore Di Giacinto
bb54f2da2b
feat(gallery): automatically install missing backends along models ( #5736 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-27 18:25:44 +02:00
Ettore Di Giacinto
e1cc7ee107
fix(ci): enable tag-latest to auto ( #5738 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-27 18:17:01 +02:00
Ettore Di Giacinto
cfc9dfa3d5
fix(ci): better handling of latest images for backends ( #5735 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-27 10:32:58 +02:00
LocalAI [bot]
6a650e68cb
chore: ⬆️ Update ggml-org/whisper.cpp to 32cf4e2aba799aff069011f37ca025401433cf9f ( #5733 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
v3.1.0
2025-06-26 22:09:26 +02:00
LocalAI [bot]
5e1373877a
chore: ⬆️ Update ggml-org/llama.cpp to 8846aace4934ad29651ea61b8c7e3f6b0556e3d2 ( #5734 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-26 22:09:03 +02:00
Ettore Di Giacinto
b5b0ab26e7
fix(ci): remove non-existant input from build matrix
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-26 21:42:27 +02:00
Ettore Di Giacinto
9725bb4bbd
chore(model gallery): add gemma-3n-e4b-it ( #5731 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-26 19:36:50 +02:00
Ettore Di Giacinto
33b4275bbc
chore(model gallery): add gemma-3n-e2b-it ( #5730 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-26 19:35:49 +02:00
Ettore Di Giacinto
6644af10c6
feat: ⚠️ reduce images size and stop bundling sources ( #5721 )
...
feat: reduce images size and stop bundling sources
Do not copy sources anymore, and reduce packages of the base images by
not using builder images.
If needed to rebuild, just build the container image from scratch by
following the docs. We will slowly try to migrate all backends to the
gallery to keep the core small.
This PR is a breaking change, it also sets the base folders to /models
and /backends instead of /build/models and /build/backends.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-06-26 18:41:38 +02:00
Ettore Di Giacinto
7c4a2e9b85
chore(ci): ⚠️ fix latest tag by using docker meta action ( #5722 )
...
chore(ci): fix latest tag by using docker meta action
Also uniform tagging names
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-26 18:40:25 +02:00
Ettore Di Giacinto
bcccee3909
fix(backends gallery): delete dangling dirs if installation failed ( #5729 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-26 17:38:03 +02:00
Ettore Di Giacinto
c6f50ddd0c
Revert "chore: ⬆️ Update leejet/stable-diffusion.cpp to 10c6501bd05a697e014f1bee3a84e5664290c489" ( #5727 )
...
Revert "chore: ⬆️ Update leejet/stable-diffusion.cpp to `10c6501bd05a…"
This reverts commit 30600dd5cb .
2025-06-26 13:25:25 +02:00
LocalAI [bot]
6613373b1b
chore: ⬆️ Update ggml-org/whisper.cpp to 4daf7050ca2bf17f5166f45ac6da651c4e33f293 ( #5725 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-26 13:00:20 +02:00
LocalAI [bot]
1659b3f795
chore: ⬆️ Update ggml-org/llama.cpp to 2bf9d539dd158345e3a3b096e16474af535265b4 ( #5724 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-26 12:59:57 +02:00
LocalAI [bot]
30600dd5cb
chore: ⬆️ Update leejet/stable-diffusion.cpp to 10c6501bd05a697e014f1bee3a84e5664290c489 ( #4925 )
...
⬆️ Update leejet/stable-diffusion.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-25 22:45:15 +00:00
Ettore Di Giacinto
179fcf5541
chore(model gallery): add menlo_jan-nano-128k ( #5723 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-25 12:14:11 +02:00
LocalAI [bot]
9cb75086bb
chore: ⬆️ Update ggml-org/whisper.cpp to 0083335ba0e9d6becbe0958903b0a27fc2ebaeed ( #5718 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-25 09:47:33 +02:00
LocalAI [bot]
594bb462ab
chore: ⬆️ Update ggml-org/llama.cpp to 73e53dc834c0a2336cd104473af6897197b96277 ( #5719 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-24 22:47:48 +00:00
Ettore Di Giacinto
aa730a7b96
chore(model gallery): add delta-vector_austral-24b-winton ( #5717 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-24 18:37:28 +02:00
Ettore Di Giacinto
0a454c527a
chore(model gallery): add astrosage-70b ( #5716 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-24 18:34:37 +02:00
Ettore Di Giacinto
cf86bcb984
chore(model gallery): add skywork_skywork-swe-32b ( #5715 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-24 18:29:36 +02:00
Ettore Di Giacinto
a6d9988e84
feat(backend gallery): add meta packages ( #5696 )
...
* feat(backend gallery): add meta packages
So we can have meta packages such as "vllm" that automatically installs
the corresponding package depending on the GPU that is being currently
detected in the system.
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat: use a metadata file
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-24 17:08:27 +02:00
Ettore Di Giacinto
f3a114342e
chore(model gallery): add mistralai_mistral-small-3.2-24b-instruct-2506 ( #5714 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-06-24 13:59:14 +02:00
LocalAI [bot]
0d275ccc03
chore: ⬆️ Update ggml-org/llama.cpp to ce82bd0117bd3598300b3a089d13d401b90279c7 ( #5712 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-06-24 08:37:32 +02:00